lucenenet-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From aro...@apache.org
Subject svn commit: r832486 [8/29] - in /incubator/lucene.net/trunk/C#/src: ./ Demo/DeleteFiles/ Demo/DemoLib/ Demo/IndexFiles/ Demo/IndexHtml/ Demo/SearchFiles/ Lucene.Net/ Lucene.Net/Analysis/ Lucene.Net/Document/ Lucene.Net/Index/ Lucene.Net/Search/ Lucene....
Date Tue, 03 Nov 2009 18:06:38 GMT
Modified: incubator/lucene.net/trunk/C#/src/Test/Index/TestIndexReader.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/C%23/src/Test/Index/TestIndexReader.cs?rev=832486&r1=832485&r2=832486&view=diff
==============================================================================
--- incubator/lucene.net/trunk/C#/src/Test/Index/TestIndexReader.cs (original)
+++ incubator/lucene.net/trunk/C#/src/Test/Index/TestIndexReader.cs Tue Nov  3 18:06:27 2009
@@ -1,4 +1,4 @@
-/*
+/* 
  * Licensed to the Apache Software Foundation (ASF) under one or more
  * contributor license agreements.  See the NOTICE file distributed with
  * this work for additional information regarding copyright ownership.
@@ -23,1556 +23,2071 @@
 using StandardAnalyzer = Lucene.Net.Analysis.Standard.StandardAnalyzer;
 using Document = Lucene.Net.Documents.Document;
 using Field = Lucene.Net.Documents.Field;
+using FieldSelector = Lucene.Net.Documents.FieldSelector;
+using Fieldable = Lucene.Net.Documents.Fieldable;
+using SetBasedFieldSelector = Lucene.Net.Documents.SetBasedFieldSelector;
 using FieldOption = Lucene.Net.Index.IndexReader.FieldOption;
-using IndexSearcher = Lucene.Net.Search.IndexSearcher;
-using ScoreDoc = Lucene.Net.Search.ScoreDoc;
-using TermQuery = Lucene.Net.Search.TermQuery;
 using AlreadyClosedException = Lucene.Net.Store.AlreadyClosedException;
 using Directory = Lucene.Net.Store.Directory;
 using FSDirectory = Lucene.Net.Store.FSDirectory;
 using LockObtainFailedException = Lucene.Net.Store.LockObtainFailedException;
 using MockRAMDirectory = Lucene.Net.Store.MockRAMDirectory;
+using NoSuchDirectoryException = Lucene.Net.Store.NoSuchDirectoryException;
 using RAMDirectory = Lucene.Net.Store.RAMDirectory;
+using FieldCache = Lucene.Net.Search.FieldCache;
+using IndexSearcher = Lucene.Net.Search.IndexSearcher;
+using ScoreDoc = Lucene.Net.Search.ScoreDoc;
+using TermQuery = Lucene.Net.Search.TermQuery;
 using LuceneTestCase = Lucene.Net.Util.LuceneTestCase;
 using _TestUtil = Lucene.Net.Util._TestUtil;
 
 namespace Lucene.Net.Index
 {
+	
     [TestFixture]
-    public class TestIndexReader : LuceneTestCase
-    {
-        /// <summary>Main for running test case by itself. </summary>
-        [STAThread]
-        public static void Main(System.String[] args)
-        {
-            // NUnit.Core.TestRunner(new NUnit.Core.TestSuite(typeof(TestIndexReader)));  // {{Aroush}} where is 'TestRunner'?
-            //        TestRunner.run (new TestIndexReader("testBasicDelete"));
-            //        TestRunner.run (new TestIndexReader("testDeleteReaderWriterConflict"));
-            //        TestRunner.run (new TestIndexReader("testDeleteReaderReaderConflict"));
-            //        TestRunner.run (new TestIndexReader("testFilesOpenClose"));
-        }
-
-        // public TestIndexReader(System.String name)
-        // {
-        // }
-
-        public virtual void TestIsCurrent()
-        {
-            RAMDirectory d = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            // set up reader:
-            IndexReader reader = IndexReader.Open(d);
-            Assert.IsTrue(reader.IsCurrent());
-            // modify index by adding another document:
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            Assert.IsFalse(reader.IsCurrent());
-            // re-create index:
-            writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            Assert.IsFalse(reader.IsCurrent());
-            reader.Close();
-            d.Close();
-        }
-
-        /// <summary> Tests the IndexReader.GetFieldNames implementation</summary>
-        /// <throws>  Exception on error </throws>
-        public virtual void TestGetFieldNames()
-        {
-            RAMDirectory d = new MockRAMDirectory();
-            // set up writer
-            IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            // set up reader
-            IndexReader reader = IndexReader.Open(d);
-            System.Collections.Generic.ICollection<string> fieldNames = reader.GetFieldNames(IndexReader.FieldOption.ALL);
-            Assert.IsTrue(fieldNames.Contains("keyword"));
-            Assert.IsTrue(fieldNames.Contains("text"));
-            Assert.IsTrue(fieldNames.Contains("unindexed"));
-            Assert.IsTrue(fieldNames.Contains("unstored"));
-            reader.Close();
-            // add more documents
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            // want to get some more segments here
-            for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
-            {
-                AddDocumentWithFields(writer);
-            }
-            // new fields are in some different segments (we hope)
-            for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
-            {
-                AddDocumentWithDifferentFields(writer);
-            }
-            // new termvector fields
-            for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
-            {
-                AddDocumentWithTermVectorFields(writer);
-            }
-
-            writer.Close();
-            // verify fields again
-            reader = IndexReader.Open(d);
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.ALL);
-            Assert.AreEqual(13, fieldNames.Count); // the following fields
-            Assert.IsTrue(fieldNames.Contains("keyword"));
-            Assert.IsTrue(fieldNames.Contains("text"));
-            Assert.IsTrue(fieldNames.Contains("unindexed"));
-            Assert.IsTrue(fieldNames.Contains("unstored"));
-            Assert.IsTrue(fieldNames.Contains("keyword2"));
-            Assert.IsTrue(fieldNames.Contains("text2"));
-            Assert.IsTrue(fieldNames.Contains("unindexed2"));
-            Assert.IsTrue(fieldNames.Contains("unstored2"));
-            Assert.IsTrue(fieldNames.Contains("tvnot"));
-            Assert.IsTrue(fieldNames.Contains("termvector"));
-            Assert.IsTrue(fieldNames.Contains("tvposition"));
-            Assert.IsTrue(fieldNames.Contains("tvoffset"));
-            Assert.IsTrue(fieldNames.Contains("tvpositionoffset"));
-
-            // verify that only indexed fields were returned
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.INDEXED);
-            Assert.AreEqual(11, fieldNames.Count); // 6 original + the 5 termvector fields 
-            Assert.IsTrue(fieldNames.Contains("keyword"));
-            Assert.IsTrue(fieldNames.Contains("text"));
-            Assert.IsTrue(fieldNames.Contains("unstored"));
-            Assert.IsTrue(fieldNames.Contains("keyword2"));
-            Assert.IsTrue(fieldNames.Contains("text2"));
-            Assert.IsTrue(fieldNames.Contains("unstored2"));
-            Assert.IsTrue(fieldNames.Contains("tvnot"));
-            Assert.IsTrue(fieldNames.Contains("termvector"));
-            Assert.IsTrue(fieldNames.Contains("tvposition"));
-            Assert.IsTrue(fieldNames.Contains("tvoffset"));
-            Assert.IsTrue(fieldNames.Contains("tvpositionoffset"));
-
-            // verify that only unindexed fields were returned
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.UNINDEXED);
-            Assert.AreEqual(2, fieldNames.Count); // the following fields
-            Assert.IsTrue(fieldNames.Contains("unindexed"));
-            Assert.IsTrue(fieldNames.Contains("unindexed2"));
-
-            // verify index term vector fields  
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR);
-            Assert.AreEqual(1, fieldNames.Count); // 1 field has term vector only
-            Assert.IsTrue(fieldNames.Contains("termvector"));
-
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_POSITION);
-            Assert.AreEqual(1, fieldNames.Count); // 4 fields are indexed with term vectors
-            Assert.IsTrue(fieldNames.Contains("tvposition"));
-
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_OFFSET);
-            Assert.AreEqual(1, fieldNames.Count); // 4 fields are indexed with term vectors
-            Assert.IsTrue(fieldNames.Contains("tvoffset"));
-
-            fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_POSITION_OFFSET);
-            Assert.AreEqual(1, fieldNames.Count); // 4 fields are indexed with term vectors
-            Assert.IsTrue(fieldNames.Contains("tvpositionoffset"));
-            reader.Close();
-            d.Close();
-        }
-
-        [Test]
-        public virtual void TestTermVectors()
-        {
-            RAMDirectory d = new MockRAMDirectory();
-            // set up writer
-            IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            // want to get some more segments here
-            // new termvector fields
-            for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
-            {
-                Document doc = new Document();
-                doc.Add(new Field("tvnot", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.NO));
-                doc.Add(new Field("termvector", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.YES));
-                doc.Add(new Field("tvoffset", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_OFFSETS));
-                doc.Add(new Field("tvposition", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS));
-                doc.Add(new Field("tvpositionoffset", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS));
-
-                writer.AddDocument(doc);
-            }
-            writer.Close();
-            IndexReader reader = IndexReader.Open(d);
-            FieldSortedTermVectorMapper mapper = new FieldSortedTermVectorMapper(new TermVectorEntryFreqSortedComparator());
-            reader.GetTermFreqVector(0, mapper);
-            System.Collections.IDictionary map = mapper.GetFieldToTerms();
-            Assert.IsTrue(map != null, "map is null and it shouldn't be");
-            Assert.IsTrue(map.Count == 4, "map Size: " + map.Count + " is not: " + 4);
-            System.Collections.IDictionary set_Renamed = (System.Collections.IDictionary)map["termvector"];
-            for (System.Collections.IEnumerator iterator = set_Renamed.Keys.GetEnumerator(); iterator.MoveNext(); )
-            {
-                TermVectorEntry entry = (TermVectorEntry)iterator.Current;
-                Assert.IsTrue(entry != null, "entry is null and it shouldn't be");
-                System.Console.Out.WriteLine("Entry: " + entry);
-            }
-        }
-
-        private void AssertTermDocsCount(System.String msg, IndexReader reader, Term term, int expected)
-        {
-            TermDocs tdocs = null;
-
-            try
-            {
-                tdocs = reader.TermDocs(term);
-                Assert.IsNotNull(tdocs, msg + ", null TermDocs");
-                int count = 0;
-                while (tdocs.Next())
-                {
-                    count++;
-                }
-                Assert.AreEqual(expected, count, msg + ", count mismatch");
-            }
-            finally
-            {
-                if (tdocs != null)
-                    tdocs.Close();
-            }
-        }
-
-
-        [Test]
-        public virtual void TestBasicDelete()
-        {
-            Directory dir = new MockRAMDirectory();
-
-            IndexWriter writer = null;
-            IndexReader reader = null;
-            Term searchTerm = new Term("content", "aaa");
-
-            //  add 100 documents with term : aaa
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 100; i++)
-            {
-                AddDoc(writer, searchTerm.Text());
-            }
-            writer.Close();
-
-            // OPEN READER AT THIS POINT - this should fix the view of the
-            // index at the point of having 100 "aaa" documents and 0 "bbb"
-            reader = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "first docFreq");
-            AssertTermDocsCount("first reader", reader, searchTerm, 100);
-            reader.Close();
-
-            // DELETE DOCUMENTS CONTAINING TERM: aaa
-            int deleted = 0;
-            reader = IndexReader.Open(dir);
-            deleted = reader.DeleteDocuments(searchTerm);
-            Assert.AreEqual(100, deleted, "deleted count");
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "deleted docFreq");
-            AssertTermDocsCount("deleted termDocs", reader, searchTerm, 0);
-
-            // open a 2nd reader to make sure first reader can
-            // commit its changes (.del) while second reader
-            // is open:
-            IndexReader reader2 = IndexReader.Open(dir);
-            reader.Close();
-
-            // CREATE A NEW READER and re-test
-            reader = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "deleted docFreq");
-            AssertTermDocsCount("deleted termDocs", reader, searchTerm, 0);
-            reader.Close();
-            reader2.Close();
-            dir.Close();
-        }
-
-        // Make sure attempts to make changes after reader is
-        // closed throws IOException:
-        [Test]
-        public virtual void TestChangesAfterClose()
-        {
-            Directory dir = new RAMDirectory();
-
-            IndexWriter writer = null;
-            IndexReader reader = null;
-            Term searchTerm = new Term("content", "aaa");
-
-            //  add 11 documents with term : aaa
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 11; i++)
-            {
-                AddDoc(writer, searchTerm.Text());
-            }
-            writer.Close();
-
-            reader = IndexReader.Open(dir);
-
-            // Close reader:
-            reader.Close();
-
-            // Then, try to make changes:
-            try
-            {
-                reader.DeleteDocument(4);
-                Assert.Fail("deleteDocument after close failed to throw IOException");
-            }
-            catch (AlreadyClosedException)
-            {
-                // expected
-            }
-
-            try
-            {
-                reader.SetNorm(5, "aaa", 2.0f);
-                Assert.Fail("setNorm after close failed to throw IOException");
-            }
-            catch (AlreadyClosedException)
-            {
-                // expected
-            }
-
-            try
-            {
-                reader.UndeleteAll();
-                Assert.Fail("undeleteAll after close failed to throw IOException");
-            }
-            catch (AlreadyClosedException)
-            {
-                // expected
-            }
-        }
-
-        // Make sure we get lock obtain failed exception with 2 writers:
-        [Test]
-        public virtual void TestLockObtainFailed()
-        {
-            Directory dir = new RAMDirectory();
-
-            IndexWriter writer = null;
-            IndexReader reader = null;
-            Term searchTerm = new Term("content", "aaa");
-
-            //  add 11 documents with term : aaa
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 11; i++)
-            {
-                AddDoc(writer, searchTerm.Text());
-            }
-
-            // Create reader:
-            reader = IndexReader.Open(dir);
-
-            // Try to make changes
-            try
-            {
-                reader.DeleteDocument(4);
-                Assert.Fail("deleteDocument should have hit LockObtainFailedException");
-            }
-            catch (LockObtainFailedException)
-            {
-                // expected
-            }
-
-            try
-            {
-                reader.SetNorm(5, "aaa", 2.0f);
-                Assert.Fail("setNorm should have hit LockObtainFailedException");
-            }
-            catch (LockObtainFailedException)
-            {
-                // expected
-            }
-
-            try
-            {
-                reader.UndeleteAll();
-                Assert.Fail("undeleteAll should have hit LockObtainFailedException");
-            }
-            catch (LockObtainFailedException)
-            {
-                // expected
-            }
-            writer.Close();
-            reader.Close();
-        }
-
-        // Make sure you can set norms & commit even if a reader
-        // is open against the index:
-        [Test]
-        public virtual void TestWritingNorms()
-        {
-            System.String tempDir = SupportClass.AppSettings.Get("tempDir", "");
-            if (tempDir == null)
-                throw new System.IO.IOException("tempDir undefined, cannot run test");
-
-            System.IO.FileInfo indexDir = new System.IO.FileInfo(tempDir + "\\" + "lucenetestnormwriter");
-            Directory dir = FSDirectory.GetDirectory(indexDir);
-            IndexWriter writer;
-            IndexReader reader;
-            Term searchTerm = new Term("content", "aaa");
-
-            //  add 1 documents with term : aaa
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDoc(writer, searchTerm.Text());
-            writer.Close();
-
-            //  now open reader & set norm for doc 0
-            reader = IndexReader.Open(dir);
-            reader.SetNorm(0, "content", (float)2.0);
-
-            // we should be holding the write lock now:
-            Assert.IsTrue(IndexReader.IsLocked(dir), "locked");
-
-            reader.Commit();
-
-            // we should not be holding the write lock now:
-            Assert.IsTrue(!IndexReader.IsLocked(dir), "not locked");
-
-            // open a 2nd reader:
-            IndexReader reader2 = IndexReader.Open(dir);
-
-            // set norm again for doc 0
-            reader.SetNorm(0, "content", (float)3.0);
-            Assert.IsTrue(IndexReader.IsLocked(dir), "locked");
-
-            reader.Close();
-
-            // we should not be holding the write lock now:
-            Assert.IsTrue(!IndexReader.IsLocked(dir), "not locked");
-
-            reader2.Close();
-            dir.Close();
-
-            RmDir(indexDir);
-        }
-
-
-        // Make sure you can set norms & commit, and there are
-        // no extra norms files left:
-        [Test]
-        public virtual void TestWritingNormsNoReader()
-        {
-            Directory dir = new MockRAMDirectory();
-            IndexWriter writer = null;
-            IndexReader reader = null;
-            Term searchTerm = new Term("content", "aaa");
-
-            //  add 1 documents with term : aaa
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            writer.SetUseCompoundFile(false);
-            AddDoc(writer, searchTerm.Text());
-            writer.Close();
-
-            //  now open reader & set norm for doc 0 (writes to
-            //  _0_1.s0)
-            reader = IndexReader.Open(dir);
-            reader.SetNorm(0, "content", (float)2.0);
-            reader.Close();
-
-            //  now open reader again & set norm for doc 0 (writes to _0_2.s0)
-            reader = IndexReader.Open(dir);
-            reader.SetNorm(0, "content", (float)2.0);
-            reader.Close();
-            Assert.IsFalse(dir.FileExists("_0_1.s0"), "failed to remove first generation norms file on writing second generation");
-
-            dir.Close();
-        }
-
-
-        [Test]
-        public virtual void TestDeleteReaderWriterConflictUnoptimized()
-        {
-            DeleteReaderWriterConflict(false);
-        }
-
-        [Test]
-        public virtual void TestOpenEmptyDirectory()
-        {
-            System.String dirName = "test.empty";
-            System.IO.FileInfo fileDirName = new System.IO.FileInfo(dirName);
-            bool tmpBool;
-            if (System.IO.File.Exists(fileDirName.FullName))
-                tmpBool = true;
-            else
-                tmpBool = System.IO.Directory.Exists(fileDirName.FullName);
-            if (!tmpBool)
-            {
-                System.IO.Directory.CreateDirectory(fileDirName.FullName);
-            }
-            try
-            {
-                IndexReader.Open(fileDirName);
-                Assert.Fail("opening IndexReader on empty directory failed to produce FileNotFoundException");
-            }
-            catch (System.IO.FileNotFoundException)
-            {
-                // GOOD
-            }
-            RmDir(fileDirName);
-        }
-
-        [Test]
-        public virtual void TestDeleteReaderWriterConflictOptimized()
-        {
-            DeleteReaderWriterConflict(true);
-        }
-
-        private void DeleteReaderWriterConflict(bool optimize)
-        {
-            //Directory dir = new RAMDirectory();
-            Directory dir = GetDirectory();
-
-            Term searchTerm = new Term("content", "aaa");
-            Term searchTerm2 = new Term("content", "bbb");
-
-            //  add 100 documents with term : aaa
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 100; i++)
-            {
-                AddDoc(writer, searchTerm.Text());
-            }
-            writer.Close();
-
-            // OPEN READER AT THIS POINT - this should fix the view of the
-            // index at the point of having 100 "aaa" documents and 0 "bbb"
-            IndexReader reader = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "first docFreq");
-            Assert.AreEqual(0, reader.DocFreq(searchTerm2), "first docFreq");
-            AssertTermDocsCount("first reader", reader, searchTerm, 100);
-            AssertTermDocsCount("first reader", reader, searchTerm2, 0);
-
-            // add 100 documents with term : bbb
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 100; i++)
-            {
-                AddDoc(writer, searchTerm2.Text());
-            }
-
-            // REQUEST OPTIMIZATION
-            // This causes a new segment to become current for all subsequent
-            // searchers. Because of this, deletions made via a previously open
-            // reader, which would be applied to that reader's segment, are lost
-            // for subsequent searchers/readers
-            if (optimize)
-                writer.Optimize();
-            writer.Close();
-
-            // The reader should not see the new data
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "first docFreq");
-            Assert.AreEqual(0, reader.DocFreq(searchTerm2), "first docFreq");
-            AssertTermDocsCount("first reader", reader, searchTerm, 100);
-            AssertTermDocsCount("first reader", reader, searchTerm2, 0);
-
-
-            // DELETE DOCUMENTS CONTAINING TERM: aaa
-            // NOTE: the reader was created when only "aaa" documents were in
-            int deleted = 0;
-            try
-            {
-                deleted = reader.DeleteDocuments(searchTerm);
-                Assert.Fail("Delete allowed on an index reader with stale segment information");
-            }
-            catch (StaleReaderException)
-            {
-                /* success */
-            }
-
-            // Re-open index reader and try again. This time it should see
-            // the new data.
-            reader.Close();
-            reader = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "first docFreq");
-            Assert.AreEqual(100, reader.DocFreq(searchTerm2), "first docFreq");
-            AssertTermDocsCount("first reader", reader, searchTerm, 100);
-            AssertTermDocsCount("first reader", reader, searchTerm2, 100);
-
-            deleted = reader.DeleteDocuments(searchTerm);
-            Assert.AreEqual(100, deleted, "deleted count");
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "deleted docFreq");
-            Assert.AreEqual(100, reader.DocFreq(searchTerm2), "deleted docFreq");
-            AssertTermDocsCount("deleted termDocs", reader, searchTerm, 0);
-            AssertTermDocsCount("deleted termDocs", reader, searchTerm2, 100);
-            reader.Close();
-
-            // CREATE A NEW READER and re-test
-            reader = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader.DocFreq(searchTerm), "deleted docFreq");
-            Assert.AreEqual(100, reader.DocFreq(searchTerm2), "deleted docFreq");
-            AssertTermDocsCount("deleted termDocs", reader, searchTerm, 0);
-            AssertTermDocsCount("deleted termDocs", reader, searchTerm2, 100);
-            reader.Close();
-        }
-
-        private Directory GetDirectory()
-        {
-            return FSDirectory.GetDirectory(new System.IO.FileInfo(System.IO.Path.Combine(SupportClass.AppSettings.Get("tempDir", ""), "testIndex")));
-        }
-
-        [Test]
-        public virtual void TestFilesOpenClose()
-        {
-            // Create initial data set
-            System.IO.FileInfo dirFile = new System.IO.FileInfo(System.IO.Path.Combine("tempDir", "testIndex"));
-            Directory dir = GetDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDoc(writer, "test");
-            writer.Close();
-            dir.Close();
-
-            // Try to erase the data - this ensures that the writer closed all files
-            _TestUtil.RmDir(dirFile);
-            dir = GetDirectory();
-
-            // Now create the data set again, just as before
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDoc(writer, "test");
-            writer.Close();
-            dir.Close();
-
-            // Now open existing directory and test that reader closes all files
-            dir = GetDirectory();
-            IndexReader reader1 = IndexReader.Open(dir);
-            reader1.Close();
-            dir.Close();
-
-            // The following will fail if reader did not Close
-            // all files
-            _TestUtil.RmDir(dirFile);
-        }
-
-        [Test]
-        public virtual void TestLastModified()
-        {
-            Assert.IsFalse(IndexReader.IndexExists("there_is_no_such_index"));
-            System.IO.FileInfo fileDir = new System.IO.FileInfo(System.IO.Path.Combine(SupportClass.AppSettings.Get("tempDir", ""), "testIndex"));
-            // can't do the filesystem version of this test, as a system level process lock prevents deletion of the index file
-            //for (int i = 0; i < 2; i++)
-            for (int i = 0; i < 1; i++)
-            {
-                try
-                {
-                    Directory dir;
-                    if (0 == i)
-                        dir = new MockRAMDirectory();
-                    else
-                        dir = GetDirectory();
-                    Assert.IsFalse(IndexReader.IndexExists(dir));
-                    IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-                    AddDocumentWithFields(writer);
-                    Assert.IsTrue(IndexReader.IsLocked(dir)); // writer open, so dir is locked
-                    writer.Close();
-                    Assert.IsTrue(IndexReader.IndexExists(dir));
-                    IndexReader reader = IndexReader.Open(dir);
-                    Assert.IsFalse(IndexReader.IsLocked(dir)); // reader only, no lock
-                    long version = IndexReader.LastModified(dir);
-                    if (i == 1)
-                    {
-                        long version2 = IndexReader.LastModified(fileDir);
-                        Assert.AreEqual(version, version2);
-                    }
-                    reader.Close();
-                    // modify index and check version has been
-                    // incremented:
-                    while (true)
-                    {
-                        try
-                        {
-                            System.Threading.Thread.Sleep(new System.TimeSpan((System.Int64)10000 * 1000));
-                            break;
-                        }
-                        catch (System.Threading.ThreadInterruptedException)
-                        {
-                            SupportClass.ThreadClass.Current().Interrupt();
-                        }
-                    }
-
-                    writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-                    AddDocumentWithFields(writer);
-                    writer.Close();
-                    reader = IndexReader.Open(dir);
-                    Assert.IsTrue(
-                        version <= IndexReader.LastModified(dir),
-                        "old lastModified is " + version + "; new lastModified is " + IndexReader.LastModified(dir)
-                    );
-                    reader.Close();
-                    dir.Close();
-                }
-                finally
-                {
-                    if (i == 1)
-                        _TestUtil.RmDir(fileDir);
-                }
-            }
-        }
-
-        [Test]
-        public virtual void TestVersion()
-        {
-            Assert.IsFalse(IndexReader.IndexExists("there_is_no_such_index"));
-            Directory dir = new MockRAMDirectory();
-            Assert.IsFalse(IndexReader.IndexExists(dir));
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            Assert.IsTrue(IndexReader.IsLocked(dir)); // writer open, so dir is locked
-            writer.Close();
-            Assert.IsTrue(IndexReader.IndexExists(dir));
-            IndexReader reader = IndexReader.Open(dir);
-            Assert.IsFalse(IndexReader.IsLocked(dir)); // reader only, no lock
-            long version = IndexReader.GetCurrentVersion(dir);
-            reader.Close();
-            // modify index and check version has been
-            // incremented:
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            reader = IndexReader.Open(dir);
-            Assert.IsTrue(version < IndexReader.GetCurrentVersion(dir), "old version is " + version + "; new version is " + IndexReader.GetCurrentVersion(dir));
-            reader.Close();
-            dir.Close();
-        }
-
-        [Test]
-        public virtual void TestLock()
-        {
-            Directory dir = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            IndexReader reader = IndexReader.Open(dir);
-            try
-            {
-                reader.DeleteDocument(0);
-                Assert.Fail("expected lock");
-            }
-            catch (System.IO.IOException)
-            {
-                // expected exception
-            }
-            IndexReader.Unlock(dir); // this should not be done in the real world! 
-            reader.DeleteDocument(0);
-            reader.Close();
-            writer.Close();
-            dir.Close();
-        }
-
-        [Test]
-        public virtual void TestUndeleteAll()
-        {
-            Directory dir = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            IndexReader reader = IndexReader.Open(dir);
-            reader.DeleteDocument(0);
-            reader.DeleteDocument(1);
-            reader.UndeleteAll();
-            reader.Close();
-            reader = IndexReader.Open(dir);
-            Assert.AreEqual(2, reader.NumDocs()); // nothing has really been deleted thanks to undeleteAll()
-            reader.Close();
-            dir.Close();
-        }
-
-        [Test]
-        public virtual void TestUndeleteAllAfterClose()
-        {
-            Directory dir = new RAMDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            IndexReader reader = IndexReader.Open(dir);
-            reader.DeleteDocument(0);
-            reader.DeleteDocument(1);
-            reader.Close();
-            reader = IndexReader.Open(dir);
-            reader.UndeleteAll();
-            Assert.AreEqual(2, reader.NumDocs()); // nothing has really been deleted thanks to undeleteAll()
-            reader.Close();
-            dir.Close();
-        }
-
-        [Test]
-        public virtual void TestUndeleteAllAfterCloseThenReopen()
-        {
-            Directory dir = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            AddDocumentWithFields(writer);
-            writer.Close();
-            IndexReader reader = IndexReader.Open(dir);
-            reader.DeleteDocument(0);
-            reader.DeleteDocument(1);
-            reader.Close();
-            reader = IndexReader.Open(dir);
-            reader.UndeleteAll();
-            reader.Close();
-            reader = IndexReader.Open(dir);
-            Assert.AreEqual(2, reader.NumDocs()); // nothing has really been deleted thanks to undeleteAll()
-            reader.Close();
-            dir.Close();
-        }
-
-        [Test]
-        public virtual void TestDeleteReaderReaderConflictUnoptimized()
-        {
-            DeleteReaderReaderConflict(false);
-        }
-
-        [Test]
-        public virtual void TestDeleteReaderReaderConflictOptimized()
-        {
-            DeleteReaderReaderConflict(true);
-        }
-
-        /// <summary> Make sure if reader tries to commit but hits disk
-        /// full that reader remains consistent and usable.
-        /// </summary>
-        [Test]
-        public virtual void TestDiskFull()
-        {
-
-            bool debug = false;
-            Term searchTerm = new Term("content", "aaa");
-            int START_COUNT = 157;
-            int END_COUNT = 144;
-
-            // First build up a starting index:
-            RAMDirectory startDir = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(startDir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 157; i++)
-            {
-                Lucene.Net.Documents.Document d = new Lucene.Net.Documents.Document();
-                d.Add(new Field("id", System.Convert.ToString(i), Field.Store.YES, Field.Index.NOT_ANALYZED));
-                d.Add(new Field("content", "aaa " + i, Field.Store.NO, Field.Index.ANALYZED));
-                writer.AddDocument(d);
-            }
-            writer.Close();
-
-            long diskUsage = startDir.SizeInBytes();
-            long diskFree = diskUsage + 100;
-
-            System.IO.IOException err = null;
-
-            bool done = false;
-
-            // Iterate w/ ever increasing free disk space:
-            while (!done)
-            {
-                MockRAMDirectory dir = new MockRAMDirectory(startDir);
-
-                // If IndexReader hits disk full, it can write to the same files again.
-                dir.SetPreventDoubleWrite(false);
-
-                IndexReader reader = IndexReader.Open(dir);
-
-                // For each disk size, first try to commit against
-                // dir that will hit random IOExceptions & disk
-                // full; after, give it infinite disk space & turn
-                // off random IOExceptions & retry w/ same reader:
-                bool success = false;
-
-                for (int x = 0; x < 2; x++)
-                {
-
-                    double rate = 0.05;
-                    double diskRatio = ((double)diskFree) / diskUsage;
-                    long thisDiskFree;
-                    System.String testName;
-
-                    if (0 == x)
-                    {
-                        thisDiskFree = diskFree;
-                        if (diskRatio >= 2.0)
-                        {
-                            rate /= 2;
-                        }
-                        if (diskRatio >= 4.0)
-                        {
-                            rate /= 2;
-                        }
-                        if (diskRatio >= 6.0)
-                        {
-                            rate = 0.0;
-                        }
-                        if (debug)
-                        {
-                            System.Console.Out.WriteLine("\ncycle: " + diskFree + " bytes");
-                        }
-                        testName = "disk full during reader.Close() @ " + thisDiskFree + " bytes";
-                    }
-                    else
-                    {
-                        thisDiskFree = 0;
-                        rate = 0.0;
-                        if (debug)
-                        {
-                            System.Console.Out.WriteLine("\ncycle: same writer: unlimited disk space");
-                        }
-                        testName = "reader re-use after disk full";
-                    }
-
-                    dir.SetMaxSizeInBytes(thisDiskFree);
-                    dir.SetRandomIOExceptionRate(rate, diskFree);
-
-                    try
-                    {
-                        if (0 == x)
-                        {
-                            int docId = 12;
-                            for (int i = 0; i < 13; i++)
-                            {
-                                reader.DeleteDocument(docId);
-                                reader.SetNorm(docId, "contents", (float)2.0);
-                                docId += 12;
-                            }
-                        }
-                        reader.Close();
-                        success = true;
-                        if (0 == x)
-                        {
-                            done = true;
-                        }
-                    }
-                    catch (System.IO.IOException e)
-                    {
-                        if (debug)
-                        {
-                            System.Console.Out.WriteLine("  hit IOException: " + e);
-                            System.Console.Out.WriteLine(e.StackTrace);
-                        }
-                        err = e;
-                        if (1 == x)
-                        {
-                            System.Console.Error.WriteLine(e.StackTrace);
-                            Assert.Fail(testName + " hit IOException after disk space was freed up");
-                        }
-                    }
-
-                    // Whether we succeeded or failed, check that all
-                    // un-referenced files were in fact deleted (ie,
-                    // we did not create garbage).  Just create a
-                    // new IndexFileDeleter, have it delete
-                    // unreferenced files, then verify that in fact
-                    // no files were deleted:
-                    System.String[] startFiles = dir.List();
-                    SegmentInfos infos = new SegmentInfos();
-                    infos.Read(dir);
-                    new IndexFileDeleter(dir, new KeepOnlyLastCommitDeletionPolicy(), infos, null, null);
-                    System.String[] endFiles = dir.List();
-
-                    System.Array.Sort(startFiles);
-                    System.Array.Sort(endFiles);
-
-                    //for(int i=0;i<startFiles.length;i++) {
-                    //  System.out.println("  startFiles: " + i + ": " + startFiles[i]);
-                    //}
-
-                    if (SupportClass.Compare.CompareStringArrays(startFiles, endFiles) == false)
-                    {
-                        System.String successStr;
-                        if (success)
-                        {
-                            successStr = "success";
-                        }
-                        else
-                        {
-                            successStr = "IOException";
-                            System.Console.Error.WriteLine(err.StackTrace);
-                        }
-                        Assert.Fail("reader.Close() failed to delete unreferenced files after " + successStr + " (" + diskFree + " bytes): before delete:\n    " + ArrayToString(startFiles) + "\n  after delete:\n    " + ArrayToString(endFiles));
-                    }
-
-                    // Finally, verify index is not corrupt, and, if
-                    // we succeeded, we see all docs changed, and if
-                    // we failed, we see either all docs or no docs
-                    // changed (transactional semantics):
-                    IndexReader newReader = null;
-                    try
-                    {
-                        newReader = IndexReader.Open(dir);
-                    }
-                    catch (System.IO.IOException e)
-                    {
-                        System.Console.Error.WriteLine(e.StackTrace);
-                        Assert.Fail(testName + ":exception when creating IndexReader after disk full during Close: " + e);
-                    }
-                    /*
-                    int result = newReader.docFreq(searchTerm);
-                    if (success) {
-                    if (result != END_COUNT) {
-                    fail(testName + ": method did not throw exception but docFreq('aaa') is " + result + " instead of expected " + END_COUNT);
-                    }
-                    } else {
-                    // On hitting exception we still may have added
-                    // all docs:
-                    if (result != START_COUNT && result != END_COUNT) {
-                    err.printStackTrace();
-                    fail(testName + ": method did throw exception but docFreq('aaa') is " + result + " instead of expected " + START_COUNT + " or " + END_COUNT);
-                    }
-                    }
-                    */
-
-                    IndexSearcher searcher = new IndexSearcher(newReader);
-                    ScoreDoc[] hits = null;
-                    try
-                    {
-                        hits = searcher.Search(new TermQuery(searchTerm), null, 1000).scoreDocs;
-                    }
-                    catch (System.IO.IOException e)
-                    {
-                        System.Console.Error.WriteLine(e.StackTrace);
-                        Assert.Fail(testName + ": exception when searching: " + e);
-                    }
-                    int result2 = hits.Length;
-                    if (success)
-                    {
-                        if (result2 != END_COUNT)
-                        {
-                            Assert.Fail(testName + ": method did not throw exception but hits.length for search on term 'aaa' is " + result2 + " instead of expected " + END_COUNT);
-                        }
-                    }
-                    else
-                    {
-                        // On hitting exception we still may have added
-                        // all docs:
-                        if (result2 != START_COUNT && result2 != END_COUNT)
-                        {
-                            System.Console.Error.WriteLine(err.StackTrace);
-                            Assert.Fail(testName + ": method did throw exception but hits.length for search on term 'aaa' is " + result2 + " instead of expected " + START_COUNT);
-                        }
-                    }
-
-                    searcher.Close();
-                    newReader.Close();
-
-                    if (result2 == END_COUNT)
-                    {
-                        break;
-                    }
-                }
-
-                dir.Close();
-
-                // Try again with 10 more bytes of free space:
-                diskFree += 10;
-            }
-
-            startDir.Close();
-        }
-
-        [Test]
-        public virtual void TestDocsOutOfOrderJIRA140()
-        {
-            Directory dir = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 11; i++)
-            {
-                AddDoc(writer, "aaa");
-            }
-            writer.Close();
-            IndexReader reader = IndexReader.Open(dir);
-
-            // Try to delete an invalid docId, yet, within range
-            // of the final bits of the BitVector:
-
-            bool gotException = false;
-            try
-            {
-                reader.DeleteDocument(11);
-            }
-            catch (System.IndexOutOfRangeException)
-            {
-                gotException = true;
-            }
-            reader.Close();
-
-            writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-
-            // We must add more docs to get a new segment written
-            for (int i = 0; i < 11; i++)
-            {
-                AddDoc(writer, "aaa");
-            }
-
-            // Without the fix for LUCENE-140 this call will
-            // [incorrectly] hit a "docs out of order"
-            // IllegalStateException because above out-of-bounds
-            // deleteDocument corrupted the index:
-            writer.Optimize();
-
-            if (!gotException)
-            {
-                Assert.Fail("delete of out-of-bounds doc number failed to hit exception");
-            }
-            dir.Close();
-        }
-
-        [Test]
-        public virtual void TestExceptionReleaseWriteLockJIRA768()
-        {
-
-            Directory dir = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDoc(writer, "aaa");
-            writer.Close();
-
-            IndexReader reader = IndexReader.Open(dir);
-            try
-            {
-                reader.DeleteDocument(1);
-                Assert.Fail("did not hit exception when deleting an invalid doc number");
-            }
-            catch (System.IndexOutOfRangeException)
-            {
-                // expected
-            }
-            reader.Close();
-            if (IndexReader.IsLocked(dir))
-            {
-                Assert.Fail("write lock is still held after Close");
-            }
-
-            reader = IndexReader.Open(dir);
-            try
-            {
-                reader.SetNorm(1, "content", (float)2.0);
-                Assert.Fail("did not hit exception when calling setNorm on an invalid doc number");
-            }
-            catch (System.IndexOutOfRangeException)
-            {
-                // expected
-            }
-            reader.Close();
-            if (IndexReader.IsLocked(dir))
-            {
-                Assert.Fail("write lock is still held after Close");
-            }
-            dir.Close();
-        }
-
-        private System.String ArrayToString(System.String[] l)
-        {
-            System.String s = "";
-            for (int i = 0; i < l.Length; i++)
-            {
-                if (i > 0)
-                {
-                    s += "\n    ";
-                }
-                s += l[i];
-            }
-            return s;
-        }
-
-        [Test]
-        public virtual void TestOpenReaderAfterDelete()
-        {
-            System.IO.FileInfo dirFile = new System.IO.FileInfo(System.IO.Path.Combine(SupportClass.AppSettings.Get("tempDir", ""), "deletetest"));
-            Directory dir = FSDirectory.GetDirectory(dirFile);
-            try
-            {
-                IndexReader.Open(dir);
-                Assert.Fail("expected FileNotFoundException");
-            }
-            catch (System.IO.FileNotFoundException)
-            {
-                // expected
-            }
-
-            System.IO.Directory.Delete(dirFile.FullName);
-
-            // Make sure we still get a CorruptIndexException (not NPE):
-            try
-            {
-                IndexReader.Open(dir);
-                Assert.Fail("expected FileNotFoundException");
-            }
-            catch (System.IO.DirectoryNotFoundException)
-            {
-                // expected
-            }
-        }
-
-        private void DeleteReaderReaderConflict(bool optimize)
-        {
-            Directory dir = GetDirectory();
-
-            Term searchTerm1 = new Term("content", "aaa");
-            Term searchTerm2 = new Term("content", "bbb");
-            Term searchTerm3 = new Term("content", "ccc");
-
-            //  add 100 documents with term : aaa
-            //  add 100 documents with term : bbb
-            //  add 100 documents with term : ccc
-            IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            for (int i = 0; i < 100; i++)
-            {
-                AddDoc(writer, searchTerm1.Text());
-                AddDoc(writer, searchTerm2.Text());
-                AddDoc(writer, searchTerm3.Text());
-            }
-            if (optimize)
-                writer.Optimize();
-            writer.Close();
-
-            // OPEN TWO READERS
-            // Both readers get segment info as exists at this time
-            IndexReader reader1 = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm1), "first opened");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm2), "first opened");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm3), "first opened");
-            AssertTermDocsCount("first opened", reader1, searchTerm1, 100);
-            AssertTermDocsCount("first opened", reader1, searchTerm2, 100);
-            AssertTermDocsCount("first opened", reader1, searchTerm3, 100);
-
-            IndexReader reader2 = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm1), "first opened");
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm2), "first opened");
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm3), "first opened");
-            AssertTermDocsCount("first opened", reader2, searchTerm1, 100);
-            AssertTermDocsCount("first opened", reader2, searchTerm2, 100);
-            AssertTermDocsCount("first opened", reader2, searchTerm3, 100);
-
-            // DELETE DOCS FROM READER 2 and CLOSE IT
-            // delete documents containing term: aaa
-            // when the reader is closed, the segment info is updated and
-            // the first reader is now stale
-            reader2.DeleteDocuments(searchTerm1);
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm1), "after delete 1");
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm2), "after delete 1");
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm3), "after delete 1");
-            AssertTermDocsCount("after delete 1", reader2, searchTerm1, 0);
-            AssertTermDocsCount("after delete 1", reader2, searchTerm2, 100);
-            AssertTermDocsCount("after delete 1", reader2, searchTerm3, 100);
-            reader2.Close();
-
-            // Make sure reader 1 is unchanged since it was open earlier
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm1), "after delete 1");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm2), "after delete 1");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm3), "after delete 1");
-            AssertTermDocsCount("after delete 1", reader1, searchTerm1, 100);
-            AssertTermDocsCount("after delete 1", reader1, searchTerm2, 100);
-            AssertTermDocsCount("after delete 1", reader1, searchTerm3, 100);
-
-
-            // ATTEMPT TO DELETE FROM STALE READER
-            // delete documents containing term: bbb
-            try
-            {
-                reader1.DeleteDocuments(searchTerm2);
-                Assert.Fail("Delete allowed from a stale index reader");
-            }
-            catch (System.IO.IOException)
-            {
-                /* success */
-            }
-
-            // RECREATE READER AND TRY AGAIN
-            reader1.Close();
-            reader1 = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm1), "reopened");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm2), "reopened");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm3), "reopened");
-            AssertTermDocsCount("reopened", reader1, searchTerm1, 0);
-            AssertTermDocsCount("reopened", reader1, searchTerm2, 100);
-            AssertTermDocsCount("reopened", reader1, searchTerm3, 100);
-
-            reader1.DeleteDocuments(searchTerm2);
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm1), "deleted 2");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm2), "deleted 2");
-            Assert.AreEqual(100, reader1.DocFreq(searchTerm3), "deleted 2");
-            AssertTermDocsCount("deleted 2", reader1, searchTerm1, 0);
-            AssertTermDocsCount("deleted 2", reader1, searchTerm2, 0);
-            AssertTermDocsCount("deleted 2", reader1, searchTerm3, 100);
-            reader1.Close();
-
-            // Open another reader to confirm that everything is deleted
-            reader2 = IndexReader.Open(dir);
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm1), "reopened 2");
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm2), "reopened 2");
-            Assert.AreEqual(100, reader2.DocFreq(searchTerm3), "reopened 2");
-            AssertTermDocsCount("reopened 2", reader2, searchTerm1, 0);
-            AssertTermDocsCount("reopened 2", reader2, searchTerm2, 0);
-            AssertTermDocsCount("reopened 2", reader2, searchTerm3, 100);
-            reader2.Close();
-
-            dir.Close();
-        }
-
-
-        private void AddDocumentWithFields(IndexWriter writer)
-        {
-            Lucene.Net.Documents.Document doc = new Lucene.Net.Documents.Document();
-            doc.Add(new Field("keyword", "test1", Field.Store.YES, Field.Index.NOT_ANALYZED));
-            doc.Add(new Field("text", "test1", Field.Store.YES, Field.Index.ANALYZED));
-            doc.Add(new Field("unindexed", "test1", Field.Store.YES, Field.Index.NO));
-            doc.Add(new Field("unstored", "test1", Field.Store.NO, Field.Index.ANALYZED));
-            writer.AddDocument(doc);
-        }
-
-        private void AddDocumentWithDifferentFields(IndexWriter writer)
-        {
-            Lucene.Net.Documents.Document doc = new Lucene.Net.Documents.Document();
-            doc.Add(new Field("keyword2", "test1", Field.Store.YES, Field.Index.NOT_ANALYZED));
-            doc.Add(new Field("text2", "test1", Field.Store.YES, Field.Index.ANALYZED));
-            doc.Add(new Field("unindexed2", "test1", Field.Store.YES, Field.Index.NO));
-            doc.Add(new Field("unstored2", "test1", Field.Store.NO, Field.Index.ANALYZED));
-            writer.AddDocument(doc);
-        }
-
-        private void AddDocumentWithTermVectorFields(IndexWriter writer)
-        {
-            Lucene.Net.Documents.Document doc = new Lucene.Net.Documents.Document();
-            doc.Add(new Field("tvnot", "tvnot", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.NO));
-            doc.Add(new Field("termvector", "termvector", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.YES));
-            doc.Add(new Field("tvoffset", "tvoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_OFFSETS));
-            doc.Add(new Field("tvposition", "tvposition", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS));
-            doc.Add(new Field("tvpositionoffset", "tvpositionoffset", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS));
-
-            writer.AddDocument(doc);
-        }
-
-        private void AddDoc(IndexWriter writer, System.String value_Renamed)
-        {
-            Lucene.Net.Documents.Document doc = new Lucene.Net.Documents.Document();
-            doc.Add(new Field("content", value_Renamed, Field.Store.NO, Field.Index.ANALYZED));
-            writer.AddDocument(doc);
-        }
-
-        private void RmDir(System.IO.FileInfo dir)
-        {
-            System.IO.FileInfo[] files = SupportClass.FileSupport.GetFiles(dir);
-            for (int i = 0; i < files.Length; i++)
-            {
-                bool tmpBool;
-                if (System.IO.File.Exists(files[i].FullName))
-                {
-                    System.IO.File.Delete(files[i].FullName);
-                    tmpBool = true;
-                }
-                else if (System.IO.Directory.Exists(files[i].FullName))
-                {
-                    System.IO.Directory.Delete(files[i].FullName);
-                    tmpBool = true;
-                }
-                else
-                    tmpBool = false;
-                bool generatedAux = tmpBool;
-            }
-            bool tmpBool2;
-            if (System.IO.File.Exists(dir.FullName))
-            {
-                System.IO.File.Delete(dir.FullName);
-                tmpBool2 = true;
-            }
-            else if (System.IO.Directory.Exists(dir.FullName))
-            {
-                System.IO.Directory.Delete(dir.FullName);
-                tmpBool2 = true;
-            }
-            else
-                tmpBool2 = false;
-            bool generatedAux2 = tmpBool2;
-        }
-
-        public static void AssertIndexEquals(IndexReader index1, IndexReader index2)
-        {
-            Assert.AreEqual(index1.NumDocs(), index2.NumDocs(), "IndexReaders have different values for numDocs.");
-            Assert.AreEqual(index1.MaxDoc(), index2.MaxDoc(), "IndexReaders have different values for maxDoc.");
-            Assert.AreEqual(index1.HasDeletions(), index2.HasDeletions(), "Only one IndexReader has deletions.");
-            Assert.AreEqual(index1.IsOptimized(), index2.IsOptimized(), "Only one index is optimized.");
-
-            // check field names
-            System.Collections.Generic.ICollection<string> fields1 = index1.GetFieldNames(FieldOption.ALL);
-            System.Collections.Generic.ICollection<string> fields2 = index2.GetFieldNames(FieldOption.ALL);
-            Assert.AreEqual(fields1.Count, fields2.Count, "IndexReaders have different numbers of fields.");
-            System.Collections.Generic.IEnumerator<string> it1 = fields1.GetEnumerator();
-            System.Collections.Generic.IEnumerator<string> it2 = fields2.GetEnumerator();
-            while (it1.MoveNext())
-            {
-                Assert.IsTrue(it2.MoveNext());
-                Assert.AreEqual(it1.Current, it2.Current, "Different field names.");
-            }
-
-            // check norms
-            it1 = fields1.GetEnumerator();
-            while (it1.MoveNext())
-            {
-                System.String curField = (System.String)it1.Current;
-                byte[] norms1 = index1.Norms(curField);
-                byte[] norms2 = index2.Norms(curField);
-                Assert.AreEqual(norms1.Length, norms2.Length);
-                for (int i = 0; i < norms1.Length; i++)
-                {
-                    if (norms1[i] == 0)
-                        System.Console.Error.WriteLine(System.Threading.Thread.CurrentThread.Name + ": new IndexReader (" + index1.GetType().ToString() + ") contains norm=0 for doc " + i);
-                    if (norms2[i] == 0)
-                        System.Console.Error.WriteLine(System.Threading.Thread.CurrentThread.Name + ": refreshed IndexReader (" + index2.GetType().ToString() + ") contains norm=0 for doc " + i);
-                    Assert.AreEqual(norms1[i], norms2[i], "Norm different for doc " + i + " and field '" + curField + "'.");
-                }
-            }
-
-            // check deletions
-            for (int i = 0; i < index1.MaxDoc(); i++)
-            {
-                Assert.AreEqual(index1.IsDeleted(i), index2.IsDeleted(i), "Doc " + i + " only deleted in one index.");
-            }
-
-            // check stored fields
-            for (int i = 0; i < index1.MaxDoc(); i++)
-            {
-                if (!index1.IsDeleted(i))
-                {
-                    Document doc1 = index1.Document(i);
-                    Document doc2 = index2.Document(i);
-                    System.Collections.IList _fields1 = doc1.GetFields();
-                    System.Collections.IList _fields2 = doc2.GetFields();
-                    Assert.AreEqual(fields1.Count, fields2.Count, "Different numbers of fields for doc " + i + ".");
-                    System.Collections.IEnumerator _it1 = _fields1.GetEnumerator();
-                    System.Collections.IEnumerator _it2 = _fields2.GetEnumerator();
-                    while (_it1.MoveNext())
-                    {
-                        Assert.IsTrue(_it2.MoveNext());
-                        Field curField1 = (Field)_it1.Current;
-                        Field curField2 = (Field)_it2.Current;
-                        Assert.AreEqual(curField1.Name(), curField2.Name(), "Different fields names for doc " + i + ".");
-                        Assert.AreEqual(curField1.StringValue(), curField2.StringValue(), "Different field values for doc " + i + ".");
-                    }
-                }
-            }
-
-            // check dictionary and posting lists
-            TermEnum enum1 = index1.Terms();
-            TermEnum enum2 = index2.Terms();
-            TermPositions tp1 = index1.TermPositions();
-            TermPositions tp2 = index2.TermPositions();
-            while (enum1.Next())
-            {
-                Assert.IsTrue(enum2.Next());
-                Assert.AreEqual(enum1.Term(), enum2.Term(), "Different term in dictionary.");
-                tp1.Seek(enum1.Term());
-                tp2.Seek(enum1.Term());
-                while (tp1.Next())
-                {
-                    Assert.IsTrue(tp2.Next());
-                    Assert.AreEqual(tp1.Doc(), tp2.Doc(), "Different doc id in postinglist of term " + enum1.Term() + ".");
-                    Assert.AreEqual(tp1.Freq(), tp2.Freq(), "Different term frequence in postinglist of term " + enum1.Term() + ".");
-                    for (int i = 0; i < tp1.Freq(); i++)
-                    {
-                        Assert.AreEqual(tp1.NextPosition(), tp2.NextPosition(), "Different positions in postinglist of term " + enum1.Term() + ".");
-                    }
-                }
-            }
-        }
-
-        [Test]
-        public void TestGetIndexCommit()
-        {
-
-            RAMDirectory d = new MockRAMDirectory();
-
-            // set up writer
-            IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            writer.SetMaxBufferedDocs(2);
-            for (int i = 0; i < 27; i++)
-                AddDocumentWithFields(writer);
-            writer.Close();
-
-            SegmentInfos sis = new SegmentInfos();
-            sis.Read(d);
-            IndexReader r = IndexReader.Open(d);
-            IndexCommit c = r.GetIndexCommit();
-
-            Assert.AreEqual(sis.GetCurrentSegmentFileName(), c.GetSegmentsFileName());
-
-            Assert.IsTrue(c.Equals(r.GetIndexCommit()));
-
-            // Change the index
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            writer.SetMaxBufferedDocs(2);
-            for (int i = 0; i < 7; i++)
-                AddDocumentWithFields(writer);
-            writer.Close();
-
-            IndexReader r2 = r.Reopen();
-            Assert.IsFalse(c.Equals(r2.GetIndexCommit()));
-            Assert.IsFalse(r2.GetIndexCommit().IsOptimized());
-            r2.Close();
-
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            writer.Optimize();
-            writer.Close();
-
-            r2 = r.Reopen();
-            Assert.IsTrue(r2.GetIndexCommit().IsOptimized());
-
-            r.Close();
-            r2.Close();
-            d.Close();
-        }
-
-        [Test]
-        public void TestReadOnly()
-        {
-            RAMDirectory d = new MockRAMDirectory();
-            IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Commit();
-            AddDocumentWithFields(writer);
-            writer.Close();
-
-            IndexReader r = IndexReader.Open(d, true);
-            try
-            {
-                r.DeleteDocument(0);
-                Assert.Fail();
-            }
-            catch (System.Exception)
-            {
-                // expected
-            }
-
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            AddDocumentWithFields(writer);
-            writer.Close();
-
-            // Make sure reopen is still readonly:
-            IndexReader r2 = r.Reopen();
-            r.Close();
-
-            Assert.IsFalse(r == r2);
-
-            try
-            {
-                r2.DeleteDocument(0);
-                Assert.Fail();
-            }
-            catch (System.Exception)
-            {
-                // expected
-            }
-
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            writer.Optimize();
-            writer.Close();
-
-            // Make sure reopen to a single segment is still readonly:
-            IndexReader r3 = r2.Reopen();
-            r2.Close();
-
-            Assert.IsFalse(r == r2);
-
-            try
-            {
-                r3.DeleteDocument(0);
-                Assert.Fail();
-            }
-            catch (System.Exception)
-            {
-                // expected
-            }
-
-            // Make sure write lock isn't held
-            writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
-            writer.Close();
-
-            r3.Close();
-        }
-    }
+	public class TestIndexReader:LuceneTestCase
+	{
+		/// <summary>Main for running test case by itself. </summary>
+		[STAThread]
+		public static void  Main(System.String[] args)
+		{
+			// TestRunner.run(new TestSuite(typeof(TestIndexReader))); // {{Aroush-2.9}} how is this done in NUnit?
+			//        TestRunner.run (new TestIndexReader("testBasicDelete"));
+			//        TestRunner.run (new TestIndexReader("testDeleteReaderWriterConflict"));
+			//        TestRunner.run (new TestIndexReader("testDeleteReaderReaderConflict"));
+			//        TestRunner.run (new TestIndexReader("testFilesOpenClose"));
+		}
+		
+		public TestIndexReader(System.String name):base(name)
+		{
+		}
+		
+		[Test]
+		public virtual void  TestCommitUserData()
+		{
+			RAMDirectory d = new MockRAMDirectory();
+			
+			System.Collections.IDictionary commitUserData = new System.Collections.Hashtable();
+			commitUserData["foo"] = "fighters";
+			
+			// set up writer
+			IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			writer.SetMaxBufferedDocs(2);
+			for (int i = 0; i < 27; i++)
+				AddDocumentWithFields(writer);
+			writer.Close();
+			
+			IndexReader r = IndexReader.Open(d);
+			r.DeleteDocument(5);
+			r.Flush(commitUserData);
+			r.Close();
+			
+			SegmentInfos sis = new SegmentInfos();
+			sis.Read(d);
+			IndexReader r2 = IndexReader.Open(d);
+			IndexCommit c = r.GetIndexCommit();
+			Assert.AreEqual(c.GetUserData(), commitUserData);
+			
+			Assert.AreEqual(sis.GetCurrentSegmentFileName(), c.GetSegmentsFileName());
+			
+			Assert.IsTrue(c.Equals(r.GetIndexCommit()));
+			
+			// Change the index
+			writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
+			writer.SetMaxBufferedDocs(2);
+			for (int i = 0; i < 7; i++)
+				AddDocumentWithFields(writer);
+			writer.Close();
+			
+			IndexReader r3 = r2.Reopen();
+			Assert.IsFalse(c.Equals(r3.GetIndexCommit()));
+			Assert.IsFalse(r2.GetIndexCommit().IsOptimized());
+			r3.Close();
+			
+			writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
+			writer.Optimize();
+			writer.Close();
+			
+			r3 = r2.Reopen();
+			Assert.IsTrue(r3.GetIndexCommit().IsOptimized());
+			r2.Close();
+			r3.Close();
+			d.Close();
+		}
+		
+		[Test]
+		public virtual void  TestIsCurrent()
+		{
+			RAMDirectory d = new MockRAMDirectory();
+			IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			AddDocumentWithFields(writer);
+			writer.Close();
+			// set up reader:
+			IndexReader reader = IndexReader.Open(d);
+			Assert.IsTrue(reader.IsCurrent());
+			// modify index by adding another document:
+			writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
+			AddDocumentWithFields(writer);
+			writer.Close();
+			Assert.IsFalse(reader.IsCurrent());
+			// re-create index:
+			writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			AddDocumentWithFields(writer);
+			writer.Close();
+			Assert.IsFalse(reader.IsCurrent());
+			reader.Close();
+			d.Close();
+		}
+		
+		/// <summary> Tests the IndexReader.getFieldNames implementation</summary>
+		/// <throws>  Exception on error </throws>
+		[Test]
+		public virtual void  TestGetFieldNames()
+		{
+			RAMDirectory d = new MockRAMDirectory();
+			// set up writer
+			IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			AddDocumentWithFields(writer);
+			writer.Close();
+			// set up reader
+			IndexReader reader = IndexReader.Open(d);
+			System.Collections.ICollection fieldNames = reader.GetFieldNames(IndexReader.FieldOption.ALL);
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "keyword"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "text"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unindexed"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unstored"));
+			reader.Close();
+			// add more documents
+			writer = new IndexWriter(d, new StandardAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
+			// want to get some more segments here
+			for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
+			{
+				AddDocumentWithFields(writer);
+			}
+			// new fields are in some different segments (we hope)
+			for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
+			{
+				AddDocumentWithDifferentFields(writer);
+			}
+			// new termvector fields
+			for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
+			{
+				AddDocumentWithTermVectorFields(writer);
+			}
+			
+			writer.Close();
+			// verify fields again
+			reader = IndexReader.Open(d);
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.ALL);
+			Assert.AreEqual(13, fieldNames.Count); // the following fields
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "keyword"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "text"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unindexed"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unstored"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "keyword2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "text2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unindexed2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unstored2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvnot"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "termvector"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvposition"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvoffset"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvpositionoffset"));
+			
+			// verify that only indexed fields were returned
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.INDEXED);
+			Assert.AreEqual(11, fieldNames.Count); // 6 original + the 5 termvector fields 
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "keyword"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "text"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unstored"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "keyword2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "text2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unstored2"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvnot"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "termvector"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvposition"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvoffset"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvpositionoffset"));
+			
+			// verify that only unindexed fields were returned
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.UNINDEXED);
+			Assert.AreEqual(2, fieldNames.Count); // the following fields
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unindexed"));
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "unindexed2"));
+			
+			// verify index term vector fields  
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR);
+			Assert.AreEqual(1, fieldNames.Count); // 1 field has term vector only
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "termvector"));
+			
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_POSITION);
+			Assert.AreEqual(1, fieldNames.Count); // 4 fields are indexed with term vectors
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvposition"));
+			
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_OFFSET);
+			Assert.AreEqual(1, fieldNames.Count); // 4 fields are indexed with term vectors
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvoffset"));
+			
+			fieldNames = reader.GetFieldNames(IndexReader.FieldOption.TERMVECTOR_WITH_POSITION_OFFSET);
+			Assert.AreEqual(1, fieldNames.Count); // 4 fields are indexed with term vectors
+			Assert.IsTrue(SupportClass.CollectionsHelper.Contains(fieldNames, "tvpositionoffset"));
+			reader.Close();
+			d.Close();
+		}
+		
+		[Test]
+		public virtual void  TestTermVectors()
+		{
+			RAMDirectory d = new MockRAMDirectory();
+			// set up writer
+			IndexWriter writer = new IndexWriter(d, new StandardAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			// want to get some more segments here
+			// new termvector fields
+			for (int i = 0; i < 5 * writer.GetMergeFactor(); i++)
+			{
+				Document doc = new Document();
+				doc.Add(new Field("tvnot", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.NO));
+				doc.Add(new Field("termvector", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.YES));
+				doc.Add(new Field("tvoffset", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_OFFSETS));
+				doc.Add(new Field("tvposition", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS));
+				doc.Add(new Field("tvpositionoffset", "one two two three three three", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS));
+				
+				writer.AddDocument(doc);
+			}
+			writer.Close();
+			IndexReader reader = IndexReader.Open(d);
+			FieldSortedTermVectorMapper mapper = new FieldSortedTermVectorMapper(new TermVectorEntryFreqSortedComparator());
+			reader.GetTermFreqVector(0, mapper);
+			System.Collections.IDictionary map = mapper.GetFieldToTerms();
+			Assert.IsTrue(map != null, "map is null and it shouldn't be");
+			Assert.IsTrue(map.Count == 4, "map Size: " + map.Count + " is not: " + 4);
+			System.Collections.Hashtable set_Renamed = (System.Collections.Hashtable) map["termvector"];
+			for (System.Collections.IEnumerator iterator = set_Renamed.GetEnumerator(); iterator.MoveNext(); )
+			{
+				TermVectorEntry entry = (TermVectorEntry) iterator.Current;
+				Assert.IsTrue(entry != null, "entry is null and it shouldn't be");
+				System.Console.Out.WriteLine("Entry: " + entry);
+			}
+		}
+		
+		private void  AssertTermDocsCount(System.String msg, IndexReader reader, Term term, int expected)
+		{
+			TermDocs tdocs = null;
+			
+			try
+			{
+				tdocs = reader.TermDocs(term);
+				Assert.IsNotNull(tdocs, msg + ", null TermDocs");
+				int count = 0;
+				while (tdocs.Next())
+				{
+					count++;
+				}
+				Assert.AreEqual(expected, count, msg + ", count mismatch");
+			}
+			finally
+			{
+				if (tdocs != null)
+					tdocs.Close();
+			}
+		}
+		
+		
+		
+		[Test]
+		public virtual void  TestBasicDelete()
+		{
+			Directory dir = new MockRAMDirectory();
+			
+			IndexWriter writer = null;
+			IndexReader reader = null;
+			Term searchTerm = new Term("content", "aaa");
+			
+			//  add 100 documents with term : aaa
+			writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			for (int i = 0; i < 100; i++)
+			{
+				AddDoc(writer, searchTerm.Text());
+			}
+			writer.Close();
+			
+			// OPEN READER AT THIS POINT - this should fix the view of the
+			// index at the point of having 100 "aaa" documents and 0 "bbb"
+			reader = IndexReader.Open(dir);
+			Assert.AreEqual(100, reader.DocFreq(searchTerm), "first docFreq");
+			AssertTermDocsCount("first reader", reader, searchTerm, 100);
+			reader.Close();
+			
+			// DELETE DOCUMENTS CONTAINING TERM: aaa
+			int deleted = 0;
+			reader = IndexReader.Open(dir);
+			deleted = reader.DeleteDocuments(searchTerm);
+			Assert.AreEqual(100, deleted, "deleted count");
+			Assert.AreEqual(100, reader.DocFreq(searchTerm), "deleted docFreq");
+			AssertTermDocsCount("deleted termDocs", reader, searchTerm, 0);
+			
+			// open a 2nd reader to make sure first reader can
+			// commit its changes (.del) while second reader
+			// is open:
+			IndexReader reader2 = IndexReader.Open(dir);
+			reader.Close();
+			
+			// CREATE A NEW READER and re-test
+			reader = IndexReader.Open(dir);
+			Assert.AreEqual(100, reader.DocFreq(searchTerm), "deleted docFreq");
+			AssertTermDocsCount("deleted termDocs", reader, searchTerm, 0);
+			reader.Close();
+			reader2.Close();
+			dir.Close();
+		}
+		
+		[Test]
+		public virtual void  TestBinaryFields()
+		{
+			Directory dir = new RAMDirectory();
+			byte[] bin = new byte[]{0, 1, 2, 3, 4, 5, 6, 7, 8, 9};
+			
+			IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.UNLIMITED);
+			
+			for (int i = 0; i < 10; i++)
+			{
+				AddDoc(writer, "document number " + (i + 1));
+				AddDocumentWithFields(writer);
+				AddDocumentWithDifferentFields(writer);
+				AddDocumentWithTermVectorFields(writer);
+			}
+			writer.Close();
+			writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
+			Document doc = new Document();
+			doc.Add(new Field("bin1", bin, Field.Store.YES));
+			doc.Add(new Field("bin2", bin, Field.Store.COMPRESS));
+			doc.Add(new Field("junk", "junk text", Field.Store.NO, Field.Index.ANALYZED));
+			writer.AddDocument(doc);
+			writer.Close();
+			IndexReader reader = IndexReader.Open(dir);
+			doc = reader.Document(reader.MaxDoc() - 1);
+			Field[] fields = doc.GetFields("bin1");
+			Assert.IsNotNull(fields);
+			Assert.AreEqual(1, fields.Length);
+			Field b1 = fields[0];
+			Assert.IsTrue(b1.IsBinary());
+			byte[] data1 = b1.GetBinaryValue();
+			Assert.AreEqual(bin.Length, b1.GetBinaryLength());
+			for (int i = 0; i < bin.Length; i++)
+			{
+				Assert.AreEqual(bin[i], data1[i + b1.GetBinaryOffset()]);
+			}
+			fields = doc.GetFields("bin2");
+			Assert.IsNotNull(fields);
+			Assert.AreEqual(1, fields.Length);
+			b1 = fields[0];
+			Assert.IsTrue(b1.IsBinary());
+			data1 = b1.GetBinaryValue();
+			Assert.AreEqual(bin.Length, b1.GetBinaryLength());
+			for (int i = 0; i < bin.Length; i++)
+			{
+				Assert.AreEqual(bin[i], data1[i + b1.GetBinaryOffset()]);
+			}
+			System.Collections.Hashtable lazyFields = new System.Collections.Hashtable();
+			SupportClass.CollectionsHelper.AddIfNotContains(lazyFields, "bin1");
+			FieldSelector sel = new SetBasedFieldSelector(new System.Collections.Hashtable(), lazyFields);
+			doc = reader.Document(reader.MaxDoc() - 1, sel);
+			Fieldable[] fieldables = doc.GetFieldables("bin1");
+			Assert.IsNotNull(fieldables);
+			Assert.AreEqual(1, fieldables.Length);
+			Fieldable fb1 = fieldables[0];
+			Assert.IsTrue(fb1.IsBinary());
+			Assert.AreEqual(bin.Length, fb1.GetBinaryLength());
+			data1 = fb1.GetBinaryValue();
+			Assert.AreEqual(bin.Length, fb1.GetBinaryLength());
+			for (int i = 0; i < bin.Length; i++)
+			{
+				Assert.AreEqual(bin[i], data1[i + fb1.GetBinaryOffset()]);
+			}
+			reader.Close();
+			// force optimize
+			
+			
+			writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false, IndexWriter.MaxFieldLength.LIMITED);
+			writer.Optimize();
+			writer.Close();
+			reader = IndexReader.Open(dir);
+			doc = reader.Document(reader.MaxDoc() - 1);
+			fields = doc.GetFields("bin1");
+			Assert.IsNotNull(fields);
+			Assert.AreEqual(1, fields.Length);
+			b1 = fields[0];
+			Assert.IsTrue(b1.IsBinary());
+			data1 = b1.GetBinaryValue();
+			Assert.AreEqual(bin.Length, b1.GetBinaryLength());
+			for (int i = 0; i < bin.Length; i++)
+			{
+				Assert.AreEqual(bin[i], data1[i + b1.GetBinaryOffset()]);
+			}
+			fields = doc.GetFields("bin2");
+			Assert.IsNotNull(fields);
+			Assert.AreEqual(1, fields.Length);
+			b1 = fields[0];
+			Assert.IsTrue(b1.IsBinary());
+			data1 = b1.GetBinaryValue();
+			Assert.AreEqual(bin.Length, b1.GetBinaryLength());
+			for (int i = 0; i < bin.Length; i++)
+			{
+				Assert.AreEqual(bin[i], data1[i + b1.GetBinaryOffset()]);
+			}
+			reader.Close();
+		}
+		
+		// Make sure attempts to make changes after reader is
+		// closed throws IOException:
+		[Test]
+		public virtual void  TestChangesAfterClose()
+		{
+			Directory dir = new RAMDirectory();
+			
+			IndexWriter writer = null;
+			IndexReader reader = null;
+			Term searchTerm = new Term("content", "aaa");
+			
+			//  add 11 documents with term : aaa
+			writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			for (int i = 0; i < 11; i++)
+			{
+				AddDoc(writer, searchTerm.Text());
+			}
+			writer.Close();
+			
+			reader = IndexReader.Open(dir);
+			
+			// Close reader:
+			reader.Close();
+			
+			// Then, try to make changes:
+			try
+			{
+				reader.DeleteDocument(4);
+				Assert.Fail("deleteDocument after close failed to throw IOException");
+			}
+			catch (AlreadyClosedException e)
+			{
+				// expected
+			}
+			
+			try
+			{
+				reader.SetNorm(5, "aaa", 2.0f);
+				Assert.Fail("setNorm after close failed to throw IOException");
+			}
+			catch (AlreadyClosedException e)
+			{
+				// expected
+			}
+			
+			try
+			{
+				reader.UndeleteAll();
+				Assert.Fail("undeleteAll after close failed to throw IOException");
+			}
+			catch (AlreadyClosedException e)
+			{
+				// expected
+			}
+		}
+		
+		// Make sure we get lock obtain failed exception with 2 writers:
+		[Test]
+		public virtual void  TestLockObtainFailed()
+		{
+			Directory dir = new RAMDirectory();
+			
+			IndexWriter writer = null;
+			IndexReader reader = null;
+			Term searchTerm = new Term("content", "aaa");
+			
+			//  add 11 documents with term : aaa
+			writer = new IndexWriter(dir, new WhitespaceAnalyzer(), true, IndexWriter.MaxFieldLength.LIMITED);
+			for (int i = 0; i < 11; i++)
+			{
+				AddDoc(writer, searchTerm.Text());
+			}
+			
+			// Create reader:
+			reader = IndexReader.Open(dir);
+			
+			// Try to make changes
+			try
+			{
+				reader.DeleteDocument(4);
+				Assert.Fail("deleteDocument should have hit LockObtainFailedException");
+			}
+			catch (LockObtainFailedException e)
+			{
+				// expected
+			}
+			
+			try
+			{
+				reader.SetNorm(5, "aaa", 2.0f);
+				Assert.Fail("setNorm should have hit LockObtainFailedException");
+			}
+			catch (LockObtainFailedException e)
+			{
+				// expected
+			}
+			
+			try
+			{
+				reader.UndeleteAll();
+				Assert.Fail("undeleteAll should have hit LockObtainFailedException");
+			}
+			catch (LockObtainFailedException e)
+			{
+				// expected
+			}
+			writer.Close();
+			reader.Close();
+		}
+		
+		// Make sure you can set norms & commit even if a reader
+		// is open against the index:
+		[Test]
+		public virtual void  TestWritingNorms()
+		{
+			System.String tempDir = SupportClass.AppSettings.Get("tempDir", "");
+			if (tempDir == null)
+				throw new System.IO.IOException("tempDir undefined, cannot run test");
+			
+			System.IO.FileInfo indexDir = new System.IO.FileInfo(System.IO.Path.Combine(tempDir, "lucenetestnormwriter"));
+			Directory dir = FSDirectory.Open(indexDir);
+			IndexWriter writer;
+			IndexReader reader;
+			Term searchTerm = new Term("content", "aaa");
+			
+			//  add 1 documents with term : aaa

[... 1529 lines stripped ...]


Mime
View raw message