Return-Path: X-Original-To: apmail-lucene-commits-archive@www.apache.org Delivered-To: apmail-lucene-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 3E2239FEC for ; Mon, 14 Nov 2011 22:37:55 +0000 (UTC) Received: (qmail 57028 invoked by uid 500); 14 Nov 2011 22:37:55 -0000 Mailing-List: contact commits-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: dev@lucene.apache.org Delivered-To: mailing list commits@lucene.apache.org Received: (qmail 57021 invoked by uid 99); 14 Nov 2011 22:37:55 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 14 Nov 2011 22:37:55 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=5.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 14 Nov 2011 22:37:39 +0000 Received: from eris.apache.org (localhost [127.0.0.1]) by eris.apache.org (Postfix) with ESMTP id 2400F2388B1B; Mon, 14 Nov 2011 22:36:52 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit Subject: svn commit: r1201946 [8/14] - in /lucene/dev/branches/solrcloud: ./ dev-tools/idea/lucene/contrib/ lucene/ lucene/contrib/ lucene/contrib/demo/src/java/org/apache/lucene/demo/ lucene/contrib/highlighter/src/java/org/apache/lucene/search/highlight/ luce... Date: Mon, 14 Nov 2011 22:36:32 -0000 To: commits@lucene.apache.org From: yonik@apache.org X-Mailer: svnmailer-1.0.8-patched Message-Id: <20111114223652.2400F2388B1B@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestCodecs.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestCodecs.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestCodecs.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestCodecs.java Mon Nov 14 22:36:20 2011 @@ -33,7 +33,6 @@ import org.apache.lucene.index.codecs.Po import org.apache.lucene.index.codecs.TermStats; import org.apache.lucene.index.codecs.TermsConsumer; import org.apache.lucene.index.codecs.lucene3x.Lucene3xCodec; -import org.apache.lucene.index.codecs.lucene3x.Lucene3xPostingsFormat; import org.apache.lucene.index.codecs.mocksep.MockSepPostingsFormat; import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.IndexSearcher; @@ -42,6 +41,7 @@ import org.apache.lucene.search.Query; import org.apache.lucene.search.ScoreDoc; import org.apache.lucene.store.Directory; import org.apache.lucene.util.BytesRef; +import org.apache.lucene.util.InfoStream; import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util.OpenBitSet; import org.apache.lucene.util.Version; @@ -370,7 +370,7 @@ public class TestCodecs extends LuceneTe assertEquals(2, results.length); assertEquals(0, results[0].doc); - writer.optimize(); + writer.forceMerge(1); // optimise to merge the segments. results = this.search(writer, pq, 5); @@ -609,7 +609,7 @@ public class TestCodecs extends LuceneTe final int termIndexInterval = _TestUtil.nextInt(random, 13, 27); final Codec codec = Codec.getDefault(); - final SegmentWriteState state = new SegmentWriteState(null, dir, SEGMENT, fieldInfos, 10000, termIndexInterval, codec, null, newIOContext(random)); + final SegmentWriteState state = new SegmentWriteState(InfoStream.getDefault(), dir, SEGMENT, fieldInfos, 10000, termIndexInterval, codec, null, newIOContext(random)); final FieldsConsumer consumer = codec.postingsFormat().fieldsConsumer(state); Arrays.sort(fields); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConcurrentMergeScheduler.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConcurrentMergeScheduler.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConcurrentMergeScheduler.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConcurrentMergeScheduler.java Mon Nov 14 22:36:20 2011 @@ -74,7 +74,6 @@ public class TestConcurrentMergeSchedule directory.failOn(failure); IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2)); - writer.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); Field idField = newField("id", "", StringField.TYPE_STORED); doc.add(idField); @@ -133,7 +132,6 @@ public class TestConcurrentMergeSchedule IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setMergePolicy(mp)); - writer.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); Field idField = newField("id", "", StringField.TYPE_STORED); @@ -172,7 +170,6 @@ public class TestConcurrentMergeSchedule IndexWriter writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setMaxBufferedDocs(2)); - writer.setInfoStream(VERBOSE ? System.out : null); for(int iter=0;iter<7;iter++) { if (VERBOSE) { @@ -192,7 +189,6 @@ public class TestConcurrentMergeSchedule writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setOpenMode(OpenMode.APPEND).setMaxBufferedDocs(2)); - writer.setInfoStream(VERBOSE ? System.out : null); } writer.close(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestConsistentFieldNumbers.java Mon Nov 14 22:36:20 2011 @@ -27,6 +27,7 @@ import org.apache.lucene.document.FieldT import org.apache.lucene.document.StringField; import org.apache.lucene.document.TextField; import org.apache.lucene.store.Directory; +import org.apache.lucene.util.FailOnNonBulkMergesInfoStream; import org.apache.lucene.util.LuceneTestCase; import org.junit.Test; @@ -76,7 +77,7 @@ public class TestConsistentFieldNumbers assertEquals("f4", fis2.fieldInfo(3).name); writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - writer.optimize(); + writer.forceMerge(1); writer.close(); sis = new SegmentInfos(); @@ -140,7 +141,7 @@ public class TestConsistentFieldNumbers assertEquals("f4", fis2.fieldInfo(3).name); writer = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - writer.optimize(); + writer.forceMerge(1); writer.close(); sis = new SegmentInfos(); @@ -250,9 +251,8 @@ public class TestConsistentFieldNumbers IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy( - new LogByteSizeMergePolicy())); - writer.optimize(); - assertFalse(" field numbers got mixed up", writer.anyNonBulkMerges); + new LogByteSizeMergePolicy()).setInfoStream(new FailOnNonBulkMergesInfoStream())); + writer.forceMerge(1); writer.close(); SegmentInfos sis = new SegmentInfos(); @@ -293,7 +293,7 @@ public class TestConsistentFieldNumbers writer.addDocument(d); } - writer.optimize(); + writer.forceMerge(1); writer.close(); SegmentInfos sis = new SegmentInfos(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDeletionPolicy.java Mon Nov 14 22:36:20 2011 @@ -74,7 +74,7 @@ public class TestDeletionPolicy extends public void onCommit(List commits) throws IOException { IndexCommit lastCommit = commits.get(commits.size()-1); IndexReader r = IndexReader.open(dir, true); - assertEquals("lastCommit.isOptimized()=" + lastCommit.isOptimized() + " vs IndexReader.isOptimized=" + r.isOptimized(), r.isOptimized(), lastCommit.isOptimized()); + assertEquals("lastCommit.segmentCount()=" + lastCommit.getSegmentCount() + " vs IndexReader.segmentCount=" + r.getSequentialSubReaders().length, r.getSequentialSubReaders().length, lastCommit.getSegmentCount()); r.close(); verifyCommitOrder(commits); numOnCommit++; @@ -317,13 +317,13 @@ public class TestDeletionPolicy extends } writer.close(); - final boolean isOptimized; + final boolean needsMerging; { IndexReader r = IndexReader.open(dir); - isOptimized = r.isOptimized(); + needsMerging = r.getSequentialSubReaders().length != 1; r.close(); } - if (!isOptimized) { + if (needsMerging) { conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode( OpenMode.APPEND).setIndexDeletionPolicy(policy); @@ -332,23 +332,22 @@ public class TestDeletionPolicy extends ((LogMergePolicy) mp).setUseCompoundFile(useCompoundFile); } if (VERBOSE) { - System.out.println("TEST: open writer for optimize"); + System.out.println("TEST: open writer for forceMerge"); } writer = new IndexWriter(dir, conf); - writer.setInfoStream(VERBOSE ? System.out : null); - writer.optimize(); + writer.forceMerge(1); writer.close(); } - assertEquals(isOptimized ? 0:1, policy.numOnInit); + assertEquals(needsMerging ? 1:0, policy.numOnInit); // If we are not auto committing then there should // be exactly 2 commits (one per close above): - assertEquals(1 + (isOptimized ? 0:1), policy.numOnCommit); + assertEquals(1 + (needsMerging ? 1:0), policy.numOnCommit); // Test listCommits Collection commits = IndexReader.listCommits(dir); // 2 from closing writer - assertEquals(1 + (isOptimized ? 0:1), commits.size()); + assertEquals(1 + (needsMerging ? 1:0), commits.size()); // Make sure we can open a reader on each commit: for (final IndexCommit commit : commits) { @@ -419,16 +418,16 @@ public class TestDeletionPolicy extends } assertTrue(lastCommit != null); - // Now add 1 doc and optimize + // Now add 1 doc and merge writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setIndexDeletionPolicy(policy)); addDoc(writer); assertEquals(11, writer.numDocs()); - writer.optimize(); + writer.forceMerge(1); writer.close(); assertEquals(6, IndexReader.listCommits(dir).size()); - // Now open writer on the commit just before optimize: + // Now open writer on the commit just before merge: writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setIndexDeletionPolicy(policy).setIndexCommit(lastCommit)); assertEquals(10, writer.numDocs()); @@ -437,8 +436,8 @@ public class TestDeletionPolicy extends writer.rollback(); IndexReader r = IndexReader.open(dir, true); - // Still optimized, still 11 docs - assertTrue(r.isOptimized()); + // Still merged, still 11 docs + assertEquals(1, r.getSequentialSubReaders().length); assertEquals(11, r.numDocs()); r.close(); @@ -452,39 +451,39 @@ public class TestDeletionPolicy extends assertEquals(7, IndexReader.listCommits(dir).size()); r = IndexReader.open(dir, true); - // Not optimized because we rolled it back, and now only + // Not fully merged because we rolled it back, and now only // 10 docs - assertTrue(!r.isOptimized()); + assertTrue(r.getSequentialSubReaders().length > 1); assertEquals(10, r.numDocs()); r.close(); - // Reoptimize + // Re-merge writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setIndexDeletionPolicy(policy)); - writer.optimize(); + writer.forceMerge(1); writer.close(); r = IndexReader.open(dir, true); - assertTrue(r.isOptimized()); + assertEquals(1, r.getSequentialSubReaders().length); assertEquals(10, r.numDocs()); r.close(); - // Now open writer on the commit just before optimize, + // Now open writer on the commit just before merging, // but this time keeping only the last commit: writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setIndexCommit(lastCommit)); assertEquals(10, writer.numDocs()); - // Reader still sees optimized index, because writer + // Reader still sees fully merged index, because writer // opened on the prior commit has not yet committed: r = IndexReader.open(dir, true); - assertTrue(r.isOptimized()); + assertEquals(1, r.getSequentialSubReaders().length); assertEquals(10, r.numDocs()); r.close(); writer.close(); - // Now reader sees unoptimized index: + // Now reader sees not-fully-merged index: r = IndexReader.open(dir, true); - assertTrue(!r.isOptimized()); + assertTrue(r.getSequentialSubReaders().length > 1); assertEquals(10, r.numDocs()); r.close(); @@ -526,7 +525,7 @@ public class TestDeletionPolicy extends ((LogMergePolicy) mp).setUseCompoundFile(true); } writer = new IndexWriter(dir, conf); - writer.optimize(); + writer.forceMerge(1); writer.close(); assertEquals(1, policy.numOnInit); @@ -570,7 +569,7 @@ public class TestDeletionPolicy extends for(int i=0;i<17;i++) { addDoc(writer); } - writer.optimize(); + writer.forceMerge(1); writer.close(); } @@ -645,7 +644,6 @@ public class TestDeletionPolicy extends ((LogMergePolicy) mp).setUseCompoundFile(useCompoundFile); } writer = new IndexWriter(dir, conf); - writer.setInfoStream(VERBOSE ? System.out : null); for(int j=0;j<17;j++) { addDoc(writer); } @@ -675,15 +673,15 @@ public class TestDeletionPolicy extends ((LogMergePolicy) mp).setUseCompoundFile(useCompoundFile); } IndexReader r = IndexReader.open(dir); - final boolean wasOptimized = r.isOptimized(); + final boolean wasFullyMerged = r.getSequentialSubReaders().length == 1 && !r.hasDeletions(); r.close(); writer = new IndexWriter(dir, conf); - writer.optimize(); + writer.forceMerge(1); // this is a commit writer.close(); assertEquals(2*(N+1)+1, policy.numOnInit); - assertEquals(2*(N+2) - (wasOptimized ? 1:0), policy.numOnCommit); + assertEquals(2*(N+2) - (wasFullyMerged ? 1:0), policy.numOnCommit); IndexSearcher searcher = new IndexSearcher(dir, false); ScoreDoc[] hits = searcher.search(query, null, 1000).scoreDocs; Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDirectoryReader.java Mon Nov 14 22:36:20 2011 @@ -184,7 +184,7 @@ public class TestDirectoryReader extends while (td.nextDoc() != td.NO_MORE_DOCS) ret += td.docID(); // really a dummy assert to ensure that we got some docs and to ensure that - // nothing is optimized out. + // nothing is eliminated by hotspot assertTrue(ret > 0); readers1[0].close(); readers1[1].close(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDoc.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDoc.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDoc.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDoc.java Mon Nov 14 22:36:20 2011 @@ -37,6 +37,7 @@ import org.apache.lucene.index.codecs.Co import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.store.Directory; import org.apache.lucene.store.IOContext; +import org.apache.lucene.util.InfoStream; import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util._TestUtil; @@ -198,16 +199,17 @@ public class TestDoc extends LuceneTestC SegmentReader r1 = SegmentReader.get(true, si1, IndexReader.DEFAULT_TERMS_INDEX_DIVISOR, context); SegmentReader r2 = SegmentReader.get(true, si2, IndexReader.DEFAULT_TERMS_INDEX_DIVISOR, context); - SegmentMerger merger = new SegmentMerger(si1.dir, IndexWriterConfig.DEFAULT_TERM_INDEX_INTERVAL, merged, null, null, new FieldInfos(), Codec.getDefault(), context); + final Codec codec = Codec.getDefault(); + SegmentMerger merger = new SegmentMerger(InfoStream.getDefault(), si1.dir, IndexWriterConfig.DEFAULT_TERM_INDEX_INTERVAL, merged, MergeState.CheckAbort.NONE, null, new FieldInfos(), codec, context); merger.add(r1); merger.add(r2); - merger.merge(); + MergeState mergeState = merger.merge(); r1.close(); r2.close(); - final FieldInfos fieldInfos = merger.fieldInfos(); + final FieldInfos fieldInfos = mergeState.fieldInfos; final SegmentInfo info = new SegmentInfo(merged, si1.docCount + si2.docCount, si1.dir, - false, merger.getCodec(), fieldInfos); + false, codec, fieldInfos); if (useCompoundFile) { Collection filesToDelete = merger.createCompoundFile(merged + ".cfs", info, newIOContext(random)); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocCount.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocCount.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocCount.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocCount.java Mon Nov 14 22:36:20 2011 @@ -42,7 +42,7 @@ public class TestDocCount extends Lucene IndexReader ir = iw.getReader(); verifyCount(ir); ir.close(); - iw.optimize(); + iw.forceMerge(1); ir = iw.getReader(); verifyCount(ir); ir.close(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestDocTermOrds.java Mon Nov 14 22:36:20 2011 @@ -17,7 +17,6 @@ package org.apache.lucene.index; * limitations under the License. */ -import java.io.IOException; import java.util.Arrays; import java.util.List; import java.util.ArrayList; @@ -31,20 +30,8 @@ import org.apache.lucene.document.Numeri import org.apache.lucene.document.StringField; import org.apache.lucene.document.TextField; import org.apache.lucene.index.DocTermOrds.TermOrdsIterator; -import org.apache.lucene.index.codecs.BlockTermsReader; -import org.apache.lucene.index.codecs.BlockTermsWriter; import org.apache.lucene.index.codecs.Codec; import org.apache.lucene.index.codecs.PostingsFormat; -import org.apache.lucene.index.codecs.FieldsConsumer; -import org.apache.lucene.index.codecs.FieldsProducer; -import org.apache.lucene.index.codecs.FixedGapTermsIndexReader; -import org.apache.lucene.index.codecs.FixedGapTermsIndexWriter; -import org.apache.lucene.index.codecs.PostingsReaderBase; -import org.apache.lucene.index.codecs.PostingsWriterBase; -import org.apache.lucene.index.codecs.TermsIndexReaderBase; -import org.apache.lucene.index.codecs.TermsIndexWriterBase; -import org.apache.lucene.index.codecs.lucene40.Lucene40PostingsReader; -import org.apache.lucene.index.codecs.lucene40.Lucene40PostingsWriter; import org.apache.lucene.search.FieldCache; import org.apache.lucene.store.Directory; import org.apache.lucene.store.MockDirectoryWrapper; @@ -317,7 +304,7 @@ public class TestDocTermOrds extends Luc _TestUtil.nextInt(random, 2, 10)); - final int[] docIDToID = FieldCache.DEFAULT.getInts(r, "id"); + final int[] docIDToID = FieldCache.DEFAULT.getInts(r, "id", false); /* for(int docID=0;docID fieldNames = new HashSet(); final int numFields = 2 + (TEST_NIGHTLY ? random.nextInt(200) : random.nextInt(20)); @@ -287,7 +285,7 @@ public class TestGlobalFieldNumbers exte FieldNumberBiMap globalFieldMap = writer.segmentInfos .getOrLoadGlobalFieldNumberMap(base); Set> entries = globalFieldMap.entries(); - writer.optimize(); + writer.forceMerge(1); writer.commit(); writer.close(); Set> afterOptmize = globalFieldMap.entries(); @@ -354,7 +352,7 @@ public class TestGlobalFieldNumbers exte IndexWriter w = new IndexWriter(base, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy( new LogByteSizeMergePolicy())); - w.optimize(); + w.forceMerge(1); w.close(); SegmentInfos sis = new SegmentInfos(); sis.read(base); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexCommit.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexCommit.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexCommit.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexCommit.java Mon Nov 14 22:36:20 2011 @@ -42,7 +42,7 @@ public class TestIndexCommit extends Luc @Override public long getTimestamp() throws IOException { return 1;} @Override public Map getUserData() throws IOException { return null; } @Override public boolean isDeleted() { return false; } - @Override public boolean isOptimized() { return false; } + @Override public int getSegmentCount() { return 2; } }; IndexCommit ic2 = new IndexCommit() { @@ -55,7 +55,7 @@ public class TestIndexCommit extends Luc @Override public long getTimestamp() throws IOException { return 1;} @Override public Map getUserData() throws IOException { return null; } @Override public boolean isDeleted() { return false; } - @Override public boolean isOptimized() { return false; } + @Override public int getSegmentCount() { return 2; } }; assertEquals(ic1, ic2); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexFileDeleter.java Mon Nov 14 22:36:20 2011 @@ -54,8 +54,6 @@ public class TestIndexFileDeleter extend setMergePolicy(mergePolicy) ); - writer.setInfoStream(VERBOSE ? System.out : null); - int i; for(i=0;i<35;i++) { addDoc(writer, i); @@ -149,7 +147,9 @@ public class TestIndexFileDeleter extend copyFile(dir, "segments_2", "segments_1"); // Create a bogus cfs file shadowing a non-cfs segment: - assertTrue(dir.fileExists("_3.fdt")); + + // TODO: assert is bogus (relies upon codec-specific filenames) + assertTrue(dir.fileExists("_3.fdt") || dir.fileExists("_3.fld")); assertTrue(!dir.fileExists("_3.cfs")); copyFile(dir, "_1.cfs", "_3.cfs"); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReader.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReader.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReader.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReader.java Mon Nov 14 22:36:20 2011 @@ -88,7 +88,6 @@ public class TestIndexReader extends Luc writer = new IndexWriter(d, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode( OpenMode.APPEND).setMaxBufferedDocs(2)); - writer.setInfoStream(VERBOSE ? System.out : null); for(int i=0;i<7;i++) addDocumentWithFields(writer); writer.close(); @@ -96,18 +95,18 @@ public class TestIndexReader extends Luc IndexReader r3 = IndexReader.openIfChanged(r2); assertNotNull(r3); assertFalse(c.equals(r3.getIndexCommit())); - assertFalse(r2.getIndexCommit().isOptimized()); + assertFalse(r2.getIndexCommit().getSegmentCount() == 1 && !r2.hasDeletions()); r3.close(); writer = new IndexWriter(d, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setOpenMode(OpenMode.APPEND)); - writer.optimize(); + writer.forceMerge(1); writer.close(); r3 = IndexReader.openIfChanged(r2); assertNotNull(r3); - assertTrue(r3.getIndexCommit().isOptimized()); + assertEquals(1, r3.getIndexCommit().getSegmentCount()); r2.close(); r3.close(); d.close(); @@ -382,11 +381,11 @@ public class TestIndexReader extends Luc assertEquals(bin[i], bytesRef.bytes[i + bytesRef.offset]); } reader.close(); - // force optimize + // force merge writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND).setMergePolicy(newLogMergePolicy())); - writer.optimize(); + writer.forceMerge(1); writer.close(); reader = IndexReader.open(dir, false); doc2 = reader.document(reader.maxDoc() - 1); @@ -722,7 +721,7 @@ public class TestIndexReader extends Luc // [incorrectly] hit a "docs out of order" // IllegalStateException because above out-of-bounds // deleteDocument corrupted the index: - writer.optimize(); + writer.forceMerge(1); writer.close(); if (!gotException) { fail("delete of out-of-bounds doc number failed to hit exception"); @@ -847,7 +846,9 @@ public class TestIndexReader extends Luc assertEquals("IndexReaders have different values for numDocs.", index1.numDocs(), index2.numDocs()); assertEquals("IndexReaders have different values for maxDoc.", index1.maxDoc(), index2.maxDoc()); assertEquals("Only one IndexReader has deletions.", index1.hasDeletions(), index2.hasDeletions()); - assertEquals("Only one index is optimized.", index1.isOptimized(), index2.isOptimized()); + if (!(index1 instanceof ParallelReader)) { + assertEquals("Single segment test differs.", index1.getSequentialSubReaders().length == 1, index2.getSequentialSubReaders().length == 1); + } // check field names Collection fields1 = index1.getFieldNames(FieldOption.ALL); @@ -971,19 +972,19 @@ public class TestIndexReader extends Luc IndexReader r2 = IndexReader.openIfChanged(r); assertNotNull(r2); assertFalse(c.equals(r2.getIndexCommit())); - assertFalse(r2.getIndexCommit().isOptimized()); + assertFalse(r2.getIndexCommit().getSegmentCount() == 1); r2.close(); writer = new IndexWriter(d, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setOpenMode(OpenMode.APPEND)); - writer.optimize(); + writer.forceMerge(1); writer.close(); r2 = IndexReader.openIfChanged(r); assertNotNull(r2); assertNull(IndexReader.openIfChanged(r2)); - assertTrue(r2.getIndexCommit().isOptimized()); + assertEquals(1, r2.getIndexCommit().getSegmentCount()); r.close(); r2.close(); @@ -1033,7 +1034,7 @@ public class TestIndexReader extends Luc writer = new IndexWriter(d, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setOpenMode(OpenMode.APPEND)); - writer.optimize(); + writer.forceMerge(1); writer.close(); // Make sure reopen to a single segment is still readonly: @@ -1142,7 +1143,7 @@ public class TestIndexReader extends Luc // Open reader IndexReader r = getOnlySegmentReader(IndexReader.open(dir, false)); - final int[] ints = FieldCache.DEFAULT.getInts(r, "number"); + final int[] ints = FieldCache.DEFAULT.getInts(r, "number", false); assertEquals(1, ints.length); assertEquals(17, ints[0]); @@ -1150,7 +1151,7 @@ public class TestIndexReader extends Luc IndexReader r2 = (IndexReader) r.clone(); r.close(); assertTrue(r2 != r); - final int[] ints2 = FieldCache.DEFAULT.getInts(r2, "number"); + final int[] ints2 = FieldCache.DEFAULT.getInts(r2, "number", false); r2.close(); assertEquals(1, ints2.length); @@ -1178,7 +1179,7 @@ public class TestIndexReader extends Luc // Open reader1 IndexReader r = IndexReader.open(dir, false); IndexReader r1 = getOnlySegmentReader(r); - final int[] ints = FieldCache.DEFAULT.getInts(r1, "number"); + final int[] ints = FieldCache.DEFAULT.getInts(r1, "number", false); assertEquals(1, ints.length); assertEquals(17, ints[0]); @@ -1191,7 +1192,7 @@ public class TestIndexReader extends Luc assertNotNull(r2); r.close(); IndexReader sub0 = r2.getSequentialSubReaders()[0]; - final int[] ints2 = FieldCache.DEFAULT.getInts(sub0, "number"); + final int[] ints2 = FieldCache.DEFAULT.getInts(sub0, "number", false); r2.close(); assertTrue(ints == ints2); @@ -1363,7 +1364,6 @@ public class TestIndexReader extends Luc Directory dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergePolicy(newLogMergePolicy())); ((LogMergePolicy) writer.getConfig().getMergePolicy()).setMergeFactor(3); - writer.setInfoStream(VERBOSE ? System.out : null); writer.addDocument(new Document()); writer.commit(); writer.addDocument(new Document()); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderClone.java Mon Nov 14 22:36:20 2011 @@ -192,15 +192,15 @@ public class TestIndexReaderClone extend } // open non-readOnly reader1 on multi-segment index, then - // optimize the index, then clone to readOnly reader2 - public void testReadOnlyCloneAfterOptimize() throws Exception { + // fully merge the index, then clone to readOnly reader2 + public void testReadOnlyCloneAfterFullMerge() throws Exception { final Directory dir1 = newDirectory(); TestIndexReaderReopen.createIndex(random, dir1, true); IndexReader reader1 = IndexReader.open(dir1, false); IndexWriter w = new IndexWriter(dir1, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - w.optimize(); + w.forceMerge(1); w.close(); IndexReader reader2 = reader1.clone(true); assertTrue(isReadOnly(reader2)); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderCloneNorms.java Mon Nov 14 22:36:20 2011 @@ -80,7 +80,7 @@ public class TestIndexReaderCloneNorms e /** * Test that norms values are preserved as the index is maintained. Including * separate norms. Including merging indexes with seprate norms. Including - * optimize. + * full merge. */ public void testNorms() throws IOException { // test with a single index: index1 @@ -112,7 +112,7 @@ public class TestIndexReaderCloneNorms e createIndex(random, dir3); if (VERBOSE) { - System.out.println("TEST: now addIndexes/optimize"); + System.out.println("TEST: now addIndexes/full merge"); } IndexWriter iw = new IndexWriter( dir3, @@ -121,9 +121,8 @@ public class TestIndexReaderCloneNorms e setMaxBufferedDocs(5). setMergePolicy(newLogMergePolicy(3)) ); - iw.setInfoStream(VERBOSE ? System.out : null); iw.addIndexes(dir1, dir2); - iw.optimize(); + iw.forceMerge(1); iw.close(); norms1.addAll(norms); @@ -136,7 +135,7 @@ public class TestIndexReaderCloneNorms e verifyIndex(dir3); doTestNorms(random, dir3); - // now with optimize + // now with full merge iw = new IndexWriter( dir3, newIndexWriterConfig(TEST_VERSION_CURRENT, anlzr). @@ -144,7 +143,7 @@ public class TestIndexReaderCloneNorms e setMaxBufferedDocs(5). setMergePolicy(newLogMergePolicy(3)) ); - iw.optimize(); + iw.forceMerge(1); iw.close(); verifyIndex(dir3); @@ -317,7 +316,6 @@ public class TestIndexReaderCloneNorms e lmp.setMergeFactor(3); lmp.setUseCompoundFile(compound); IndexWriter iw = new IndexWriter(dir, conf); - iw.setInfoStream(VERBOSE ? System.out : null); for (int i = 0; i < ndocs; i++) { iw.addDocument(newDoc()); } Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderDelete.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderDelete.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderDelete.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderDelete.java Mon Nov 14 22:36:20 2011 @@ -33,7 +33,7 @@ import static org.apache.lucene.index.Te import static org.apache.lucene.index.TestIndexReader.createDocument; public class TestIndexReaderDelete extends LuceneTestCase { - private void deleteReaderReaderConflict(boolean optimize) throws IOException { + private void deleteReaderReaderConflict(boolean doFullMerge) throws IOException { Directory dir = newDirectory(); Term searchTerm1 = new Term("content", "aaa"); @@ -49,8 +49,9 @@ public class TestIndexReaderDelete exten addDoc(writer, searchTerm2.text()); addDoc(writer, searchTerm3.text()); } - if(optimize) - writer.optimize(); + if (doFullMerge) { + writer.forceMerge(1); + } writer.close(); // OPEN TWO READERS @@ -131,7 +132,7 @@ public class TestIndexReaderDelete exten dir.close(); } - private void deleteReaderWriterConflict(boolean optimize) throws IOException { + private void deleteReaderWriterConflict(boolean doFullMerge) throws IOException { //Directory dir = new RAMDirectory(); Directory dir = newDirectory(); @@ -159,13 +160,14 @@ public class TestIndexReaderDelete exten addDoc(writer, searchTerm2.text()); } - // REQUEST OPTIMIZATION + // REQUEST full merge // This causes a new segment to become current for all subsequent // searchers. Because of this, deletions made via a previously open // reader, which would be applied to that reader's segment, are lost // for subsequent searchers/readers - if(optimize) - writer.optimize(); + if (doFullMerge) { + writer.forceMerge(1); + } writer.close(); // The reader should not see the new data @@ -220,7 +222,6 @@ public class TestIndexReaderDelete exten // add 100 documents with term : aaa writer = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))); - writer.setInfoStream(VERBOSE ? System.out : null); for (int i = 0; i < 100; i++) { addDoc(writer, searchTerm.text()); } @@ -256,19 +257,19 @@ public class TestIndexReaderDelete exten dir.close(); } - public void testDeleteReaderReaderConflictUnoptimized() throws IOException { + public void testDeleteReaderReaderConflictNoFullMerge() throws IOException { deleteReaderReaderConflict(false); } - public void testDeleteReaderReaderConflictOptimized() throws IOException { + public void testDeleteReaderReaderConflictFullMerge() throws IOException { deleteReaderReaderConflict(true); } - public void testDeleteReaderWriterConflictUnoptimized() throws IOException { + public void testDeleteReaderWriterConflictNoFullMerge() throws IOException { deleteReaderWriterConflict(false); } - public void testDeleteReaderWriterConflictOptimized() throws IOException { + public void testDeleteReaderWriterConflictFullMerge() throws IOException { deleteReaderWriterConflict(true); } Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderOnDiskFull.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderOnDiskFull.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderOnDiskFull.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderOnDiskFull.java Mon Nov 14 22:36:20 2011 @@ -49,7 +49,6 @@ public class TestIndexReaderOnDiskFull e IndexWriter writer = new IndexWriter(startDir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))); if (VERBOSE) { System.out.println("TEST: create initial index"); - writer.setInfoStream(System.out); } for(int i=0;i<157;i++) { Document d = new Document(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderReopen.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderReopen.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderReopen.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexReaderReopen.java Mon Nov 14 22:36:20 2011 @@ -711,7 +711,7 @@ public class TestIndexReaderReopen exten for (int i = 0; i < n; i++) { writer.addDocument(createDocument(i, 3)); } - writer.optimize(); + writer.forceMerge(1); writer.close(); final TestReopen test = new TestReopen() { @@ -961,7 +961,7 @@ public class TestIndexReaderReopen exten } if (!multiSegment) { - w.optimize(); + w.forceMerge(1); } w.close(); @@ -1003,7 +1003,6 @@ public class TestIndexReaderReopen exten System.out.println("TEST: modify index"); } IndexWriter w = new IndexWriter(dir, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))); - w.setInfoStream(VERBOSE ? System.out : null); w.deleteDocuments(new Term("field2", "a11")); w.deleteDocuments(new Term("field2", "b30")); w.close(); @@ -1020,14 +1019,14 @@ public class TestIndexReaderReopen exten } case 2: { IndexWriter w = new IndexWriter(dir, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))); - w.optimize(); + w.forceMerge(1); w.close(); break; } case 3: { IndexWriter w = new IndexWriter(dir, new IndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random))); w.addDocument(createDocument(101, 4)); - w.optimize(); + w.forceMerge(1); w.addDocument(createDocument(102, 4)); w.addDocument(createDocument(103, 4)); w.close(); @@ -1280,7 +1279,7 @@ public class TestIndexReaderReopen exten IndexReader r = IndexReader.open(dir, false); assertTrue(r instanceof DirectoryReader); IndexReader r1 = getOnlySegmentReader(r); - final int[] ints = FieldCache.DEFAULT.getInts(r1, "number"); + final int[] ints = FieldCache.DEFAULT.getInts(r1, "number", false); assertEquals(1, ints.length); assertEquals(17, ints[0]); @@ -1300,7 +1299,7 @@ public class TestIndexReaderReopen exten r.close(); assertTrue(((DirectoryReader) r2).readOnly); IndexReader[] subs = r2.getSequentialSubReaders(); - final int[] ints2 = FieldCache.DEFAULT.getInts(subs[0], "number"); + final int[] ints2 = FieldCache.DEFAULT.getInts(subs[0], "number", false); r2.close(); assertTrue(((SegmentReader) subs[0]).readOnly); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriter.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriter.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriter.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriter.java Mon Nov 14 22:36:20 2011 @@ -19,7 +19,6 @@ package org.apache.lucene.index; import java.io.ByteArrayOutputStream; import java.io.IOException; -import java.io.PrintStream; import java.io.Reader; import java.io.StringReader; import java.util.ArrayList; @@ -41,6 +40,8 @@ import org.apache.lucene.document.FieldT import org.apache.lucene.document.StringField; import org.apache.lucene.document.TextField; import org.apache.lucene.index.IndexWriterConfig.OpenMode; +import org.apache.lucene.index.codecs.Codec; +import org.apache.lucene.index.codecs.simpletext.SimpleTextCodec; import org.apache.lucene.search.DocIdSetIterator; import org.apache.lucene.search.FieldCache; import org.apache.lucene.search.IndexSearcher; @@ -52,16 +53,13 @@ import org.apache.lucene.search.TopDocs; import org.apache.lucene.search.spans.SpanTermQuery; import org.apache.lucene.store.AlreadyClosedException; import org.apache.lucene.store.Directory; -import org.apache.lucene.store.FSDirectory; import org.apache.lucene.store.IndexOutput; import org.apache.lucene.store.Lock; import org.apache.lucene.store.LockFactory; import org.apache.lucene.store.LockObtainFailedException; import org.apache.lucene.store.MockDirectoryWrapper; -import org.apache.lucene.store.NativeFSLockFactory; import org.apache.lucene.store.NoLockFactory; import org.apache.lucene.store.RAMDirectory; -import org.apache.lucene.store.SimpleFSDirectory; import org.apache.lucene.store.SimpleFSLockFactory; import org.apache.lucene.store.SingleInstanceLockFactory; import org.apache.lucene.util.BytesRef; @@ -106,10 +104,10 @@ public class TestIndexWriter extends Luc assertEquals(60, reader.numDocs()); reader.close(); - // optimize the index and check that the new doc count is correct + // merge the index down and check that the new doc count is correct writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); assertEquals(60, writer.numDocs()); - writer.optimize(); + writer.forceMerge(1); assertEquals(60, writer.maxDoc()); assertEquals(60, writer.numDocs()); writer.close(); @@ -623,7 +621,6 @@ public class TestIndexWriter extends Luc public void testEmptyDocAfterFlushingRealDoc() throws IOException { Directory dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - writer.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); FieldType customType = new FieldType(TextField.TYPE_STORED); customType.setStoreTermVectors(true); @@ -698,7 +695,6 @@ public class TestIndexWriter extends Luc System.out.println("TEST: iter=" + i); } IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2).setMergePolicy(newLogMergePolicy())); - writer.setInfoStream(VERBOSE ? System.out : null); //LogMergePolicy lmp = (LogMergePolicy) writer.getConfig().getMergePolicy(); //lmp.setMergeFactor(2); //lmp.setUseCompoundFile(false); @@ -733,7 +729,7 @@ public class TestIndexWriter extends Luc writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); //LogMergePolicy lmp2 = (LogMergePolicy) writer.getConfig().getMergePolicy(); //lmp2.setUseCompoundFile(false); - writer.optimize(); + writer.forceMerge(1); writer.close(); } } @@ -766,7 +762,6 @@ public class TestIndexWriter extends Luc IndexWriter writer = new IndexWriter(directory, conf); ((LogMergePolicy) writer.getConfig().getMergePolicy()).setMergeFactor(100); - writer.setInfoStream(VERBOSE ? System.out : null); for(int iter=0;iter<10;iter++) { if (VERBOSE) { @@ -831,7 +826,6 @@ public class TestIndexWriter extends Luc // Reopen writer = new IndexWriter(directory, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND).setMergePolicy(newLogMergePolicy())); - writer.setInfoStream(VERBOSE ? System.out : null); } writer.close(); } @@ -1159,7 +1153,6 @@ public class TestIndexWriter extends Luc IndexWriterConfig conf = newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2); w = new IndexWriter(dir, conf); - w.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); doc.add(newField("field", "some text contents", storedTextType)); @@ -1305,7 +1298,7 @@ public class TestIndexWriter extends Luc w.addDocument(doc); w.commit(); - w.optimize(); // force segment merge. + w.forceMerge(1); // force segment merge. w.close(); IndexReader ir = IndexReader.open(dir, true); @@ -1378,7 +1371,6 @@ public class TestIndexWriter extends Luc IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); ByteArrayOutputStream bos = new ByteArrayOutputStream(1024); - writer.setInfoStream(new PrintStream(bos)); writer.addDocument(new Document()); writer.close(); @@ -1443,7 +1435,7 @@ public class TestIndexWriter extends Luc List files = Arrays.asList(dir.listAll()); assertTrue(files.contains("_0.cfs")); w.addDocument(doc); - w.optimize(); + w.forceMerge(1); if (iter == 1) { w.commit(); } @@ -1454,10 +1446,10 @@ public class TestIndexWriter extends Luc // NOTE: here we rely on "Windows" behavior, ie, even // though IW wanted to delete _0.cfs since it was - // optimized away, because we have a reader open + // merged away, because we have a reader open // against this file, it should still be here: assertTrue(files.contains("_0.cfs")); - // optimize created this + // forceMerge created this //assertTrue(files.contains("_2.cfs")); w.deleteUnusedFiles(); @@ -1537,6 +1529,8 @@ public class TestIndexWriter extends Luc } public void testEmptyDirRollback() throws Exception { + // TODO: generalize this test + assumeFalse("test makes assumptions about file counts", Codec.getDefault() instanceof SimpleTextCodec); // Tests that if IW is created over an empty Directory, some documents are // indexed, flushed (but not committed) and then IW rolls back, then no // files are left in the Directory. @@ -1546,8 +1540,6 @@ public class TestIndexWriter extends Luc .setMaxBufferedDocs(2).setMergePolicy(newLogMergePolicy())); String[] files = dir.listAll(); - writer.setInfoStream(VERBOSE ? System.out : null); - // Creating over empty dir should not create any files, // or, at most the write.lock file final int extraFileCount; @@ -1615,11 +1607,7 @@ public class TestIndexWriter extends Luc Directory dir = newDirectory(); Random rand = random; RandomIndexWriter w = new RandomIndexWriter(rand, dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(_TestUtil.nextInt(rand, 5, 20))); - //w.w.setInfoStream(System.out); //w.w.setUseCompoundFile(false); - if (VERBOSE) { - w.w.setInfoStream(System.out); - } final int docCount = atLeast(200); final int fieldCount = _TestUtil.nextInt(rand, 1, 5); @@ -1705,7 +1693,7 @@ public class TestIndexWriter extends Luc } s.close(); r.close(); - w.optimize(); + w.forceMerge(1); } } w.close(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterCommit.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterCommit.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterCommit.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterCommit.java Mon Nov 14 22:36:20 2011 @@ -224,7 +224,7 @@ public class TestIndexWriterCommit exten } long midDiskUsage = dir.getMaxUsedSizeInBytes(); dir.resetMaxUsedSizeInBytes(); - writer.optimize(); + writer.forceMerge(1); writer.close(); IndexReader.open(dir, true).close(); @@ -246,11 +246,11 @@ public class TestIndexWriterCommit exten /* - * Verify that calling optimize when writer is open for + * Verify that calling forceMerge when writer is open for * "commit on close" works correctly both for rollback() * and close(). */ - public void testCommitOnCloseOptimize() throws IOException { + public void testCommitOnCloseForceMerge() throws IOException { MockDirectoryWrapper dir = newDirectory(); // Must disable throwing exc on double-write: this // test uses IW.rollback which easily results in @@ -268,51 +268,44 @@ public class TestIndexWriterCommit exten writer.close(); writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND)); - writer.optimize(); - - if (VERBOSE) { - writer.setInfoStream(System.out); - } + writer.forceMerge(1); // Open a reader before closing (commiting) the writer: IndexReader reader = IndexReader.open(dir, true); - // Reader should see index as unoptimized at this + // Reader should see index as multi-seg at this // point: - assertFalse("Reader incorrectly sees that the index is optimized", reader.isOptimized()); + assertTrue("Reader incorrectly sees one segment", reader.getSequentialSubReaders().length > 1); reader.close(); // Abort the writer: writer.rollback(); - TestIndexWriter.assertNoUnreferencedFiles(dir, "aborted writer after optimize"); + TestIndexWriter.assertNoUnreferencedFiles(dir, "aborted writer after forceMerge"); // Open a reader after aborting writer: reader = IndexReader.open(dir, true); - // Reader should still see index as unoptimized: - assertFalse("Reader incorrectly sees that the index is optimized", reader.isOptimized()); + // Reader should still see index as multi-segment + assertTrue("Reader incorrectly sees one segment", reader.getSequentialSubReaders().length > 1); reader.close(); if (VERBOSE) { - System.out.println("TEST: do real optimize"); + System.out.println("TEST: do real full merge"); } writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setOpenMode(OpenMode.APPEND)); - if (VERBOSE) { - writer.setInfoStream(System.out); - } - writer.optimize(); + writer.forceMerge(1); writer.close(); if (VERBOSE) { System.out.println("TEST: writer closed"); } - TestIndexWriter.assertNoUnreferencedFiles(dir, "aborted writer after optimize"); + TestIndexWriter.assertNoUnreferencedFiles(dir, "aborted writer after forceMerge"); // Open a reader after aborting writer: reader = IndexReader.open(dir, true); - // Reader should still see index as unoptimized: - assertTrue("Reader incorrectly sees that the index is unoptimized", reader.isOptimized()); + // Reader should see index as one segment + assertEquals("Reader incorrectly sees more than one segment", 1, reader.getSequentialSubReaders().length); reader.close(); dir.close(); } @@ -664,7 +657,7 @@ public class TestIndexWriterCommit exten r.close(); w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - w.optimize(); + w.forceMerge(1); w.close(); assertEquals("test1", IndexReader.getCommitUserData(dir).get("label")); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterConfig.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterConfig.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterConfig.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterConfig.java Mon Nov 14 22:36:20 2011 @@ -29,6 +29,7 @@ import org.apache.lucene.index.IndexWrit import org.apache.lucene.index.codecs.Codec; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.similarities.DefaultSimilarityProvider; +import org.apache.lucene.util.InfoStream; import org.apache.lucene.util.LuceneTestCase; import org.junit.Test; @@ -73,6 +74,7 @@ public class TestIndexWriterConfig exten assertEquals(FlushByRamOrCountsPolicy.class, conf.getFlushPolicy().getClass()); assertEquals(IndexWriterConfig.DEFAULT_RAM_PER_THREAD_HARD_LIMIT_MB, conf.getRAMPerThreadHardLimitMB()); assertEquals(Codec.getDefault(), conf.getCodec()); + assertEquals(InfoStream.getDefault(), conf.getInfoStream()); // Sanity check - validate that all getters are covered. Set getters = new HashSet(); getters.add("getAnalyzer"); @@ -98,6 +100,7 @@ public class TestIndexWriterConfig exten getters.add("getFlushPolicy"); getters.add("getRAMPerThreadHardLimitMB"); getters.add("getCodec"); + getters.add("getInfoStream"); for (Method m : IndexWriterConfig.class.getDeclaredMethods()) { if (m.getDeclaringClass() == IndexWriterConfig.class && m.getName().startsWith("get")) { Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterDelete.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterDelete.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterDelete.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterDelete.java Mon Nov 14 22:36:20 2011 @@ -64,7 +64,7 @@ public class TestIndexWriterDelete exten doc.add(newField("city", text[i], TextField.TYPE_STORED)); modifier.addDocument(doc); } - modifier.optimize(); + modifier.forceMerge(1); modifier.commit(); Term term = new Term("city", "Amsterdam"); @@ -93,7 +93,6 @@ public class TestIndexWriterDelete exten IndexWriter modifier = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false)).setMaxBufferedDocs(2) .setMaxBufferedDeleteTerms(2)); - modifier.setInfoStream(VERBOSE ? System.out : null); int id = 0; int value = 100; @@ -127,7 +126,6 @@ public class TestIndexWriterDelete exten IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false)).setMaxBufferedDeleteTerms(1)); - writer.setInfoStream(VERBOSE ? System.out : null); writer.addDocument(new Document()); writer.deleteDocuments(new Term("foobar", "1")); writer.deleteDocuments(new Term("foobar", "1")); @@ -147,7 +145,6 @@ public class TestIndexWriterDelete exten IndexWriter modifier = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false)).setMaxBufferedDocs(4) .setMaxBufferedDeleteTerms(4)); - modifier.setInfoStream(VERBOSE ? System.out : null); int id = 0; int value = 100; @@ -461,7 +458,6 @@ public class TestIndexWriterDelete exten .setMaxBufferedDeleteTerms(1000) .setMergeScheduler(new ConcurrentMergeScheduler())); ((ConcurrentMergeScheduler) modifier.getConfig().getMergeScheduler()).setSuppressExceptions(); - modifier.setInfoStream(VERBOSE ? System.out : null); // For each disk size, first try to commit against // dir that will hit random IOExceptions & disk @@ -696,7 +692,6 @@ public class TestIndexWriterDelete exten MockDirectoryWrapper dir = newDirectory(); IndexWriter modifier = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random, MockTokenizer.WHITESPACE, false)).setMaxBufferedDeleteTerms(2).setReaderPooling(false).setMergePolicy(newLogMergePolicy())); - modifier.setInfoStream(VERBOSE ? System.out : null); LogMergePolicy lmp = (LogMergePolicy) modifier.getConfig().getMergePolicy(); lmp.setUseCompoundFile(true); @@ -716,10 +711,10 @@ public class TestIndexWriterDelete exten // flush (and commit if ac) if (VERBOSE) { - System.out.println("TEST: now optimize"); + System.out.println("TEST: now full merge"); } - modifier.optimize(); + modifier.forceMerge(1); if (VERBOSE) { System.out.println("TEST: now commit"); } @@ -909,7 +904,6 @@ public class TestIndexWriterDelete exten } }; IndexWriter w = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer).setRAMBufferSizeMB(1.0).setMaxBufferedDocs(IndexWriterConfig.DISABLE_AUTO_FLUSH).setMaxBufferedDeleteTerms(IndexWriterConfig.DISABLE_AUTO_FLUSH)); - w.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); doc.add(newField("field", "go 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20", TextField.TYPE_UNSTORED)); int num = atLeast(3); @@ -952,7 +946,6 @@ public class TestIndexWriterDelete exten IndexWriter w = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setRAMBufferSizeMB(0.2f).setMaxBufferedDocs(1000).setMergePolicy(NoMergePolicy.NO_COMPOUND_FILES).setReaderPooling(false)); - w.setInfoStream(VERBOSE ? System.out : null); int count = 0; while(true) { Document doc = new Document(); @@ -998,7 +991,6 @@ public class TestIndexWriterDelete exten IndexWriter w = new IndexWriter(dir, newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)). setMaxBufferedDeleteTerms(flushAtDelCount).setMaxBufferedDocs(1000).setRAMBufferSizeMB(IndexWriterConfig.DISABLE_AUTO_FLUSH).setMergePolicy(NoMergePolicy.NO_COMPOUND_FILES).setReaderPooling(false)); - w.setInfoStream(VERBOSE ? System.out : null); int count = 0; while(true) { Document doc = new Document(); @@ -1045,7 +1037,6 @@ public class TestIndexWriterDelete exten sawAfterFlush.set(true); } }; - w.setInfoStream(VERBOSE ? System.out : null); int id = 0; while(true) { StringBuilder sb = new StringBuilder(); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterExceptions.java Mon Nov 14 22:36:20 2011 @@ -17,13 +17,10 @@ package org.apache.lucene.index; * limitations under the License. */ -import java.io.ByteArrayOutputStream; import java.io.IOException; -import java.io.PrintStream; import java.io.Reader; import java.io.StringReader; import java.util.ArrayList; -import java.util.Arrays; import java.util.Iterator; import java.util.List; import java.util.Random; @@ -45,6 +42,7 @@ import org.apache.lucene.store.MockDirec import org.apache.lucene.store.RAMDirectory; import org.apache.lucene.util.Bits; import org.apache.lucene.util.BytesRef; +import org.apache.lucene.util.InfoStream; import org.apache.lucene.util.LuceneTestCase; import org.apache.lucene.util._TestUtil; @@ -234,10 +232,6 @@ public class TestIndexWriterExceptions e } writer.commit(); - if (VERBOSE) { - writer.setInfoStream(System.out); - } - IndexerThread thread = new IndexerThread(0, writer); thread.run(); if (thread.failure != null) { @@ -278,10 +272,6 @@ public class TestIndexWriterExceptions e //writer.setMaxBufferedDocs(10); writer.commit(); - if (VERBOSE) { - writer.setInfoStream(System.out); - } - final int NUM_THREADS = 4; final IndexerThread[] threads = new IndexerThread[NUM_THREADS]; @@ -363,7 +353,6 @@ public class TestIndexWriterExceptions e public void testExceptionDocumentsWriterInit() throws IOException { Directory dir = newDirectory(); MockIndexWriter2 w = new MockIndexWriter2(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - w.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); doc.add(newField("field", "a field", TextField.TYPE_STORED)); w.addDocument(doc); @@ -382,7 +371,6 @@ public class TestIndexWriterExceptions e public void testExceptionJustBeforeFlush() throws IOException { Directory dir = newDirectory(); MockIndexWriter w = new MockIndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2)); - w.setInfoStream(VERBOSE ? System.out : null); Document doc = new Document(); doc.add(newField("field", "a field", TextField.TYPE_STORED)); w.addDocument(doc); @@ -607,7 +595,6 @@ public class TestIndexWriterExceptions e } MockDirectoryWrapper dir = newDirectory(); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, analyzer).setMergePolicy(newLogMergePolicy())); - writer.setInfoStream(VERBOSE ? System.out : null); // don't allow a sudden merge to clean up the deleted // doc below: @@ -667,7 +654,7 @@ public class TestIndexWriterExceptions e doc.add(newField("contents", "here are some contents", DocCopyIterator.custom5)); for(int j=0;j<17;j++) writer.addDocument(doc); - writer.optimize(); + writer.forceMerge(1); writer.close(); reader = IndexReader.open(dir, true); @@ -781,7 +768,7 @@ public class TestIndexWriterExceptions e doc.add(newField("contents", "here are some contents", DocCopyIterator.custom5)); for(int j=0;j<17;j++) writer.addDocument(doc); - writer.optimize(); + writer.forceMerge(1); writer.close(); reader = IndexReader.open(dir, true); @@ -937,7 +924,7 @@ public class TestIndexWriterExceptions e } } - public void testOptimizeExceptions() throws IOException { + public void testForceMergeExceptions() throws IOException { Directory startDir = newDirectory(); IndexWriterConfig conf = newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMaxBufferedDocs(2).setMergePolicy(newLogMergePolicy()); ((LogMergePolicy) conf.getMergePolicy()).setMergeFactor(100); @@ -955,13 +942,12 @@ public class TestIndexWriterExceptions e conf = newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setMergeScheduler(new ConcurrentMergeScheduler()); ((ConcurrentMergeScheduler) conf.getMergeScheduler()).setSuppressExceptions(); w = new IndexWriter(dir, conf); - w.setInfoStream(VERBOSE ? System.out : null); dir.setRandomIOExceptionRate(0.5); try { - w.optimize(); + w.forceMerge(1); } catch (IOException ioe) { if (ioe.getCause() == null) - fail("optimize threw IOException without root cause"); + fail("forceMerge threw IOException without root cause"); } dir.setRandomIOExceptionRate(0); w.close(); @@ -976,18 +962,19 @@ public class TestIndexWriterExceptions e final List thrown = new ArrayList(); final Directory dir = newDirectory(); final IndexWriter writer = new IndexWriter(dir, - newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))) { + newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random)).setInfoStream(new InfoStream() { @Override - public void message(final String message) { + public void message(String component, final String message) { if (message.startsWith("now flush at close") && 0 == thrown.size()) { thrown.add(null); throw new OutOfMemoryError("fake OOME at " + message); } } - }; - // need to set an info stream so message is called - writer.setInfoStream(new PrintStream(new ByteArrayOutputStream())); + @Override + public void close() throws IOException {} + })); + try { writer.close(); fail("OutOfMemoryError expected"); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMergePolicy.java Mon Nov 14 22:36:20 2011 @@ -108,8 +108,6 @@ public class TestIndexWriterMergePolicy setMergeScheduler(new SerialMergeScheduler()) ); - writer.setInfoStream(VERBOSE ? System.out : null); - for (int i = 0; i < 250; i++) { addDoc(writer); checkInvariants(writer); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterMerging.java Mon Nov 14 22:36:20 2011 @@ -63,9 +63,8 @@ public class TestIndexWriterMerging exte newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)). setMergePolicy(newLogMergePolicy(2)) ); - writer.setInfoStream(VERBOSE ? System.out : null); writer.addIndexes(indexA, indexB); - writer.optimize(); + writer.forceMerge(1); writer.close(); fail = verifyIndex(merged, 0); @@ -125,7 +124,6 @@ public class TestIndexWriterMerging exte TEST_VERSION_CURRENT, new MockAnalyzer(random)) .setMaxBufferedDocs(2).setRAMBufferSizeMB( IndexWriterConfig.DISABLE_AUTO_FLUSH)); - writer.setInfoStream(VERBOSE ? System.out : null); Document document = new Document(); document = new Document(); @@ -233,8 +231,6 @@ public class TestIndexWriterMerging exte setMergePolicy(newLogMergePolicy(50)) ); - writer.setInfoStream(VERBOSE ? System.out : null); - FieldType customType = new FieldType(); customType.setStored(true); @@ -308,7 +304,6 @@ public class TestIndexWriterMerging exte lmp.setMaxMergeDocs(20); lmp.setMergeFactor(2); IndexWriter iw = new IndexWriter(dir, conf); - iw.setInfoStream(VERBOSE ? System.out : null); Document document = new Document(); FieldType customType = new FieldType(TextField.TYPE_UNSTORED); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterNRTIsCurrent.java Mon Nov 14 22:36:20 2011 @@ -41,9 +41,6 @@ public class TestIndexWriterNRTIsCurrent IndexWriterConfig conf = newIndexWriterConfig(TEST_VERSION_CURRENT, new MockAnalyzer(random)); IndexWriter writer = new IndexWriter(dir, conf); - if (VERBOSE) { - writer.setInfoStream(System.out); - } ReaderHolder holder = new ReaderHolder(); ReaderThread[] threads = new ReaderThread[atLeast(3)]; final CountDownLatch latch = new CountDownLatch(1); Modified: lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java URL: http://svn.apache.org/viewvc/lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java?rev=1201946&r1=1201945&r2=1201946&view=diff ============================================================================== --- lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java (original) +++ lucene/dev/branches/solrcloud/lucene/src/test/org/apache/lucene/index/TestIndexWriterOnDiskFull.java Mon Nov 14 22:36:20 2011 @@ -24,7 +24,6 @@ import org.apache.lucene.document.Docume import org.apache.lucene.document.FieldType; import org.apache.lucene.document.TextField; import org.apache.lucene.index.IndexWriterConfig.OpenMode; -import org.apache.lucene.index.codecs.Codec; import org.apache.lucene.search.IndexSearcher; import org.apache.lucene.search.ScoreDoc; import org.apache.lucene.search.TermQuery; @@ -61,7 +60,6 @@ public class TestIndexWriterOnDiskFull e MockDirectoryWrapper dir = new MockDirectoryWrapper(random, new RAMDirectory()); dir.setMaxSizeInBytes(diskFree); IndexWriter writer = new IndexWriter(dir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); - writer.setInfoStream(VERBOSE ? System.out : null); MergeScheduler ms = writer.getConfig().getMergeScheduler(); if (ms instanceof ConcurrentMergeScheduler) { // This test intentionally produces exceptions @@ -181,7 +179,7 @@ public class TestIndexWriterOnDiskFull e } // Now, build a starting index that has START_COUNT docs. We - // will then try to addIndexesNoOptimize into a copy of this: + // will then try to addIndexes into a copy of this: MockDirectoryWrapper startDir = newDirectory(); IndexWriter writer = new IndexWriter(startDir, newIndexWriterConfig( TEST_VERSION_CURRENT, new MockAnalyzer(random))); for(int j=0;j