lucene-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Michael McCandless <luc...@mikemccandless.com>
Subject Re: svn commit: r822284 - /lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibility.java
Date Tue, 06 Oct 2009 14:36:36 GMT
OK will do.

Mike

On Tue, Oct 6, 2009 at 10:23 AM, Uwe Schindler <uwe@thetaphi.de> wrote:
> Can you add this patch to backwards, too? I forgot, that some of the
> backwards-changes also applied to BW, but for completeness, not sure, if a
> tag is also needed.
>
> -----
> Uwe Schindler
> H.-H.-Meier-Allee 63, D-28213 Bremen
> http://www.thetaphi.de
> eMail: uwe@thetaphi.de
>
>
>> -----Original Message-----
>> From: mikemccand@apache.org [mailto:mikemccand@apache.org]
>> Sent: Tuesday, October 06, 2009 4:13 PM
>> To: java-commits@lucene.apache.org
>> Subject: svn commit: r822284 -
>> /lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibi
>> lity.java
>>
>> Author: mikemccand
>> Date: Tue Oct  6 14:12:46 2009
>> New Revision: 822284
>>
>> URL: http://svn.apache.org/viewvc?rev=822284&view=rev
>> Log:
>> fix TestBackwardsCompability to not use deprecated APIs
>>
>> Modified:
>>
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> ity.java
>>
>> Modified:
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> ity.java
>> URL:
>> http://svn.apache.org/viewvc/lucene/java/trunk/src/test/org/apache/lucene/
>> index/TestBackwardsCompatibility.java?rev=822284&r1=822283&r2=822284&view=
>> diff
>> ==========================================================================
>> ====
>> ---
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> ity.java (original)
>> +++
>> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil
>> ity.java Tue Oct  6 14:12:46 2009
>> @@ -158,11 +158,7 @@
>>      for(int i=0;i<oldNames.length;i++) {
>>        String dirName = "src/test/org/apache/lucene/index/index." +
>> oldNames[i];
>>        unzip(dirName, oldNames[i]);
>> -      changeIndexNoAdds(oldNames[i], true);
>> -      rmDir(oldNames[i]);
>> -
>> -      unzip(dirName, oldNames[i]);
>> -      changeIndexNoAdds(oldNames[i], false);
>> +      changeIndexNoAdds(oldNames[i]);
>>        rmDir(oldNames[i]);
>>      }
>>    }
>> @@ -171,11 +167,7 @@
>>      for(int i=0;i<oldNames.length;i++) {
>>        String dirName = "src/test/org/apache/lucene/index/index." +
>> oldNames[i];
>>        unzip(dirName, oldNames[i]);
>> -      changeIndexWithAdds(oldNames[i], true);
>> -      rmDir(oldNames[i]);
>> -
>> -      unzip(dirName, oldNames[i]);
>> -      changeIndexWithAdds(oldNames[i], false);
>> +      changeIndexWithAdds(oldNames[i]);
>>        rmDir(oldNames[i]);
>>      }
>>    }
>> @@ -196,7 +188,7 @@
>>      dirName = fullDir(dirName);
>>
>>      Directory dir = FSDirectory.open(new File(dirName));
>> -    IndexSearcher searcher = new IndexSearcher(dir);
>> +    IndexSearcher searcher = new IndexSearcher(dir, true);
>>      IndexReader reader = searcher.getIndexReader();
>>
>>      _TestUtil.checkIndex(dir);
>> @@ -267,14 +259,14 @@
>>
>>    /* Open pre-lockless index, add docs, do a delete &
>>     * setNorm, and search */
>> -  public void changeIndexWithAdds(String dirName, boolean autoCommit)
>> throws IOException {
>> +  public void changeIndexWithAdds(String dirName) throws IOException {
>>      String origDirName = dirName;
>>      dirName = fullDir(dirName);
>>
>>      Directory dir = FSDirectory.open(new File(dirName));
>>
>>      // open writer
>> -    IndexWriter writer = new IndexWriter(dir, autoCommit, new
>> WhitespaceAnalyzer(), false);
>> +    IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(),
>> false, IndexWriter.MaxFieldLength.UNLIMITED);
>>
>>      // add 10 docs
>>      for(int i=0;i<10;i++) {
>> @@ -288,11 +280,11 @@
>>      } else {
>>        expected = 46;
>>      }
>> -    assertEquals("wrong doc count", expected, writer.docCount());
>> +    assertEquals("wrong doc count", expected, writer.maxDoc());
>>      writer.close();
>>
>>      // make sure searching sees right # hits
>> -    IndexSearcher searcher = new IndexSearcher(dir);
>> +    IndexSearcher searcher = new IndexSearcher(dir, true);
>>      ScoreDoc[] hits = searcher.search(new TermQuery(new Term("content",
>> "aaa")), null, 1000).scoreDocs;
>>      Document d = searcher.doc(hits[0].doc);
>>      assertEquals("wrong first document", "21", d.get("id"));
>> @@ -301,7 +293,7 @@
>>
>>      // make sure we can do delete & setNorm against this
>>      // pre-lockless segment:
>> -    IndexReader reader = IndexReader.open(dir);
>> +    IndexReader reader = IndexReader.open(dir, false);
>>      Term searchTerm = new Term("id", "6");
>>      int delCount = reader.deleteDocuments(searchTerm);
>>      assertEquals("wrong delete count", 1, delCount);
>> @@ -309,7 +301,7 @@
>>      reader.close();
>>
>>      // make sure they "took":
>> -    searcher = new IndexSearcher(dir);
>> +    searcher = new IndexSearcher(dir, true);
>>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> null, 1000).scoreDocs;
>>      assertEquals("wrong number of hits", 43, hits.length);
>>      d = searcher.doc(hits[0].doc);
>> @@ -318,11 +310,11 @@
>>      searcher.close();
>>
>>      // optimize
>> -    writer = new IndexWriter(dir, autoCommit, new WhitespaceAnalyzer(),
>> false);
>> +    writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false,
>> IndexWriter.MaxFieldLength.UNLIMITED);
>>      writer.optimize();
>>      writer.close();
>>
>> -    searcher = new IndexSearcher(dir);
>> +    searcher = new IndexSearcher(dir, true);
>>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> null, 1000).scoreDocs;
>>      assertEquals("wrong number of hits", 43, hits.length);
>>      d = searcher.doc(hits[0].doc);
>> @@ -335,14 +327,14 @@
>>
>>    /* Open pre-lockless index, add docs, do a delete &
>>     * setNorm, and search */
>> -  public void changeIndexNoAdds(String dirName, boolean autoCommit)
>> throws IOException {
>> +  public void changeIndexNoAdds(String dirName) throws IOException {
>>
>>      dirName = fullDir(dirName);
>>
>>      Directory dir = FSDirectory.open(new File(dirName));
>>
>>      // make sure searching sees right # hits
>> -    IndexSearcher searcher = new IndexSearcher(dir);
>> +    IndexSearcher searcher = new IndexSearcher(dir, true);
>>      ScoreDoc[] hits = searcher.search(new TermQuery(new Term("content",
>> "aaa")), null, 1000).scoreDocs;
>>      assertEquals("wrong number of hits", 34, hits.length);
>>      Document d = searcher.doc(hits[0].doc);
>> @@ -351,7 +343,7 @@
>>
>>      // make sure we can do a delete & setNorm against this
>>      // pre-lockless segment:
>> -    IndexReader reader = IndexReader.open(dir);
>> +    IndexReader reader = IndexReader.open(dir, false);
>>      Term searchTerm = new Term("id", "6");
>>      int delCount = reader.deleteDocuments(searchTerm);
>>      assertEquals("wrong delete count", 1, delCount);
>> @@ -359,7 +351,7 @@
>>      reader.close();
>>
>>      // make sure they "took":
>> -    searcher = new IndexSearcher(dir);
>> +    searcher = new IndexSearcher(dir, true);
>>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> null, 1000).scoreDocs;
>>      assertEquals("wrong number of hits", 33, hits.length);
>>      d = searcher.doc(hits[0].doc);
>> @@ -368,11 +360,11 @@
>>      searcher.close();
>>
>>      // optimize
>> -    IndexWriter writer = new IndexWriter(dir, autoCommit, new
>> WhitespaceAnalyzer(), false);
>> +    IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(),
>> false, IndexWriter.MaxFieldLength.UNLIMITED);
>>      writer.optimize();
>>      writer.close();
>>
>> -    searcher = new IndexSearcher(dir);
>> +    searcher = new IndexSearcher(dir, true);
>>      hits = searcher.search(new TermQuery(new Term("content", "aaa")),
>> null, 1000).scoreDocs;
>>      assertEquals("wrong number of hits", 33, hits.length);
>>      d = searcher.doc(hits[0].doc);
>> @@ -397,7 +389,7 @@
>>      for(int i=0;i<35;i++) {
>>        addDoc(writer, i);
>>      }
>> -    assertEquals("wrong doc count", 35, writer.docCount());
>> +    assertEquals("wrong doc count", 35, writer.maxDoc());
>>      writer.close();
>>
>>      // open fresh writer so we get no prx file in the added segment
>> @@ -408,7 +400,7 @@
>>      writer.close();
>>
>>      // Delete one doc so we get a .del file:
>> -    IndexReader reader = IndexReader.open(dir);
>> +    IndexReader reader = IndexReader.open(dir, false);
>>      Term searchTerm = new Term("id", "7");
>>      int delCount = reader.deleteDocuments(searchTerm);
>>      assertEquals("didn't delete the right number of documents", 1,
>> delCount);
>> @@ -422,71 +414,66 @@
>>
>>    public void testExactFileNames() throws IOException {
>>
>> -    for(int pass=0;pass<2;pass++) {
>> -
>> -      String outputDir = "lucene.backwardscompat0.index";
>> -      rmDir(outputDir);
>> +    String outputDir = "lucene.backwardscompat0.index";
>> +    rmDir(outputDir);
>>
>> -      try {
>> -        Directory dir = FSDirectory.open(new File(fullDir(outputDir)));
>> +    try {
>> +      Directory dir = FSDirectory.open(new File(fullDir(outputDir)));
>>
>> -        boolean autoCommit = 0 == pass;
>> -
>> -        IndexWriter writer = new IndexWriter(dir, autoCommit, new
>> WhitespaceAnalyzer(), true);
>> -        writer.setRAMBufferSizeMB(16.0);
>> -        for(int i=0;i<35;i++) {
>> -          addDoc(writer, i);
>> -        }
>> -        assertEquals("wrong doc count", 35, writer.docCount());
>> -        writer.close();
>> +      IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(),
>> true, IndexWriter.MaxFieldLength.UNLIMITED);
>> +      writer.setRAMBufferSizeMB(16.0);
>> +      for(int i=0;i<35;i++) {
>> +        addDoc(writer, i);
>> +      }
>> +      assertEquals("wrong doc count", 35, writer.maxDoc());
>> +      writer.close();
>>
>> -        // Delete one doc so we get a .del file:
>> -        IndexReader reader = IndexReader.open(dir);
>> -        Term searchTerm = new Term("id", "7");
>> -        int delCount = reader.deleteDocuments(searchTerm);
>> -        assertEquals("didn't delete the right number of documents", 1,
>> delCount);
>> -
>> -        // Set one norm so we get a .s0 file:
>> -        reader.setNorm(21, "content", (float) 1.5);
>> -        reader.close();
>> -
>> -        // The numbering of fields can vary depending on which
>> -        // JRE is in use.  On some JREs we see content bound to
>> -        // field 0; on others, field 1.  So, here we have to
>> -        // figure out which field number corresponds to
>> -        // "content", and then set our expected file names below
>> -        // accordingly:
>> -        CompoundFileReader cfsReader = new CompoundFileReader(dir,
>> "_0.cfs");
>> -        FieldInfos fieldInfos = new FieldInfos(cfsReader, "_0.fnm");
>> -        int contentFieldIndex = -1;
>> -        for(int i=0;i<fieldInfos.size();i++) {
>> -          FieldInfo fi = fieldInfos.fieldInfo(i);
>> -          if (fi.name.equals("content")) {
>> -            contentFieldIndex = i;
>> -            break;
>> -          }
>> +      // Delete one doc so we get a .del file:
>> +      IndexReader reader = IndexReader.open(dir, false);
>> +      Term searchTerm = new Term("id", "7");
>> +      int delCount = reader.deleteDocuments(searchTerm);
>> +      assertEquals("didn't delete the right number of documents", 1,
>> delCount);
>> +
>> +      // Set one norm so we get a .s0 file:
>> +      reader.setNorm(21, "content", (float) 1.5);
>> +      reader.close();
>> +
>> +      // The numbering of fields can vary depending on which
>> +      // JRE is in use.  On some JREs we see content bound to
>> +      // field 0; on others, field 1.  So, here we have to
>> +      // figure out which field number corresponds to
>> +      // "content", and then set our expected file names below
>> +      // accordingly:
>> +      CompoundFileReader cfsReader = new CompoundFileReader(dir,
>> "_0.cfs");
>> +      FieldInfos fieldInfos = new FieldInfos(cfsReader, "_0.fnm");
>> +      int contentFieldIndex = -1;
>> +      for(int i=0;i<fieldInfos.size();i++) {
>> +        FieldInfo fi = fieldInfos.fieldInfo(i);
>> +        if (fi.name.equals("content")) {
>> +          contentFieldIndex = i;
>> +          break;
>>          }
>> -        cfsReader.close();
>> -        assertTrue("could not locate the 'content' field number in the
>> _2.cfs segment", contentFieldIndex != -1);
>> +      }
>> +      cfsReader.close();
>> +      assertTrue("could not locate the 'content' field number in the
>> _2.cfs segment", contentFieldIndex != -1);
>>
>> -        // Now verify file names:
>> -        String[] expected;
>> -        expected = new String[] {"_0.cfs",
>> -                    "_0_1.del",
>> -                    "_0_1.s" + contentFieldIndex,
>> -                    "segments_3",
>> -                    "segments.gen"};
>> -
>> -        String[] actual = dir.listAll();
>> -        Arrays.sort(expected);
>> -        Arrays.sort(actual);
>> -        if (!Arrays.equals(expected, actual)) {
>> -          fail("incorrect filenames in index: expected:\n    " +
>> asString(expected) + "\n  actual:\n    " + asString(actual));
>> -        }
>> -        dir.close();
>> -      } finally {
>> -        rmDir(outputDir);
>> +      // Now verify file names:
>> +      String[] expected;
>> +      expected = new String[] {"_0.cfs",
>> +                               "_0_1.del",
>> +                               "_0_1.s" + contentFieldIndex,
>> +                               "segments_3",
>> +                               "segments.gen"};
>> +
>> +      String[] actual = dir.listAll();
>> +      Arrays.sort(expected);
>> +      Arrays.sort(actual);
>> +      if (!Arrays.equals(expected, actual)) {
>> +        fail("incorrect filenames in index: expected:\n    " +
>> asString(expected) + "\n  actual:\n    " + asString(actual));
>>        }
>> +      dir.close();
>> +    } finally {
>> +      rmDir(outputDir);
>>      }
>>    }
>>
>> @@ -509,17 +496,17 @@
>>      doc.add(new Field("autf8", "Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000
>> \u2620 ab\ud917\udc17cd", Field.Store.YES, Field.Index.ANALYZED,
>> Field.TermVector.WITH_POSITIONS_OFFSETS));
>>      doc.add(new Field("utf8", "Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000
>> \u2620 ab\ud917\udc17cd", Field.Store.YES, Field.Index.ANALYZED,
>> Field.TermVector.WITH_POSITIONS_OFFSETS));
>>      doc.add(new Field("content2", "here is more content with aaa aaa
>> aaa", Field.Store.YES, Field.Index.ANALYZED,
>> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> -    doc.add(new Field("fie\u2C77ld", "field with non-ascii name",
>> Field.Store.YES, Field.Index.TOKENIZED,
>> Field.TermVector.WITH_POSITIONS_OFFSETS));
>> +    doc.add(new Field("fie\u2C77ld", "field with non-ascii name",
>> Field.Store.YES, Field.Index.ANALYZED,
>> Field.TermVector.WITH_POSITIONS_OFFSETS));
>>      writer.addDocument(doc);
>>    }
>>
>>    private void addNoProxDoc(IndexWriter writer) throws IOException {
>>      Document doc = new Document();
>>      Field f = new Field("content3", "aaa", Field.Store.YES,
>> Field.Index.ANALYZED);
>> -    f.setOmitTf(true);
>> +    f.setOmitTermFreqAndPositions(true);
>>      doc.add(f);
>>      f = new Field("content4", "aaa", Field.Store.YES, Field.Index.NO);
>> -    f.setOmitTf(true);
>> +    f.setOmitTermFreqAndPositions(true);
>>      doc.add(f);
>>      writer.addDocument(doc);
>>    }
>>
>
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: java-dev-unsubscribe@lucene.apache.org
> For additional commands, e-mail: java-dev-help@lucene.apache.org
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: java-dev-unsubscribe@lucene.apache.org
For additional commands, e-mail: java-dev-help@lucene.apache.org


Mime
View raw message