Return-Path: Delivered-To: apmail-lucene-java-commits-archive@www.apache.org Received: (qmail 71411 invoked from network); 31 Jan 2010 16:01:52 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.3) by minotaur.apache.org with SMTP; 31 Jan 2010 16:01:52 -0000 Received: (qmail 23798 invoked by uid 500); 31 Jan 2010 16:01:52 -0000 Delivered-To: apmail-lucene-java-commits-archive@lucene.apache.org Received: (qmail 23706 invoked by uid 500); 31 Jan 2010 16:01:52 -0000 Mailing-List: contact java-commits-help@lucene.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: java-dev@lucene.apache.org Delivered-To: mailing list java-commits@lucene.apache.org Received: (qmail 23697 invoked by uid 99); 31 Jan 2010 16:01:52 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 31 Jan 2010 16:01:52 +0000 X-ASF-Spam-Status: No, hits=-2000.0 required=10.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.4] (HELO eris.apache.org) (140.211.11.4) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 31 Jan 2010 16:01:41 +0000 Received: by eris.apache.org (Postfix, from userid 65534) id 47AB423889C5; Sun, 31 Jan 2010 16:01:19 +0000 (UTC) Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Subject: svn commit: r905065 - in /lucene/java/trunk/contrib/analyzers/common/src: java/org/apache/lucene/analysis/ar/ java/org/apache/lucene/analysis/fa/ java/org/apache/lucene/analysis/ru/ test/org/apache/lucene/analysis/ar/ test/org/apache/lucene/analysis/bg... Date: Sun, 31 Jan 2010 16:01:18 -0000 To: java-commits@lucene.apache.org From: simonw@apache.org X-Mailer: svnmailer-1.0.8 Message-Id: <20100131160119.47AB423889C5@eris.apache.org> X-Virus-Checked: Checked by ClamAV on apache.org Author: simonw Date: Sun Jan 31 16:01:17 2010 New Revision: 905065 URL: http://svn.apache.org/viewvc?rev=905065&view=rev Log: LUCENE-2242: Contrib CharTokenizer classes should be instantiated using their new Version based ctors Modified: lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicAnalyzer.java lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/fa/PersianAnalyzer.java lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/RussianAnalyzer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicNormalizationFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicStemFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/bg/TestBulgarianStemmer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/br/TestBrazilianStemmer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cn/TestChineseTokenizer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/compound/TestCompoundWordTokenFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cz/TestCzechStemmer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/de/TestGermanStemFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/fa/TestPersianNormalizationFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAndSuffixAwareTokenFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAwareTokenFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/NGramTokenFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/NumericPayloadTokenFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TokenOffsetPayloadTokenFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TypeAsPayloadTokenFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/query/QueryAutoStopWordAnalyzerTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianAnalyzer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianLetterTokenizer.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapperTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/TestShingleMatrixFilter.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/DateRecognizerSinkTokenizerTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenRangeSinkTokenizerTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenTypeSinkTokenizerTest.java lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/tr/TestTurkishLowerCaseFilter.java Modified: lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicAnalyzer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicAnalyzer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicAnalyzer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ar/ArabicAnalyzer.java Sun Jan 31 16:01:17 2010 @@ -174,7 +174,7 @@ @Override protected TokenStreamComponents createComponents(String fieldName, Reader reader) { - final Tokenizer source = new ArabicLetterTokenizer(reader); + final Tokenizer source = new ArabicLetterTokenizer(matchVersion, reader); TokenStream result = new LowerCaseFilter(matchVersion, source); // the order here is important: the stopword list is not normalized! result = new StopFilter( matchVersion, result, stopwords); Modified: lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/fa/PersianAnalyzer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/fa/PersianAnalyzer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/fa/PersianAnalyzer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/fa/PersianAnalyzer.java Sun Jan 31 16:01:17 2010 @@ -147,7 +147,7 @@ @Override protected TokenStreamComponents createComponents(String fieldName, Reader reader) { - final Tokenizer source = new ArabicLetterTokenizer(reader); + final Tokenizer source = new ArabicLetterTokenizer(matchVersion, reader); TokenStream result = new LowerCaseFilter(matchVersion, source); result = new ArabicNormalizationFilter(result); /* additional persian-specific normalization */ Modified: lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/RussianAnalyzer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/RussianAnalyzer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/RussianAnalyzer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/java/org/apache/lucene/analysis/ru/RussianAnalyzer.java Sun Jan 31 16:01:17 2010 @@ -139,7 +139,7 @@ @Override protected TokenStreamComponents createComponents(String fieldName, Reader reader) { - final Tokenizer source = new RussianLetterTokenizer(reader); + final Tokenizer source = new RussianLetterTokenizer(matchVersion, reader); TokenStream result = new LowerCaseFilter(matchVersion, source); result = new StopFilter(matchVersion, result, stopwords); if(!stemExclusionSet.isEmpty()) Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicNormalizationFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicNormalizationFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicNormalizationFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicNormalizationFilter.java Sun Jan 31 16:01:17 2010 @@ -22,6 +22,7 @@ import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.tokenattributes.TermAttribute; +import org.apache.lucene.util.Version; /** * Test the Arabic Normalization Filter @@ -86,7 +87,7 @@ } private void check(final String input, final String expected) throws IOException { - ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(new StringReader(input)); + ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(Version.LUCENE_CURRENT, new StringReader(input)); ArabicNormalizationFilter filter = new ArabicNormalizationFilter(tokenStream); assertTokenStreamContents(filter, new String[]{expected}); } Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicStemFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicStemFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicStemFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ar/TestArabicStemFilter.java Sun Jan 31 16:01:17 2010 @@ -118,14 +118,14 @@ public void testWithKeywordAttribute() throws IOException { CharArraySet set = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set.add("ساهدهات"); - ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(new StringReader("ساهدهات")); + ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(Version.LUCENE_CURRENT, new StringReader("ساهدهات")); ArabicStemFilter filter = new ArabicStemFilter(new KeywordMarkerTokenFilter(tokenStream, set)); assertTokenStreamContents(filter, new String[]{"ساهدهات"}); } private void check(final String input, final String expected) throws IOException { - ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(new StringReader(input)); + ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(Version.LUCENE_CURRENT, new StringReader(input)); ArabicStemFilter filter = new ArabicStemFilter(tokenStream); assertTokenStreamContents(filter, new String[]{expected}); } Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/bg/TestBulgarianStemmer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/bg/TestBulgarianStemmer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/bg/TestBulgarianStemmer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/bg/TestBulgarianStemmer.java Sun Jan 31 16:01:17 2010 @@ -215,7 +215,7 @@ public void testWithKeywordAttribute() throws IOException { CharArraySet set = new CharArraySet(Version.LUCENE_31, 1, true); set.add("строеве"); - WhitespaceTokenizer tokenStream = new WhitespaceTokenizer( + WhitespaceTokenizer tokenStream = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("строевете строеве")); BulgarianStemFilter filter = new BulgarianStemFilter( Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/br/TestBrazilianStemmer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/br/TestBrazilianStemmer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/br/TestBrazilianStemmer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/br/TestBrazilianStemmer.java Sun Jan 31 16:01:17 2010 @@ -145,7 +145,7 @@ CharArraySet set = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set.add("Brasília"); BrazilianStemFilter filter = new BrazilianStemFilter( - new LowerCaseTokenizer(new StringReader("Brasília Brasilia")), set); + new LowerCaseTokenizer(Version.LUCENE_CURRENT, new StringReader("Brasília Brasilia")), set); assertTokenStreamContents(filter, new String[] { "brasília", "brasil" }); } @@ -153,7 +153,7 @@ CharArraySet set = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set.add("Brasília"); BrazilianStemFilter filter = new BrazilianStemFilter( - new KeywordMarkerTokenFilter(new LowerCaseTokenizer(new StringReader( + new KeywordMarkerTokenFilter(new LowerCaseTokenizer(Version.LUCENE_CURRENT, new StringReader( "Brasília Brasilia")), set)); assertTokenStreamContents(filter, new String[] { "brasília", "brasil" }); } @@ -164,7 +164,7 @@ CharArraySet set1 = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set1.add("Brasilia"); BrazilianStemFilter filter = new BrazilianStemFilter( - new KeywordMarkerTokenFilter(new LowerCaseTokenizer(new StringReader( + new KeywordMarkerTokenFilter(new LowerCaseTokenizer(Version.LUCENE_CURRENT, new StringReader( "Brasília Brasilia")), set), set1); assertTokenStreamContents(filter, new String[] { "brasília", "brasilia" }); } Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cn/TestChineseTokenizer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cn/TestChineseTokenizer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cn/TestChineseTokenizer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cn/TestChineseTokenizer.java Sun Jan 31 16:01:17 2010 @@ -26,6 +26,7 @@ import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.tokenattributes.OffsetAttribute; +import org.apache.lucene.util.Version; /** @deprecated Remove this test when ChineseAnalyzer is removed. */ @@ -79,7 +80,7 @@ private class JustChineseFilterAnalyzer extends Analyzer { @Override public TokenStream tokenStream(String fieldName, Reader reader) { - return new ChineseFilter(new WhitespaceTokenizer(reader)); + return new ChineseFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, reader)); } } Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/compound/TestCompoundWordTokenFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/compound/TestCompoundWordTokenFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/compound/TestCompoundWordTokenFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/compound/TestCompoundWordTokenFilter.java Sun Jan 31 16:01:17 2010 @@ -28,6 +28,7 @@ import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.compound.hyphenation.HyphenationTree; import org.apache.lucene.analysis.tokenattributes.TermAttribute; +import org.apache.lucene.util.Version; public class TestCompoundWordTokenFilter extends BaseTokenStreamTestCase { static final File dataDir = new File(System.getProperty("dataDir", "./bin")); @@ -46,8 +47,8 @@ HyphenationTree hyphenator = HyphenationCompoundWordTokenFilter .getHyphenationTree(reader); - HyphenationCompoundWordTokenFilter tf = new HyphenationCompoundWordTokenFilter( - new WhitespaceTokenizer(new StringReader( + HyphenationCompoundWordTokenFilter tf = new HyphenationCompoundWordTokenFilter(Version.LUCENE_CURRENT, + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "min veninde som er lidt af en læsehest")), hyphenator, dict, CompoundWordTokenFilterBase.DEFAULT_MIN_WORD_SIZE, CompoundWordTokenFilterBase.DEFAULT_MIN_SUBWORD_SIZE, @@ -66,8 +67,8 @@ .getHyphenationTree(reader); // the word basket will not be added due to the longest match option - HyphenationCompoundWordTokenFilter tf = new HyphenationCompoundWordTokenFilter( - new WhitespaceTokenizer(new StringReader( + HyphenationCompoundWordTokenFilter tf = new HyphenationCompoundWordTokenFilter(Version.LUCENE_CURRENT, + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "basketballkurv")), hyphenator, dict, CompoundWordTokenFilterBase.DEFAULT_MIN_WORD_SIZE, CompoundWordTokenFilterBase.DEFAULT_MIN_SUBWORD_SIZE, 40, true); @@ -83,8 +84,8 @@ "Pelar", "Glas", "Ögon", "Fodral", "Bas", "Fiol", "Makare", "Gesäll", "Sko", "Vind", "Rute", "Torkare", "Blad" }; - DictionaryCompoundWordTokenFilter tf = new DictionaryCompoundWordTokenFilter( - new WhitespaceTokenizer( + DictionaryCompoundWordTokenFilter tf = new DictionaryCompoundWordTokenFilter(Version.LUCENE_CURRENT, + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "Bildörr Bilmotor Biltak Slagborr Hammarborr Pelarborr Glasögonfodral Basfiolsfodral Basfiolsfodralmakaregesäll Skomakare Vindrutetorkare Vindrutetorkarblad abba")), dict); @@ -112,8 +113,8 @@ "Pelar", "Glas", "Ögon", "Fodral", "Bas", "Fiols", "Makare", "Gesäll", "Sko", "Vind", "Rute", "Torkare", "Blad", "Fiolsfodral" }; - DictionaryCompoundWordTokenFilter tf = new DictionaryCompoundWordTokenFilter( - new WhitespaceTokenizer(new StringReader("Basfiolsfodralmakaregesäll")), + DictionaryCompoundWordTokenFilter tf = new DictionaryCompoundWordTokenFilter(Version.LUCENE_CURRENT, + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("Basfiolsfodralmakaregesäll")), dict, CompoundWordTokenFilterBase.DEFAULT_MIN_WORD_SIZE, CompoundWordTokenFilterBase.DEFAULT_MIN_SUBWORD_SIZE, CompoundWordTokenFilterBase.DEFAULT_MAX_SUBWORD_SIZE, true); @@ -128,9 +129,9 @@ String[] dict = { "Rind", "Fleisch", "Draht", "Schere", "Gesetz", "Aufgabe", "Überwachung" }; - Tokenizer wsTokenizer = new WhitespaceTokenizer(new StringReader( + Tokenizer wsTokenizer = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "Rindfleischüberwachungsgesetz")); - DictionaryCompoundWordTokenFilter tf = new DictionaryCompoundWordTokenFilter( + DictionaryCompoundWordTokenFilter tf = new DictionaryCompoundWordTokenFilter(Version.LUCENE_CURRENT, wsTokenizer, dict, CompoundWordTokenFilterBase.DEFAULT_MIN_WORD_SIZE, CompoundWordTokenFilterBase.DEFAULT_MIN_SUBWORD_SIZE, Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cz/TestCzechStemmer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cz/TestCzechStemmer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cz/TestCzechStemmer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/cz/TestCzechStemmer.java Sun Jan 31 16:01:17 2010 @@ -279,7 +279,7 @@ CharArraySet set = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set.add("hole"); CzechStemFilter filter = new CzechStemFilter(new KeywordMarkerTokenFilter( - new WhitespaceTokenizer(new StringReader("hole desek")), set)); + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("hole desek")), set)); assertTokenStreamContents(filter, new String[] { "hole", "desk" }); } Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/de/TestGermanStemFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/de/TestGermanStemFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/de/TestGermanStemFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/de/TestGermanStemFilter.java Sun Jan 31 16:01:17 2010 @@ -70,7 +70,7 @@ } public void testExclusionTableBWCompat() throws IOException { - GermanStemFilter filter = new GermanStemFilter(new LowerCaseTokenizer( + GermanStemFilter filter = new GermanStemFilter(new LowerCaseTokenizer(Version.LUCENE_CURRENT, new StringReader("Fischen Trinken"))); CharArraySet set = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set.add("fischen"); @@ -82,7 +82,7 @@ CharArraySet set = new CharArraySet(Version.LUCENE_CURRENT, 1, true); set.add("fischen"); GermanStemFilter filter = new GermanStemFilter( - new KeywordMarkerTokenFilter(new LowerCaseTokenizer(new StringReader( + new KeywordMarkerTokenFilter(new LowerCaseTokenizer(Version.LUCENE_CURRENT, new StringReader( "Fischen Trinken")), set)); assertTokenStreamContents(filter, new String[] { "fischen", "trink" }); } @@ -94,7 +94,7 @@ set1.add("trinken"); set1.add("fischen"); GermanStemFilter filter = new GermanStemFilter( - new KeywordMarkerTokenFilter(new LowerCaseTokenizer(new StringReader( + new KeywordMarkerTokenFilter(new LowerCaseTokenizer(Version.LUCENE_CURRENT, new StringReader( "Fischen Trinken")), set)); filter.setExclusionSet(set1); assertTokenStreamContents(filter, new String[] { "fischen", "trinken" }); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/fa/TestPersianNormalizationFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/fa/TestPersianNormalizationFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/fa/TestPersianNormalizationFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/fa/TestPersianNormalizationFilter.java Sun Jan 31 16:01:17 2010 @@ -22,6 +22,7 @@ import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.ar.ArabicLetterTokenizer; +import org.apache.lucene.util.Version; /** * Test the Persian Normalization Filter @@ -54,7 +55,7 @@ } private void check(final String input, final String expected) throws IOException { - ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer( + ArabicLetterTokenizer tokenStream = new ArabicLetterTokenizer(Version.LUCENE_CURRENT, new StringReader(input)); PersianNormalizationFilter filter = new PersianNormalizationFilter( tokenStream); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAndSuffixAwareTokenFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAndSuffixAwareTokenFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAndSuffixAwareTokenFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAndSuffixAwareTokenFilter.java Sun Jan 31 16:01:17 2010 @@ -20,6 +20,7 @@ import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.Token; import org.apache.lucene.analysis.WhitespaceTokenizer; +import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; @@ -30,7 +31,7 @@ PrefixAndSuffixAwareTokenFilter ts = new PrefixAndSuffixAwareTokenFilter( new SingleTokenTokenStream(createToken("^", 0, 0)), - new WhitespaceTokenizer(new StringReader("hello world")), + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("hello world")), new SingleTokenTokenStream(createToken("$", 0, 0))); assertTokenStreamContents(ts, Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAwareTokenFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAwareTokenFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAwareTokenFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/miscellaneous/TestPrefixAwareTokenFilter.java Sun Jan 31 16:01:17 2010 @@ -20,6 +20,7 @@ import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.Token; import org.apache.lucene.analysis.WhitespaceTokenizer; +import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; @@ -40,7 +41,8 @@ // prefix and suffix using 2x prefix - ts = new PrefixAwareTokenFilter(new SingleTokenTokenStream(createToken("^", 0, 0)), new WhitespaceTokenizer(new StringReader("hello world"))); + ts = new PrefixAwareTokenFilter(new SingleTokenTokenStream(createToken("^", 0, 0)), + new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("hello world"))); ts = new PrefixAwareTokenFilter(ts, new SingleTokenTokenStream(createToken("$", 0, 0))); assertTokenStreamContents(ts, Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/EdgeNGramTokenFilterTest.java Sun Jan 31 16:01:17 2010 @@ -20,6 +20,7 @@ import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.BaseTokenStreamTestCase; +import org.apache.lucene.util.Version; import java.io.StringReader; @@ -32,7 +33,7 @@ @Override public void setUp() throws Exception { super.setUp(); - input = new WhitespaceTokenizer(new StringReader("abcde")); + input = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("abcde")); } public void testInvalidInput() throws Exception { @@ -91,13 +92,13 @@ } public void testSmallTokenInStream() throws Exception { - input = new WhitespaceTokenizer(new StringReader("abc de fgh")); + input = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("abc de fgh")); EdgeNGramTokenFilter tokenizer = new EdgeNGramTokenFilter(input, EdgeNGramTokenFilter.Side.FRONT, 3, 3); assertTokenStreamContents(tokenizer, new String[]{"abc","fgh"}, new int[]{0,0}, new int[]{3,3}); } public void testReset() throws Exception { - WhitespaceTokenizer tokenizer = new WhitespaceTokenizer(new StringReader("abcde")); + WhitespaceTokenizer tokenizer = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("abcde")); EdgeNGramTokenFilter filter = new EdgeNGramTokenFilter(tokenizer, EdgeNGramTokenFilter.Side.FRONT, 1, 3); assertTokenStreamContents(filter, new String[]{"a","ab","abc"}, new int[]{0,0,0}, new int[]{1,2,3}); tokenizer.reset(new StringReader("abcde")); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/NGramTokenFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/NGramTokenFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/NGramTokenFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ngram/NGramTokenFilterTest.java Sun Jan 31 16:01:17 2010 @@ -20,6 +20,7 @@ import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.BaseTokenStreamTestCase; +import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; @@ -33,7 +34,7 @@ @Override public void setUp() throws Exception { super.setUp(); - input = new WhitespaceTokenizer(new StringReader("abcde")); + input = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("abcde")); } public void testInvalidInput() throws Exception { @@ -81,13 +82,13 @@ } public void testSmallTokenInStream() throws Exception { - input = new WhitespaceTokenizer(new StringReader("abc de fgh")); + input = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("abc de fgh")); NGramTokenFilter filter = new NGramTokenFilter(input, 3, 3); assertTokenStreamContents(filter, new String[]{"abc","fgh"}, new int[]{0,0}, new int[]{3,3}); } public void testReset() throws Exception { - WhitespaceTokenizer tokenizer = new WhitespaceTokenizer(new StringReader("abcde")); + WhitespaceTokenizer tokenizer = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("abcde")); NGramTokenFilter filter = new NGramTokenFilter(tokenizer, 1, 1); assertTokenStreamContents(filter, new String[]{"a","b","c","d","e"}, new int[]{0,1,2,3,4}, new int[]{1,2,3,4,5}); tokenizer.reset(new StringReader("abcde")); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/DelimitedPayloadTokenFilterTest.java Sun Jan 31 16:01:17 2010 @@ -22,6 +22,7 @@ import org.apache.lucene.analysis.tokenattributes.TermAttribute; import org.apache.lucene.index.Payload; import org.apache.lucene.util.LuceneTestCase; +import org.apache.lucene.util.Version; import java.io.StringReader; @@ -35,7 +36,7 @@ public void testPayloads() throws Exception { String test = "The quick|JJ red|JJ fox|NN jumped|VB over the lazy|JJ brown|JJ dogs|NN"; DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter - (new WhitespaceTokenizer(new StringReader(test)), + (new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test)), DelimitedPayloadTokenFilter.DEFAULT_DELIMITER, new IdentityEncoder()); TermAttribute termAtt = filter.getAttribute(TermAttribute.class); PayloadAttribute payAtt = filter.getAttribute(PayloadAttribute.class); @@ -56,7 +57,7 @@ String test = "The quick|JJ red|JJ fox|NN jumped|VB over the lazy|JJ brown|JJ dogs|NN"; DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter - (new WhitespaceTokenizer(new StringReader(test)), + (new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test)), DelimitedPayloadTokenFilter.DEFAULT_DELIMITER, new IdentityEncoder()); assertTermEquals("The", filter, null); assertTermEquals("quick", filter, "JJ".getBytes("UTF-8")); @@ -74,7 +75,7 @@ public void testFloatEncoding() throws Exception { String test = "The quick|1.0 red|2.0 fox|3.5 jumped|0.5 over the lazy|5 brown|99.3 dogs|83.7"; - DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter(new WhitespaceTokenizer(new StringReader(test)), '|', new FloatEncoder()); + DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test)), '|', new FloatEncoder()); TermAttribute termAtt = filter.getAttribute(TermAttribute.class); PayloadAttribute payAtt = filter.getAttribute(PayloadAttribute.class); assertTermEquals("The", filter, termAtt, payAtt, null); @@ -92,7 +93,7 @@ public void testIntEncoding() throws Exception { String test = "The quick|1 red|2 fox|3 jumped over the lazy|5 brown|99 dogs|83"; - DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter(new WhitespaceTokenizer(new StringReader(test)), '|', new IntegerEncoder()); + DelimitedPayloadTokenFilter filter = new DelimitedPayloadTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test)), '|', new IntegerEncoder()); TermAttribute termAtt = filter.getAttribute(TermAttribute.class); PayloadAttribute payAtt = filter.getAttribute(PayloadAttribute.class); assertTermEquals("The", filter, termAtt, payAtt, null); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/NumericPayloadTokenFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/NumericPayloadTokenFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/NumericPayloadTokenFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/NumericPayloadTokenFilterTest.java Sun Jan 31 16:01:17 2010 @@ -23,6 +23,7 @@ import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; import org.apache.lucene.analysis.tokenattributes.TermAttribute; import org.apache.lucene.analysis.tokenattributes.TypeAttribute; +import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; @@ -37,7 +38,7 @@ public void test() throws IOException { String test = "The quick red fox jumped over the lazy brown dogs"; - NumericPayloadTokenFilter nptf = new NumericPayloadTokenFilter(new WordTokenFilter(new WhitespaceTokenizer(new StringReader(test))), 3, "D"); + NumericPayloadTokenFilter nptf = new NumericPayloadTokenFilter(new WordTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test))), 3, "D"); boolean seenDogs = false; TermAttribute termAtt = nptf.getAttribute(TermAttribute.class); TypeAttribute typeAtt = nptf.getAttribute(TypeAttribute.class); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TokenOffsetPayloadTokenFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TokenOffsetPayloadTokenFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TokenOffsetPayloadTokenFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TokenOffsetPayloadTokenFilterTest.java Sun Jan 31 16:01:17 2010 @@ -21,6 +21,7 @@ import org.apache.lucene.analysis.tokenattributes.OffsetAttribute; import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; import org.apache.lucene.index.Payload; +import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; @@ -35,7 +36,7 @@ public void test() throws IOException { String test = "The quick red fox jumped over the lazy brown dogs"; - TokenOffsetPayloadTokenFilter nptf = new TokenOffsetPayloadTokenFilter(new WhitespaceTokenizer(new StringReader(test))); + TokenOffsetPayloadTokenFilter nptf = new TokenOffsetPayloadTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test))); int count = 0; PayloadAttribute payloadAtt = nptf.getAttribute(PayloadAttribute.class); OffsetAttribute offsetAtt = nptf.getAttribute(OffsetAttribute.class); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TypeAsPayloadTokenFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TypeAsPayloadTokenFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TypeAsPayloadTokenFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/payloads/TypeAsPayloadTokenFilterTest.java Sun Jan 31 16:01:17 2010 @@ -23,6 +23,7 @@ import org.apache.lucene.analysis.tokenattributes.PayloadAttribute; import org.apache.lucene.analysis.tokenattributes.TermAttribute; import org.apache.lucene.analysis.tokenattributes.TypeAttribute; +import org.apache.lucene.util.Version; import java.io.IOException; import java.io.StringReader; @@ -37,7 +38,7 @@ public void test() throws IOException { String test = "The quick red fox jumped over the lazy brown dogs"; - TypeAsPayloadTokenFilter nptf = new TypeAsPayloadTokenFilter(new WordTokenFilter(new WhitespaceTokenizer(new StringReader(test)))); + TypeAsPayloadTokenFilter nptf = new TypeAsPayloadTokenFilter(new WordTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test)))); int count = 0; TermAttribute termAtt = nptf.getAttribute(TermAttribute.class); TypeAttribute typeAtt = nptf.getAttribute(TypeAttribute.class); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/query/QueryAutoStopWordAnalyzerTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/query/QueryAutoStopWordAnalyzerTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/query/QueryAutoStopWordAnalyzerTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/query/QueryAutoStopWordAnalyzerTest.java Sun Jan 31 16:01:17 2010 @@ -51,7 +51,7 @@ protected void setUp() throws Exception { super.setUp(); dir = new RAMDirectory(); - appAnalyzer = new WhitespaceAnalyzer(); + appAnalyzer = new WhitespaceAnalyzer(Version.LUCENE_CURRENT); IndexWriter writer = new IndexWriter(dir, appAnalyzer, true, IndexWriter.MaxFieldLength.UNLIMITED); int numDocs = 200; for (int i = 0; i < numDocs; i++) { @@ -157,9 +157,9 @@ @Override public TokenStream tokenStream(String fieldName, Reader reader) { if (++invocationCount % 2 == 0) - return new WhitespaceTokenizer(reader); + return new WhitespaceTokenizer(Version.LUCENE_CURRENT, reader); else - return new LetterTokenizer(reader); + return new LetterTokenizer(Version.LUCENE_CURRENT, reader); } } @@ -173,7 +173,7 @@ } public void testTokenStream() throws Exception { - QueryAutoStopWordAnalyzer a = new QueryAutoStopWordAnalyzer(Version.LUCENE_CURRENT, new WhitespaceAnalyzer()); + QueryAutoStopWordAnalyzer a = new QueryAutoStopWordAnalyzer(Version.LUCENE_CURRENT, new WhitespaceAnalyzer(Version.LUCENE_CURRENT)); a.addStopWords(reader, 10); TokenStream ts = a.tokenStream("repetitiveField", new StringReader("this boring")); TermAttribute termAtt = ts.getAttribute(TermAttribute.class); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/reverse/TestReverseStringFilter.java Sun Jan 31 16:01:17 2010 @@ -27,9 +27,9 @@ public class TestReverseStringFilter extends BaseTokenStreamTestCase { public void testFilter() throws Exception { - TokenStream stream = new WhitespaceTokenizer( + TokenStream stream = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("Do have a nice day")); // 1-4 length string - ReverseStringFilter filter = new ReverseStringFilter(stream); + ReverseStringFilter filter = new ReverseStringFilter(Version.LUCENE_CURRENT, stream); TermAttribute text = filter.getAttribute(TermAttribute.class); assertTrue(filter.incrementToken()); assertEquals("oD", text.term()); @@ -45,9 +45,9 @@ } public void testFilterWithMark() throws Exception { - TokenStream stream = new WhitespaceTokenizer(new StringReader( + TokenStream stream = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "Do have a nice day")); // 1-4 length string - ReverseStringFilter filter = new ReverseStringFilter(stream, '\u0001'); + ReverseStringFilter filter = new ReverseStringFilter(Version.LUCENE_CURRENT, stream, '\u0001'); TermAttribute text = filter .getAttribute(TermAttribute.class); assertTrue(filter.incrementToken()); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianAnalyzer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianAnalyzer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianAnalyzer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianAnalyzer.java Sun Jan 31 16:01:17 2010 @@ -66,7 +66,7 @@ TokenStream in = ra.tokenStream("all", inWords); RussianLetterTokenizer sample = - new RussianLetterTokenizer( + new RussianLetterTokenizer(Version.LUCENE_CURRENT, sampleUnicode); TermAttribute text = in.getAttribute(TermAttribute.class); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianLetterTokenizer.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianLetterTokenizer.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianLetterTokenizer.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/ru/TestRussianLetterTokenizer.java Sun Jan 31 16:01:17 2010 @@ -30,7 +30,7 @@ public void testRussianLetterTokenizer() throws IOException { StringReader reader = new StringReader("1234567890 Вместе \ud801\udc1ctest"); - RussianLetterTokenizer tokenizer = new RussianLetterTokenizer(Version.LUCENE_31, + RussianLetterTokenizer tokenizer = new RussianLetterTokenizer(Version.LUCENE_CURRENT, reader); assertTokenStreamContents(tokenizer, new String[] {"1234567890", "Вместе", "\ud801\udc1ctest"}); Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapperTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapperTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapperTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleAnalyzerWrapperTest.java Sun Jan 31 16:01:17 2010 @@ -106,7 +106,7 @@ */ public void testShingleAnalyzerWrapperQueryParsing() throws Exception { ScoreDoc[] hits = queryParsingTest(new ShingleAnalyzerWrapper - (new WhitespaceAnalyzer(), 2), + (new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2), "test sentence"); int[] ranks = new int[] { 1, 2, 0 }; compareRanks(hits, ranks); @@ -117,7 +117,7 @@ */ public void testShingleAnalyzerWrapperPhraseQueryParsingFails() throws Exception { ScoreDoc[] hits = queryParsingTest(new ShingleAnalyzerWrapper - (new WhitespaceAnalyzer(), 2), + (new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2), "\"this sentence\""); int[] ranks = new int[] { 0 }; compareRanks(hits, ranks); @@ -128,7 +128,7 @@ */ public void testShingleAnalyzerWrapperPhraseQueryParsing() throws Exception { ScoreDoc[] hits = queryParsingTest(new ShingleAnalyzerWrapper - (new WhitespaceAnalyzer(), 2), + (new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2), "\"test sentence\""); int[] ranks = new int[] { 1 }; compareRanks(hits, ranks); @@ -139,7 +139,7 @@ */ public void testShingleAnalyzerWrapperRequiredQueryParsing() throws Exception { ScoreDoc[] hits = queryParsingTest(new ShingleAnalyzerWrapper - (new WhitespaceAnalyzer(), 2), + (new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2), "+test +sentence"); int[] ranks = new int[] { 1, 2 }; compareRanks(hits, ranks); @@ -149,7 +149,7 @@ * This shows how to construct a phrase query containing shingles. */ public void testShingleAnalyzerWrapperPhraseQuery() throws Exception { - Analyzer analyzer = new ShingleAnalyzerWrapper(new WhitespaceAnalyzer(), 2); + Analyzer analyzer = new ShingleAnalyzerWrapper(new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2); searcher = setUpSearcher(analyzer); PhraseQuery q = new PhraseQuery(); @@ -178,7 +178,7 @@ * in the right order and adjacent to each other. */ public void testShingleAnalyzerWrapperBooleanQuery() throws Exception { - Analyzer analyzer = new ShingleAnalyzerWrapper(new WhitespaceAnalyzer(), 2); + Analyzer analyzer = new ShingleAnalyzerWrapper(new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2); searcher = setUpSearcher(analyzer); BooleanQuery q = new BooleanQuery(); @@ -200,7 +200,7 @@ } public void testReusableTokenStream() throws Exception { - Analyzer a = new ShingleAnalyzerWrapper(new WhitespaceAnalyzer(), 2); + Analyzer a = new ShingleAnalyzerWrapper(new WhitespaceAnalyzer(Version.LUCENE_CURRENT), 2); assertAnalyzesToReuse(a, "please divide into shingles", new String[] { "please", "please divide", "divide", "divide into", "into", "into shingles", "shingles" }, new int[] { 0, 0, 7, 7, 14, 14, 19 }, @@ -222,9 +222,9 @@ @Override public TokenStream tokenStream(String fieldName, Reader reader) { if (++invocationCount % 2 == 0) - return new WhitespaceTokenizer(reader); + return new WhitespaceTokenizer(Version.LUCENE_CURRENT, reader); else - return new LetterTokenizer(reader); + return new LetterTokenizer(Version.LUCENE_CURRENT, reader); } } Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/ShingleFilterTest.java Sun Jan 31 16:01:17 2010 @@ -26,6 +26,7 @@ import org.apache.lucene.analysis.Tokenizer; import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.tokenattributes.*; +import org.apache.lucene.util.Version; public class ShingleFilterTest extends BaseTokenStreamTestCase { @@ -835,7 +836,7 @@ public void testReset() throws Exception { - Tokenizer wsTokenizer = new WhitespaceTokenizer(new StringReader("please divide this sentence")); + Tokenizer wsTokenizer = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("please divide this sentence")); TokenStream filter = new ShingleFilter(wsTokenizer, 2); assertTokenStreamContents(filter, new String[]{"please","please divide","divide","divide this","this","this sentence","sentence"}, Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/TestShingleMatrixFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/TestShingleMatrixFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/TestShingleMatrixFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/shingle/TestShingleMatrixFilter.java Sun Jan 31 16:01:17 2010 @@ -34,6 +34,7 @@ import org.apache.lucene.analysis.shingle.ShingleMatrixFilter.Matrix; import org.apache.lucene.analysis.shingle.ShingleMatrixFilter.Matrix.Column; import org.apache.lucene.analysis.tokenattributes.*; +import org.apache.lucene.util.Version; public class TestShingleMatrixFilter extends BaseTokenStreamTestCase { @@ -43,7 +44,7 @@ public void testIterator() throws IOException { - WhitespaceTokenizer wst = new WhitespaceTokenizer(new StringReader("one two three four five")); + WhitespaceTokenizer wst = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader("one two three four five")); ShingleMatrixFilter smf = new ShingleMatrixFilter(wst, 2, 2, '_', false, new ShingleMatrixFilter.OneDimensionalNonWeightedTokenSettingsCodec()); int i; Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/DateRecognizerSinkTokenizerTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/DateRecognizerSinkTokenizerTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/DateRecognizerSinkTokenizerTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/DateRecognizerSinkTokenizerTest.java Sun Jan 31 16:01:17 2010 @@ -25,6 +25,7 @@ import org.apache.lucene.analysis.TeeSinkTokenFilter; import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.TeeSinkTokenFilter.SinkTokenStream; +import org.apache.lucene.util.Version; public class DateRecognizerSinkTokenizerTest extends BaseTokenStreamTestCase { @@ -36,7 +37,7 @@ public void test() throws IOException { DateRecognizerSinkFilter sinkFilter = new DateRecognizerSinkFilter(new SimpleDateFormat("MM/dd/yyyy", Locale.US)); String test = "The quick red fox jumped over the lazy brown dogs on 7/11/2006 The dogs finally reacted on 7/12/2006"; - TeeSinkTokenFilter tee = new TeeSinkTokenFilter(new WhitespaceTokenizer(new StringReader(test))); + TeeSinkTokenFilter tee = new TeeSinkTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test))); SinkTokenStream sink = tee.newSinkTokenStream(sinkFilter); int count = 0; Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenRangeSinkTokenizerTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenRangeSinkTokenizerTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenRangeSinkTokenizerTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenRangeSinkTokenizerTest.java Sun Jan 31 16:01:17 2010 @@ -23,6 +23,7 @@ import org.apache.lucene.analysis.TeeSinkTokenFilter; import org.apache.lucene.analysis.WhitespaceTokenizer; import org.apache.lucene.analysis.TeeSinkTokenFilter.SinkTokenStream; +import org.apache.lucene.util.Version; public class TokenRangeSinkTokenizerTest extends BaseTokenStreamTestCase { @@ -34,7 +35,7 @@ public void test() throws IOException { TokenRangeSinkFilter sinkFilter = new TokenRangeSinkFilter(2, 4); String test = "The quick red fox jumped over the lazy brown dogs"; - TeeSinkTokenFilter tee = new TeeSinkTokenFilter(new WhitespaceTokenizer(new StringReader(test))); + TeeSinkTokenFilter tee = new TeeSinkTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test))); SinkTokenStream rangeToks = tee.newSinkTokenStream(sinkFilter); int count = 0; Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenTypeSinkTokenizerTest.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenTypeSinkTokenizerTest.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenTypeSinkTokenizerTest.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/sinks/TokenTypeSinkTokenizerTest.java Sun Jan 31 16:01:17 2010 @@ -27,6 +27,7 @@ import org.apache.lucene.analysis.TeeSinkTokenFilter.SinkTokenStream; import org.apache.lucene.analysis.tokenattributes.TermAttribute; import org.apache.lucene.analysis.tokenattributes.TypeAttribute; +import org.apache.lucene.util.Version; public class TokenTypeSinkTokenizerTest extends BaseTokenStreamTestCase { @@ -39,7 +40,7 @@ TokenTypeSinkFilter sinkFilter = new TokenTypeSinkFilter("D"); String test = "The quick red fox jumped over the lazy brown dogs"; - TeeSinkTokenFilter ttf = new TeeSinkTokenFilter(new WordTokenFilter(new WhitespaceTokenizer(new StringReader(test)))); + TeeSinkTokenFilter ttf = new TeeSinkTokenFilter(new WordTokenFilter(new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader(test)))); SinkTokenStream sink = ttf.newSinkTokenStream(sinkFilter); boolean seenDogs = false; Modified: lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/tr/TestTurkishLowerCaseFilter.java URL: http://svn.apache.org/viewvc/lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/tr/TestTurkishLowerCaseFilter.java?rev=905065&r1=905064&r2=905065&view=diff ============================================================================== --- lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/tr/TestTurkishLowerCaseFilter.java (original) +++ lucene/java/trunk/contrib/analyzers/common/src/test/org/apache/lucene/analysis/tr/TestTurkishLowerCaseFilter.java Sun Jan 31 16:01:17 2010 @@ -22,6 +22,7 @@ import org.apache.lucene.analysis.BaseTokenStreamTestCase; import org.apache.lucene.analysis.TokenStream; import org.apache.lucene.analysis.WhitespaceTokenizer; +import org.apache.lucene.util.Version; /** * Test the Turkish lowercase filter. @@ -32,7 +33,7 @@ * Test composed forms */ public void testTurkishLowerCaseFilter() throws Exception { - TokenStream stream = new WhitespaceTokenizer(new StringReader( + TokenStream stream = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "\u0130STANBUL \u0130ZM\u0130R ISPARTA")); TurkishLowerCaseFilter filter = new TurkishLowerCaseFilter(stream); assertTokenStreamContents(filter, new String[] {"istanbul", "izmir", @@ -43,7 +44,7 @@ * Test decomposed forms */ public void testDecomposed() throws Exception { - TokenStream stream = new WhitespaceTokenizer(new StringReader( + TokenStream stream = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "\u0049\u0307STANBUL \u0049\u0307ZM\u0049\u0307R ISPARTA")); TurkishLowerCaseFilter filter = new TurkishLowerCaseFilter(stream); assertTokenStreamContents(filter, new String[] {"istanbul", "izmir", @@ -56,7 +57,7 @@ * to U+0130 + U+0316, and is lowercased the same way. */ public void testDecomposed2() throws Exception { - TokenStream stream = new WhitespaceTokenizer(new StringReader( + TokenStream stream = new WhitespaceTokenizer(Version.LUCENE_CURRENT, new StringReader( "\u0049\u0316\u0307STANBUL \u0049\u0307ZM\u0049\u0307R I\u0316SPARTA")); TurkishLowerCaseFilter filter = new TurkishLowerCaseFilter(stream); assertTokenStreamContents(filter, new String[] {"i\u0316stanbul", "izmir",