Return-Path: X-Original-To: apmail-lucenenet-commits-archive@www.apache.org Delivered-To: apmail-lucenenet-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 1552617C7B for ; Sun, 9 Nov 2014 03:32:27 +0000 (UTC) Received: (qmail 16348 invoked by uid 500); 9 Nov 2014 03:32:27 -0000 Delivered-To: apmail-lucenenet-commits-archive@lucenenet.apache.org Received: (qmail 16305 invoked by uid 500); 9 Nov 2014 03:32:27 -0000 Mailing-List: contact commits-help@lucenenet.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: lucene-net-dev@lucenenet.apache.org Delivered-To: mailing list commits@lucenenet.apache.org Received: (qmail 16294 invoked by uid 99); 9 Nov 2014 03:32:27 -0000 Received: from tyr.zones.apache.org (HELO tyr.zones.apache.org) (140.211.11.114) by apache.org (qpsmtpd/0.29) with ESMTP; Sun, 09 Nov 2014 03:32:27 +0000 Received: by tyr.zones.apache.org (Postfix, from userid 65534) id 2D646934832; Sun, 9 Nov 2014 03:32:27 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit From: synhershko@apache.org To: commits@lucenenet.apache.org Message-Id: <5f8e4133c1c34af78eae1b17018ac1bd@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: lucenenet git commit: More porting fixes Date: Sun, 9 Nov 2014 03:32:27 +0000 (UTC) Repository: lucenenet Updated Branches: refs/heads/master bcff24dd1 -> 6f6f938ef More porting fixes Project: http://git-wip-us.apache.org/repos/asf/lucenenet/repo Commit: http://git-wip-us.apache.org/repos/asf/lucenenet/commit/6f6f938e Tree: http://git-wip-us.apache.org/repos/asf/lucenenet/tree/6f6f938e Diff: http://git-wip-us.apache.org/repos/asf/lucenenet/diff/6f6f938e Branch: refs/heads/master Commit: 6f6f938ef0ba0cce5079a22782926a9f855f023d Parents: bcff24d Author: Itamar Syn-Hershko Authored: Sun Nov 9 05:31:49 2014 +0200 Committer: Itamar Syn-Hershko Committed: Sun Nov 9 05:32:16 2014 +0200 ---------------------------------------------------------------------- .../CharFilter/MappingCharFilterFactory.cs | 20 +++--- .../Analysis/CommonGrams/CommonGramsFilter.cs | 66 ++++++++------------ .../CommonGrams/CommonGramsFilterFactory.cs | 4 +- .../CommonGrams/CommonGramsQueryFilter.cs | 26 +++----- .../Analysis/Core/KeywordAnalyzer.cs | 2 +- .../Analysis/Core/WhitespaceAnalyzer.cs | 2 +- .../Analysis/Util/FilteringTokenFilter.cs | 4 +- .../Collation/CollationKeyAnalyzer.cs | 19 +++--- src/Lucene.Net.Core/Analysis/Analyzer.cs | 2 +- src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs | 2 +- .../Analysis/CachingTokenFilter.cs | 4 +- src/Lucene.Net.Core/Analysis/TokenFilter.cs | 10 +-- .../Suggest/Analyzing/SuggestStopFilter.cs | 6 +- .../Analysis/LookaheadTokenFilter.cs | 2 +- .../Analysis/MockAnalyzer.cs | 2 +- .../Analysis/MockBytesAnalyzer.cs | 2 +- .../Analysis/MockFixedLengthPayloadFilter.cs | 2 +- .../Analysis/MockHoleInjectingTokenFilter.cs | 2 +- .../Analysis/MockPayloadAnalyzer.cs | 4 +- .../Analysis/MockTokenFilter.cs | 2 +- .../Analysis/MockTokenizer.cs | 6 +- .../Analysis/MockVariableLengthPayloadFilter.cs | 2 +- .../Analysis/ValidatingTokenFilter.cs | 2 +- .../core/Analysis/TestGraphTokenizers.cs | 18 +++--- .../core/Analysis/TestLookaheadTokenFilter.cs | 6 +- .../core/Analysis/TestMockAnalyzer.cs | 4 +- .../core/Analysis/TestMockCharFilter.cs | 2 +- .../Codecs/Lucene41/TestBlockPostingsFormat3.cs | 2 +- .../Index/TestDocInverterPerFieldErrorInfo.cs | 2 +- .../core/Index/TestDocumentWriter.cs | 6 +- .../core/Index/TestIndexWriter.cs | 8 +-- .../core/Index/TestIndexWriterCommit.cs | 4 +- .../core/Index/TestIndexWriterDelete.cs | 2 +- .../core/Index/TestIndexWriterExceptions.cs | 12 ++-- .../core/Index/TestLazyProxSkipping.cs | 2 +- .../core/Index/TestMultiLevelSkipList.cs | 4 +- src/Lucene.Net.Tests/core/Index/TestPayloads.cs | 4 +- .../core/Index/TestTermVectorsReader.cs | 2 +- .../core/Index/TestTermdocPerf.cs | 2 +- .../core/Search/FuzzyTermOnShortTermsTest.cs | 2 +- .../core/Search/Payloads/PayloadHelper.cs | 4 +- .../Search/Payloads/TestPayloadNearQuery.cs | 4 +- .../Search/Payloads/TestPayloadTermQuery.cs | 4 +- .../core/Search/Spans/TestBasics.cs | 4 +- .../core/Search/Spans/TestPayloadSpans.cs | 6 +- .../core/Search/TestPhraseQuery.cs | 2 +- .../core/Search/TestPositionIncrement.cs | 2 +- .../core/Search/TestTermRangeQuery.cs | 2 +- .../core/Util/TestQueryBuilder.cs | 10 +-- 49 files changed, 143 insertions(+), 170 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs index 4489b7c..1d8f008 100644 --- a/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs +++ b/src/Lucene.Net.Analysis.Common/Analysis/CharFilter/MappingCharFilterFactory.cs @@ -1,7 +1,11 @@ using System.Collections.Generic; +using System.IO; using Lucene.Net.Analysis.Util; +using Lucene.Net.Support; +using org.apache.lucene.analysis.charfilter; +using org.apache.lucene.analysis.util; -namespace org.apache.lucene.analysis.charfilter +namespace Lucene.Net.Analysis.CharFilter { /* @@ -20,15 +24,7 @@ namespace org.apache.lucene.analysis.charfilter * See the License for the specific language governing permissions and * limitations under the License. */ - - - using AbstractAnalysisFactory = AbstractAnalysisFactory; - using CharFilterFactory = org.apache.lucene.analysis.util.CharFilterFactory; - using MultiTermAwareComponent = org.apache.lucene.analysis.util.MultiTermAwareComponent; - using ResourceLoader = org.apache.lucene.analysis.util.ResourceLoader; - using ResourceLoaderAware = org.apache.lucene.analysis.util.ResourceLoaderAware; - - /// + /// /// Factory for . ///
 	/// <fieldType name="text_map" class="solr.TextField" positionIncrementGap="100">
@@ -80,8 +76,6 @@ namespace org.apache.lucene.analysis.charfilter
 			  wlist.AddRange(lines);
 			}
 		  }
-//JAVA TO C# CONVERTER WARNING: The original Java variable was marked 'final':
-//ORIGINAL LINE: final NormalizeCharMap.Builder builder = new NormalizeCharMap.Builder();
 		  NormalizeCharMap.Builder builder = new NormalizeCharMap.Builder();
 		  parseRules(wlist, builder);
 		  normMap = builder.build();
@@ -94,7 +88,7 @@ namespace org.apache.lucene.analysis.charfilter
 		}
 	  }
 
-	  public override Reader create(Reader input)
+	  public override TextReader Create(TextReader input)
 	  {
 		// if the map is null, it means there's actually no mappings... just return the original stream
 		// as there is nothing to do here.

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
index 2b97da8..06deccf 100644
--- a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
+++ b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilter.cs
@@ -1,6 +1,4 @@
-using System.Text;
-
-/*
+/*
  * Licensed to the Apache Software Foundation (ASF) under one or more
  * contributor license agreements.  See the NOTICE file distributed with
  * this work for additional information regarding copyright ownership.
@@ -16,19 +14,14 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
+using System.Text;
+using Lucene.Net.Analysis.Tokenattributes;
+using Lucene.Net.Util;
+using org.apache.lucene.analysis.util;
 
-namespace org.apache.lucene.analysis.commongrams
+namespace Lucene.Net.Analysis.CommonGrams
 {
-
-	using OffsetAttribute = org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
-	using PositionIncrementAttribute = org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute;
-	using CharTermAttribute = org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
-	using PositionLengthAttribute = org.apache.lucene.analysis.tokenattributes.PositionLengthAttribute;
-	using TypeAttribute = org.apache.lucene.analysis.tokenattributes.TypeAttribute;
-	using CharArraySet = org.apache.lucene.analysis.util.CharArraySet;
-	using Version = org.apache.lucene.util.Version;
-
-	/*
+    /*
 	 * TODO: Consider implementing https://issues.apache.org/jira/browse/LUCENE-1688 changes to stop list and associated constructors 
 	 */
 
@@ -98,16 +91,14 @@ namespace org.apache.lucene.analysis.commongrams
 	  /// eliminate the middle bigram "of-the"and save a disk seek and a whole set of
 	  /// position lookups.
 	  /// 
-//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET: -//ORIGINAL LINE: @Override public boolean incrementToken() throws java.io.IOException - public override bool incrementToken() + public override bool IncrementToken() { // get the next piece of input if (savedState != null) { - restoreState(savedState); + RestoreState(savedState); savedState = null; - saveTermBuffer(); + SaveTermBuffer(); return true; } else if (!input.incrementToken()) @@ -121,23 +112,21 @@ namespace org.apache.lucene.analysis.commongrams */ if (lastWasCommon || (Common && buffer.Length > 0)) { - savedState = captureState(); - gramToken(); + savedState = CaptureState(); + GramToken(); return true; } - saveTermBuffer(); + SaveTermBuffer(); return true; } /// /// {@inheritDoc} /// -//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET: -//ORIGINAL LINE: @Override public void reset() throws java.io.IOException - public override void reset() + public override void Reset() { - base.reset(); + base.Reset(); lastWasCommon = false; savedState = null; buffer.Length = 0; @@ -153,47 +142,46 @@ namespace org.apache.lucene.analysis.commongrams { get { - return commonWords != null && commonWords.contains(termAttribute.buffer(), 0, termAttribute.length()); + return commonWords != null && commonWords.contains(termAttribute.Buffer(), 0, termAttribute.Length); } } /// /// Saves this information to form the left part of a gram /// - private void saveTermBuffer() + private void SaveTermBuffer() { buffer.Length = 0; - buffer.Append(termAttribute.buffer(), 0, termAttribute.length()); + buffer.Append(termAttribute.Buffer(), 0, termAttribute.Length); buffer.Append(SEPARATOR); - lastStartOffset = offsetAttribute.startOffset(); + lastStartOffset = offsetAttribute.StartOffset(); lastWasCommon = Common; } /// /// Constructs a compound token. /// - private void gramToken() + private void GramToken() { - buffer.Append(termAttribute.buffer(), 0, termAttribute.length()); - int endOffset = offsetAttribute.endOffset(); + buffer.Append(termAttribute.Buffer(), 0, termAttribute.Length); + int endOffset = offsetAttribute.EndOffset(); - clearAttributes(); + ClearAttributes(); int length = buffer.Length; - char[] termText = termAttribute.buffer(); + char[] termText = termAttribute.Buffer(); if (length > termText.Length) { - termText = termAttribute.resizeBuffer(length); + termText = termAttribute.ResizeBuffer(length); } - buffer.getChars(0, length, termText, 0); + buffer.GetChars(0, length, termText, 0); termAttribute.Length = length; posIncAttribute.PositionIncrement = 0; posLenAttribute.PositionLength = 2; // bigram - offsetAttribute.setOffset(lastStartOffset, endOffset); + offsetAttribute.SetOffset(lastStartOffset, endOffset); typeAttribute.Type = GRAM_TYPE; buffer.Length = 0; } } - } \ No newline at end of file http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs index 2233e83..fe531cb 100644 --- a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs +++ b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsFilterFactory.cs @@ -92,9 +92,9 @@ namespace Lucene.Net.Analysis.CommonGrams } } - public override TokenFilter Create(TokenStream input) + public override TokenStream Create(TokenStream input) { - CommonGramsFilter commonGrams = new CommonGramsFilter(luceneMatchVersion, input, commonWords); + var commonGrams = new CommonGramsFilter(luceneMatchVersion, input, commonWords); return commonGrams; } } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs index b787bde..e708697 100644 --- a/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs +++ b/src/Lucene.Net.Analysis.Common/Analysis/CommonGrams/CommonGramsQueryFilter.cs @@ -14,15 +14,11 @@ * See the License for the specific language governing permissions and * limitations under the License. */ -namespace org.apache.lucene.analysis.commongrams -{ - - using PositionIncrementAttribute = org.apache.lucene.analysis.tokenattributes.PositionIncrementAttribute; - using TypeAttribute = org.apache.lucene.analysis.tokenattributes.TypeAttribute; -//JAVA TO C# CONVERTER TODO TASK: This Java 'import static' statement cannot be converted to C#: -// import static org.apache.lucene.analysis.commongrams.CommonGramsFilter.GRAM_TYPE; +using Lucene.Net.Analysis.Tokenattributes; +namespace Lucene.Net.Analysis.CommonGrams +{ /// /// Wrap a CommonGramsFilter optimizing phrase queries by only returning single /// words when they are not a member of a bigram. @@ -62,11 +58,9 @@ namespace org.apache.lucene.analysis.commongrams /// /// {@inheritDoc} /// -//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET: -//ORIGINAL LINE: @Override public void reset() throws java.io.IOException - public override void reset() + public override void Reset() { - base.reset(); + base.Reset(); previous = null; previousType = null; exhausted = false; @@ -80,17 +74,15 @@ namespace org.apache.lucene.analysis.commongrams ///
  • output:"the-rain", "rain-in" ,"in-spain", "falls", "mainly" /// ///
  • -//JAVA TO C# CONVERTER WARNING: Method 'throws' clauses are not available in .NET: -//ORIGINAL LINE: @Override public boolean incrementToken() throws java.io.IOException - public override bool incrementToken() + public override bool IncrementToken() { while (!exhausted && input.incrementToken()) { - State current = captureState(); + State current = CaptureState(); if (previous != null && !GramType) { - restoreState(previous); + RestoreState(previous); previous = current; previousType = typeAttribute.type(); @@ -111,7 +103,7 @@ namespace org.apache.lucene.analysis.commongrams return false; } - restoreState(previous); + RestoreState(previous); previous = null; if (GramType) http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs index 1f1a42b..4d13021 100644 --- a/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs +++ b/src/Lucene.Net.Analysis.Common/Analysis/Core/KeywordAnalyzer.cs @@ -32,7 +32,7 @@ namespace Lucene.Net.Analysis.Core { } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new KeywordTokenizer(reader)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs index 7e77c8d..4ef12c7 100644 --- a/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs +++ b/src/Lucene.Net.Analysis.Common/Analysis/Core/WhitespaceAnalyzer.cs @@ -50,7 +50,7 @@ namespace Lucene.Net.Analysis.Core this.matchVersion = matchVersion; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new WhitespaceTokenizer(matchVersion, reader)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs b/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs index 4d55a25..0feb6b8 100644 --- a/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs +++ b/src/Lucene.Net.Analysis.Common/Analysis/Util/FilteringTokenFilter.cs @@ -81,7 +81,7 @@ namespace Lucene.Net.Analysis.Util if (enablePositionIncrements) { skippedPositions = 0; - while (Input.IncrementToken()) + while (input.IncrementToken()) { if (Accept()) { @@ -96,7 +96,7 @@ namespace Lucene.Net.Analysis.Util } else { - while (Input.IncrementToken()) + while (input.IncrementToken()) { if (Accept()) { http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs index 06fb9e0..0c4df1f 100644 --- a/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs +++ b/src/Lucene.Net.Analysis.Common/Collation/CollationKeyAnalyzer.cs @@ -1,8 +1,10 @@ using System; +using System.IO; +using Lucene.Net.Analysis; using Lucene.Net.Analysis.Core; -using Lucene.Net.Collation; +using org.apache.lucene.collation; -namespace org.apache.lucene.collation +namespace Lucene.Net.Collation { /* @@ -23,13 +25,10 @@ namespace org.apache.lucene.collation */ - using Analyzer = org.apache.lucene.analysis.Analyzer; - using KeywordTokenizer = KeywordTokenizer; - using IndexableBinaryStringTools = org.apache.lucene.util.IndexableBinaryStringTools; // javadoc @link - using Version = org.apache.lucene.util.Version; + // javadoc @link - /// + /// /// /// Configures with . /// @@ -106,14 +105,14 @@ namespace org.apache.lucene.collation /// @deprecated Use /// and specify a version instead. This ctor will be removed in Lucene 5.0 - [Obsolete("Use ")] + [Obsolete("Use ")] public CollationKeyAnalyzer(Collator collator) : this(Version.LUCENE_31, collator) { } - protected internal override TokenStreamComponents createComponents(string fieldName, Reader reader) + public override Analyzer.TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { - if (matchVersion.onOrAfter(Version.LUCENE_40)) + if (matchVersion.OnOrAfter(Version.LUCENE_40)) { KeywordTokenizer tokenizer = new KeywordTokenizer(factory, reader, KeywordTokenizer.DEFAULT_BUFFER_SIZE); return new TokenStreamComponents(tokenizer, tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/Analyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Core/Analysis/Analyzer.cs b/src/Lucene.Net.Core/Analysis/Analyzer.cs index 0ae0c8e..c183885 100644 --- a/src/Lucene.Net.Core/Analysis/Analyzer.cs +++ b/src/Lucene.Net.Core/Analysis/Analyzer.cs @@ -104,7 +104,7 @@ namespace Lucene.Net.Analysis /// /// the reader passed to the constructor /// the for this analyzer. - protected abstract TokenStreamComponents CreateComponents(string fieldName, TextReader reader); + public abstract TokenStreamComponents CreateComponents(string fieldName, TextReader reader); /// /// Returns a TokenStream suitable for fieldName, tokenizing http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs b/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs index f58467b..a5e8660 100644 --- a/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs +++ b/src/Lucene.Net.Core/Analysis/AnalyzerWrapper.cs @@ -95,7 +95,7 @@ namespace Lucene.Net.Analysis return reader; } - protected override sealed TokenStreamComponents CreateComponents(string fieldName, TextReader aReader) + public override sealed TokenStreamComponents CreateComponents(string fieldName, TextReader aReader) { return WrapComponents(fieldName, GetWrappedAnalyzer(fieldName).CreateComponents(fieldName, aReader)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs b/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs index 77994aa..3a1706d 100644 --- a/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs +++ b/src/Lucene.Net.Core/Analysis/CachingTokenFilter.cs @@ -91,12 +91,12 @@ namespace Lucene.Net.Analysis private void FillCache() { - while (Input.IncrementToken()) + while (input.IncrementToken()) { Cache.AddLast(CaptureState()); } // capture final state - Input.End(); + input.End(); FinalState = CaptureState(); } } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Core/Analysis/TokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Core/Analysis/TokenFilter.cs b/src/Lucene.Net.Core/Analysis/TokenFilter.cs index 70f1373..6ed98a5 100644 --- a/src/Lucene.Net.Core/Analysis/TokenFilter.cs +++ b/src/Lucene.Net.Core/Analysis/TokenFilter.cs @@ -26,14 +26,14 @@ namespace Lucene.Net.Analysis { /// /// The source of tokens for this filter. - protected internal readonly TokenStream Input; + protected internal readonly TokenStream input; /// /// Construct a token stream filtering the given input. protected internal TokenFilter(TokenStream input) : base(input) { - this.Input = input; + this.input = input; } /// @@ -45,7 +45,7 @@ namespace Lucene.Net.Analysis /// public override void End() { - Input.End(); + input.End(); } /// @@ -57,7 +57,7 @@ namespace Lucene.Net.Analysis /// public override void Dispose() { - Input.Dispose(); + input.Dispose(); } /// @@ -69,7 +69,7 @@ namespace Lucene.Net.Analysis /// public override void Reset() { - Input.Reset(); + input.Reset(); } } } \ No newline at end of file http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs b/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs index 3f2f72c..1451ac9 100644 --- a/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs +++ b/src/Lucene.Net.Suggest/Suggest/Analyzing/SuggestStopFilter.cs @@ -83,7 +83,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing return false; } - if (!Input.IncrementToken()) + if (!input.IncrementToken()) { return false; } @@ -97,7 +97,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing int endOffset = offsetAtt.EndOffset(); // This token may be a stopword, if it's not end: State sav = CaptureState(); - if (Input.IncrementToken()) + if (input.IncrementToken()) { // It was a stopword; skip it skippedPositions += posInc; @@ -105,7 +105,7 @@ namespace Lucene.Net.Search.Suggest.Analyzing else { ClearAttributes(); - Input.End(); + input.End(); endState = CaptureState(); int finalEndOffset = offsetAtt.EndOffset(); Debug.Assert(finalEndOffset >= endOffset); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs index 0bc4bbe..9b2aee1 100644 --- a/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs +++ b/src/Lucene.Net.TestFramework/Analysis/LookaheadTokenFilter.cs @@ -180,7 +180,7 @@ namespace Lucene.Net.Analysis positions.Get(InputPos).Add(CaptureState()); TokenPending = false; } - bool gotToken = Input.IncrementToken(); + bool gotToken = input.IncrementToken(); if (DEBUG) { Console.WriteLine(" input.incrToken() returned " + gotToken); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs b/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs index 37fe8cd..9a1b883 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockAnalyzer.cs @@ -89,7 +89,7 @@ namespace Lucene.Net.Analysis { } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader, RunAutomaton, LowerCase, MaxTokenLength_Renamed); tokenizer.EnableChecks = EnableChecks_Renamed; http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs b/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs index 12934df..6967b75 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockBytesAnalyzer.cs @@ -26,7 +26,7 @@ namespace Lucene.Net.Analysis { private readonly MockBytesAttributeFactory Factory = new MockBytesAttributeFactory(); - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(Factory, reader, MockTokenizer.KEYWORD, false, MockTokenizer.DEFAULT_MAX_TOKEN_LENGTH); return new TokenStreamComponents(t); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs index 7b510ba..d4a6646 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockFixedLengthPayloadFilter.cs @@ -47,7 +47,7 @@ namespace Lucene.Net.Analysis public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { Random.NextBytes((byte[])(Array)Bytes); PayloadAtt.Payload = Payload; http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs index c5c2577..fceb767 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockHoleInjectingTokenFilter.cs @@ -56,7 +56,7 @@ namespace Lucene.Net.Analysis public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { int posInc = PosIncAtt.PositionIncrement; http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs b/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs index 60ee80e..b348670 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockPayloadAnalyzer.cs @@ -31,7 +31,7 @@ namespace Lucene.Net.Analysis /// public sealed class MockPayloadAnalyzer : Analyzer { - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer result = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true); return new TokenStreamComponents(result, new MockPayloadFilter(result, fieldName)); @@ -66,7 +66,7 @@ namespace Lucene.Net.Analysis public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { PayloadAttr.Payload = new BytesRef(("pos: " + Pos)/*.getBytes(IOUtils.CHARSET_UTF_8)*/); int posIncr; http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs index f8467dc..f692726 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockTokenFilter.cs @@ -69,7 +69,7 @@ namespace Lucene.Net.Analysis // return the first non-stop word found SkippedPositions = 0; - while (Input.IncrementToken()) + while (input.IncrementToken()) { if (!Filter.Run(TermAtt.Buffer(), 0, TermAtt.Length)) { http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs b/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs index 5fa4ae3..7cb9f28 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockTokenizer.cs @@ -247,14 +247,14 @@ namespace Lucene.Net.Analysis { // read(char[]) char[] c = new char[1]; - int ret = Input.Read(c, 0, c.Length); + int ret = input.Read(c, 0, c.Length); return ret == 0 ? -1 : c[0]; } case 1: { // read(char[], int, int) char[] c = new char[2]; - int ret = Input.Read(c, 1, 1); + int ret = input.Read(c, 1, 1); return ret == 0 ? -1 : c[1]; } /* LUCENE TO-DO not sure if needed, CharBuffer not supported @@ -268,7 +268,7 @@ namespace Lucene.Net.Analysis }*/ default: // read() - return Input.Read(); + return input.Read(); } } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs b/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs index 40f2151..07631bf 100644 --- a/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs +++ b/src/Lucene.Net.TestFramework/Analysis/MockVariableLengthPayloadFilter.cs @@ -44,7 +44,7 @@ namespace Lucene.Net.Analysis public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { Random.NextBytes((byte[])(Array)Bytes); Payload.Length = Random.Next(MAXLENGTH); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs b/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs index 3970a3e..bb5cee6 100644 --- a/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs +++ b/src/Lucene.Net.TestFramework/Analysis/ValidatingTokenFilter.cs @@ -99,7 +99,7 @@ namespace Lucene.Net.Analysis public override bool IncrementToken() { - if (!Input.IncrementToken()) + if (!input.IncrementToken()) { return false; } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs b/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs index 4fe19b1..bf85d9a 100644 --- a/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs +++ b/src/Lucene.Net.Tests/core/Analysis/TestGraphTokenizers.cs @@ -204,7 +204,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t2 = new MockGraphTokenFilter(Random(), t); @@ -239,7 +239,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new GraphTokenizer(reader); TokenStream t2 = new MockGraphTokenFilter(Random(), t); @@ -278,7 +278,7 @@ namespace Lucene.Net.Analysis { while (true) { - bool gotOne = Input.IncrementToken(); + bool gotOne = input.IncrementToken(); if (!gotOne) { return false; @@ -328,7 +328,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t2 = new MockGraphTokenFilter(Random(), t); @@ -368,7 +368,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t2 = new RemoveATokens(t); @@ -405,7 +405,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t2 = new MockGraphTokenFilter(Random(), t); @@ -449,7 +449,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t1 = new MockGraphTokenFilter(Random(), t); @@ -486,7 +486,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t1 = new MockGraphTokenFilter(Random(), t); @@ -523,7 +523,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream t1 = new MockHoleInjectingTokenFilter(Random(), t); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs b/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs index db48b98..04622b8 100644 --- a/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs +++ b/src/Lucene.Net.Tests/core/Analysis/TestLookaheadTokenFilter.cs @@ -41,7 +41,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Random random = Random(); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, random.NextBoolean()); @@ -84,7 +84,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, Random().NextBoolean()); TokenStream output = new NeverPeeksLookaheadTokenFilter(tokenizer); @@ -109,7 +109,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TrivialLookaheadFilter filter = new TrivialLookaheadFilter(source); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs b/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs index e39be6a..e185484 100644 --- a/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs +++ b/src/Lucene.Net.Tests/core/Analysis/TestMockAnalyzer.cs @@ -182,7 +182,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false, 5); return new TokenStreamComponents(t, t); @@ -259,7 +259,7 @@ namespace Lucene.Net.Analysis this.Limit = limit; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader, Dfa, Lowercase, Limit); return new TokenStreamComponents(t, t); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs b/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs index cabd3f2..03df580 100644 --- a/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs +++ b/src/Lucene.Net.Tests/core/Analysis/TestMockCharFilter.cs @@ -44,7 +44,7 @@ namespace Lucene.Net.Analysis this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs b/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs index ca67a17..138d941 100644 --- a/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs +++ b/src/Lucene.Net.Tests/core/Codecs/Lucene41/TestBlockPostingsFormat3.cs @@ -150,7 +150,7 @@ namespace Lucene.Net.Codecs.Lucene41 this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader); if (fieldName.Contains("payloadsFixed")) http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs b/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs index e2eb471..8ec9e62 100644 --- a/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs +++ b/src/Lucene.Net.Tests/core/Index/TestDocInverterPerFieldErrorInfo.cs @@ -54,7 +54,7 @@ namespace Lucene.Net.Index private class ThrowingAnalyzer : Analyzer { - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader input) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader input) { Tokenizer tokenizer = new MockTokenizer(input); if (fieldName.Equals("distinctiveFieldName")) http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs b/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs index e66a2fe..997f4ff 100644 --- a/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs +++ b/src/Lucene.Net.Tests/core/Index/TestDocumentWriter.cs @@ -146,7 +146,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, false)); } @@ -195,7 +195,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new TokenFilterAnonymousInnerClassHelper(this, tokenizer)); @@ -230,7 +230,7 @@ namespace Lucene.Net.Index return true; } - bool hasNext = Input.IncrementToken(); + bool hasNext = input.IncrementToken(); if (!hasNext) { return false; http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs index a9801f3..8631da7 100644 --- a/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs +++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriter.cs @@ -1866,7 +1866,7 @@ namespace Lucene.Net.Index internal sealed class StringSplitAnalyzer : Analyzer { - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new StringSplitTokenizer(reader)); } @@ -2226,7 +2226,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { throw new InvalidOperationException("don't invoke me!"); } @@ -2301,7 +2301,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader); TokenStream stream = new MockTokenFilter(tokenizer, MockTokenFilter.ENGLISH_STOPSET); @@ -2346,7 +2346,7 @@ namespace Lucene.Net.Index this.SecondSet = secondSet; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader); TokenStream stream = new MockTokenFilter(tokenizer, MockTokenFilter.ENGLISH_STOPSET); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs index 1a4b4d3..4580112 100644 --- a/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs +++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriterCommit.cs @@ -258,7 +258,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true)); } @@ -276,7 +276,7 @@ namespace Lucene.Net.Index this.Length = length; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true); return new TokenStreamComponents(tokenizer, new MockFixedLengthPayloadFilter(Random(), tokenizer, Length)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs index d5f8513..60dff7e 100644 --- a/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs +++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriterDelete.cs @@ -1192,7 +1192,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs b/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs index 58b706e..6fc21f6 100644 --- a/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs +++ b/src/Lucene.Net.Tests/core/Index/TestIndexWriterExceptions.cs @@ -448,7 +448,7 @@ namespace Lucene.Net.Index { throw new IOException(CRASH_FAIL_MESSAGE); } - return Input.IncrementToken(); + return input.IncrementToken(); } public override void Reset() @@ -519,7 +519,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases. @@ -637,7 +637,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases. @@ -663,7 +663,7 @@ namespace Lucene.Net.Index { throw new IOException(); } - return Input.IncrementToken(); + return input.IncrementToken(); } public override void Reset() @@ -873,7 +873,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases. @@ -972,7 +972,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); tokenizer.EnableChecks = false; // disable workflow checking as we forcefully close() in exceptional cases. http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs b/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs index 59e3e02..4010ec4 100644 --- a/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs +++ b/src/Lucene.Net.Tests/core/Index/TestLazyProxSkipping.cs @@ -121,7 +121,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, true)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs b/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs index 3caf34c..aa94dc8 100644 --- a/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs +++ b/src/Lucene.Net.Tests/core/Index/TestMultiLevelSkipList.cs @@ -127,7 +127,7 @@ namespace Lucene.Net.Index { internal readonly AtomicInteger PayloadCount = new AtomicInteger(-1); - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true); return new TokenStreamComponents(tokenizer, new PayloadFilter(PayloadCount, tokenizer)); @@ -148,7 +148,7 @@ namespace Lucene.Net.Index public override bool IncrementToken() { - bool hasNext = Input.IncrementToken(); + bool hasNext = input.IncrementToken(); if (hasNext) { PayloadAtt.Payload = new BytesRef(new sbyte[] { (sbyte)PayloadCount.IncrementAndGet() }); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestPayloads.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestPayloads.cs b/src/Lucene.Net.Tests/core/Index/TestPayloads.cs index 903225a..dde563c 100644 --- a/src/Lucene.Net.Tests/core/Index/TestPayloads.cs +++ b/src/Lucene.Net.Tests/core/Index/TestPayloads.cs @@ -387,7 +387,7 @@ namespace Lucene.Net.Index FieldToData[field] = new PayloadData(data, offset, length); } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { PayloadData payload; FieldToData.TryGetValue(fieldName, out payload); @@ -436,7 +436,7 @@ namespace Lucene.Net.Index public override bool IncrementToken() { - bool hasNext = Input.IncrementToken(); + bool hasNext = input.IncrementToken(); if (!hasNext) { return false; http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs b/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs index 24deec6..7995318 100644 --- a/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs +++ b/src/Lucene.Net.Tests/core/Index/TestTermVectorsReader.cs @@ -228,7 +228,7 @@ namespace Lucene.Net.Index this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MyTokenizer(OuterInstance, reader)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs b/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs index fe83c63..5160a75 100644 --- a/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs +++ b/src/Lucene.Net.Tests/core/Index/TestTermdocPerf.cs @@ -119,7 +119,7 @@ namespace Lucene.Net.Index this.PercentDocs = percentDocs; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new RepeatingTokenizer(reader, Val, Random, PercentDocs, MaxTF)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs b/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs index 1b9fb84..b668aad 100644 --- a/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs +++ b/src/Lucene.Net.Tests/core/Search/FuzzyTermOnShortTermsTest.cs @@ -89,7 +89,7 @@ namespace Lucene.Net.Search { } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(tokenizer, tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs b/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs index a88b90e..71ea1e1 100644 --- a/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs +++ b/src/Lucene.Net.Tests/core/Search/Payloads/PayloadHelper.cs @@ -63,7 +63,7 @@ namespace Lucene.Net.Search.Payloads this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(result, new PayloadFilter(OuterInstance, result, fieldName)); @@ -88,7 +88,7 @@ namespace Lucene.Net.Search.Payloads public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { if (FieldName.Equals(FIELD)) { http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs index 9df0229..82ef6d1 100644 --- a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs +++ b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadNearQuery.cs @@ -51,7 +51,7 @@ namespace Lucene.Net.Search.Payloads private class PayloadAnalyzer : Analyzer { - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(result, new PayloadFilter(result, fieldName)); @@ -74,7 +74,7 @@ namespace Lucene.Net.Search.Payloads public override bool IncrementToken() { bool result = false; - if (Input.IncrementToken()) + if (input.IncrementToken()) { if (NumSeen % 2 == 0) { http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs index 2f0ced7..8f49550 100644 --- a/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs +++ b/src/Lucene.Net.Tests/core/Search/Payloads/TestPayloadTermQuery.cs @@ -58,7 +58,7 @@ namespace Lucene.Net.Search.Payloads { } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(result, new PayloadFilter(result, fieldName)); @@ -81,7 +81,7 @@ namespace Lucene.Net.Search.Payloads public override bool IncrementToken() { - bool hasNext = Input.IncrementToken(); + bool hasNext = input.IncrementToken(); if (hasNext) { if (FieldName.Equals("field")) http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs b/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs index 0aef0b3..b013b58 100644 --- a/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs +++ b/src/Lucene.Net.Tests/core/Search/Spans/TestBasics.cs @@ -72,7 +72,7 @@ namespace Lucene.Net.Search.Spans public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { PayloadAttr.Payload = new BytesRef(("pos: " + Pos).GetBytes(IOUtils.CHARSET_UTF_8)); Pos++; @@ -118,7 +118,7 @@ namespace Lucene.Net.Search.Spans { } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(tokenizer, new SimplePayloadFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs b/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs index 45ce22a..6017223 100644 --- a/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs +++ b/src/Lucene.Net.Tests/core/Search/Spans/TestPayloadSpans.cs @@ -510,7 +510,7 @@ namespace Lucene.Net.Search.Spans this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(result, new PayloadFilter(OuterInstance, result)); @@ -544,7 +544,7 @@ namespace Lucene.Net.Search.Spans public override bool IncrementToken() { - if (Input.IncrementToken()) + if (input.IncrementToken()) { string token = TermAtt.ToString(); @@ -581,7 +581,7 @@ namespace Lucene.Net.Search.Spans this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer result = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); return new TokenStreamComponents(result, new PayloadFilter(OuterInstance, result)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs b/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs index 2e1ac1d..de56994 100644 --- a/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs +++ b/src/Lucene.Net.Tests/core/Search/TestPhraseQuery.cs @@ -90,7 +90,7 @@ namespace Lucene.Net.Search { } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MockTokenizer(reader, MockTokenizer.WHITESPACE, false)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs b/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs index 83b0970..982c244 100644 --- a/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs +++ b/src/Lucene.Net.Tests/core/Search/TestPositionIncrement.cs @@ -173,7 +173,7 @@ namespace Lucene.Net.Search this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new TokenizerAnonymousInnerClassHelper(this, reader)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs b/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs index e6541b1..54054e0 100644 --- a/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs +++ b/src/Lucene.Net.Tests/core/Search/TestTermRangeQuery.cs @@ -259,7 +259,7 @@ namespace Lucene.Net.Search } } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new SingleCharTokenizer(reader)); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/6f6f938e/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs b/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs index 0140fc0..421a203 100644 --- a/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs +++ b/src/Lucene.Net.Tests/core/Util/TestQueryBuilder.cs @@ -135,7 +135,7 @@ namespace Lucene.Net.Util /// adds synonym of "dog" for "dogs". internal class MockSynonymAnalyzer : Analyzer { - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader); return new TokenStreamComponents(tokenizer, new MockSynonymFilter(tokenizer)); @@ -169,7 +169,7 @@ namespace Lucene.Net.Util return true; } - if (Input.IncrementToken()) + if (input.IncrementToken()) { AddSynonym = TermAtt.ToString().Equals("dogs"); return true; @@ -240,7 +240,7 @@ namespace Lucene.Net.Util this.OuterInstance = outerInstance; } - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new SimpleCJKTokenizer(reader)); } @@ -316,7 +316,7 @@ namespace Lucene.Net.Util return true; } - if (Input.IncrementToken()) + if (input.IncrementToken()) { AddSynonym = TermAtt.ToString().Equals("国"); return true; @@ -330,7 +330,7 @@ namespace Lucene.Net.Util internal class MockCJKSynonymAnalyzer : Analyzer { - protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + public override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new SimpleCJKTokenizer(reader); return new TokenStreamComponents(tokenizer, new MockCJKSynonymFilter(tokenizer));