lucenenet-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From ccurr...@apache.org
Subject svn commit: r1303294 [4/6] - in /incubator/lucene.net/trunk: src/contrib/Analyzers/AR/ src/contrib/Analyzers/BR/ src/contrib/Analyzers/CJK/ src/contrib/Analyzers/Cn/ src/contrib/Analyzers/Compound/ src/contrib/Analyzers/Cz/ src/contrib/Analyzers/De/ sr...
Date Wed, 21 Mar 2012 06:04:35 GMT
Modified: incubator/lucene.net/trunk/src/core/Search/Spans/TermSpans.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/Spans/TermSpans.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/Spans/TermSpans.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/Spans/TermSpans.cs Wed Mar 21 06:04:26 2012
@@ -97,16 +97,13 @@ namespace Lucene.Net.Search.Spans
 		
 		// TODO: Remove warning after API has been finalized
 
-	    public override ICollection<byte[]> Payload
+	    public override ICollection<byte[]> GetPayload()
 	    {
-	        get
-	        {
-	            byte[] bytes = new byte[positions.PayloadLength];
-	            bytes = positions.GetPayload(bytes, 0);
-	            System.Collections.Generic.List<byte[]> val = new System.Collections.Generic.List<byte[]>();
-	            val.Add(bytes);
-	            return val;
-	        }
+	        byte[] bytes = new byte[positions.PayloadLength];
+	        bytes = positions.GetPayload(bytes, 0);
+	        System.Collections.Generic.List<byte[]> val = new System.Collections.Generic.List<byte[]>();
+	        val.Add(bytes);
+	        return val;
 	    }
 
 	    // TODO: Remove warning after API has been finalized

Modified: incubator/lucene.net/trunk/src/core/Search/TermQuery.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/TermQuery.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/TermQuery.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/TermQuery.cs Wed Mar 21 06:04:26 2012
@@ -80,13 +80,10 @@ namespace Lucene.Net.Search
 		        get { return value_Renamed; }
 		    }
 
-		    public override float SumOfSquaredWeights
+		    public override float GetSumOfSquaredWeights()
 		    {
-		        get
-		        {
-		            queryWeight = idf*Enclosing_Instance.Boost; // compute query weight
-		            return queryWeight*queryWeight; // square it
-		        }
+		        queryWeight = idf*Enclosing_Instance.Boost; // compute query weight
+		        return queryWeight*queryWeight; // square it
 		    }
 
 		    public override void  Normalize(float queryNorm)

Modified: incubator/lucene.net/trunk/src/core/Search/TermScorer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/TermScorer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/TermScorer.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/TermScorer.cs Wed Mar 21 06:04:26 2012
@@ -182,7 +182,7 @@ namespace Lucene.Net.Search
 		}
 		static TermScorer()
 		{
-			SIM_NORM_DECODER = Similarity.GetNormDecoder();
+			SIM_NORM_DECODER = Search.Similarity.GetNormDecoder();
 		}
 	}
 }
\ No newline at end of file

Modified: incubator/lucene.net/trunk/src/core/Search/TimeLimitingCollector.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/TimeLimitingCollector.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/TimeLimitingCollector.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/TimeLimitingCollector.cs Wed Mar 21 06:04:26 2012
@@ -224,12 +224,13 @@ namespace Lucene.Net.Search
 		{
 			collector.SetScorer(scorer);
 		}
-		
-		public override bool AcceptsDocsOutOfOrder()
-		{
-			return collector.AcceptsDocsOutOfOrder();
-		}
-		static TimeLimitingCollector()
+
+	    public override bool AcceptsDocsOutOfOrder
+	    {
+	        get { return collector.AcceptsDocsOutOfOrder; }
+	    }
+
+	    static TimeLimitingCollector()
 		{
 			{
 				TIMER_THREAD.Start();

Modified: incubator/lucene.net/trunk/src/core/Search/TopDocs.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/TopDocs.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/TopDocs.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/TopDocs.cs Wed Mar 21 06:04:26 2012
@@ -40,6 +40,7 @@ namespace Lucene.Net.Search
         }
 
         /// <summary>The top hits for the query. </summary>
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Performance", "CA1819:PropertiesShouldNotReturnArrays")]
         public ScoreDoc[] ScoreDocs
         {
             get { return _scoreDocs; }

Modified: incubator/lucene.net/trunk/src/core/Search/TopFieldCollector.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/TopFieldCollector.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/TopFieldCollector.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/TopFieldCollector.cs Wed Mar 21 06:04:26 2012
@@ -147,11 +147,11 @@ namespace Lucene.Net.Search
 					}
 				}
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		/*
@@ -267,11 +267,11 @@ namespace Lucene.Net.Search
 					}
 				}
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		//
@@ -387,11 +387,11 @@ namespace Lucene.Net.Search
 					}
 				}
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		/*
@@ -571,11 +571,11 @@ namespace Lucene.Net.Search
 					}
 				}
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		/*
@@ -753,11 +753,11 @@ namespace Lucene.Net.Search
 					}
 				}
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		/*
@@ -939,11 +939,11 @@ namespace Lucene.Net.Search
 				this.scorer = scorer;
 				base.SetScorer(scorer);
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		private static readonly ScoreDoc[] EMPTY_SCOREDOCS = new ScoreDoc[0];
@@ -1128,10 +1128,10 @@ namespace Lucene.Net.Search
 			// If this is a maxScoring tracking collector and there were no results, 
 			return new TopFieldDocs(totalHits, results, ((FieldValueHitQueue) pq).GetFields(), maxScore);
 		}
-		
-		public override bool AcceptsDocsOutOfOrder()
-		{
-			return false;
-		}
+
+	    public override bool AcceptsDocsOutOfOrder
+	    {
+	        get { return false; }
+	    }
 	}
 }
\ No newline at end of file

Modified: incubator/lucene.net/trunk/src/core/Search/TopScoreDocCollector.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/TopScoreDocCollector.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/TopScoreDocCollector.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/TopScoreDocCollector.cs Wed Mar 21 06:04:26 2012
@@ -64,11 +64,11 @@ namespace Lucene.Net.Search
 				pqTop.Score = score;
 				pqTop = pq.UpdateTop();
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return false;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return false; }
+		    }
 		}
 		
 		// Assumes docs are scored out of order.
@@ -95,11 +95,11 @@ namespace Lucene.Net.Search
 				pqTop.Score = score;
 				pqTop = pq.UpdateTop();
 			}
-			
-			public override bool AcceptsDocsOutOfOrder()
-			{
-				return true;
-			}
+
+		    public override bool AcceptsDocsOutOfOrder
+		    {
+		        get { return true; }
+		    }
 		}
 		
 		/// <summary> Creates a new <see cref="TopScoreDocCollector" /> given the number of hits to
@@ -111,7 +111,7 @@ namespace Lucene.Net.Search
 		/// <c>numHits</c>, and fill the array with sentinel
 		/// objects.
 		/// </summary>
-		public static TopScoreDocCollector create(int numHits, bool docsScoredInOrder)
+		public static TopScoreDocCollector Create(int numHits, bool docsScoredInOrder)
 		{
 			
 			if (docsScoredInOrder)
@@ -164,12 +164,12 @@ namespace Lucene.Net.Search
 			return new TopDocs(totalHits, results, maxScore);
 		}
 		
-		public override void  SetNextReader(IndexReader reader, int base_Renamed)
+		public override void SetNextReader(IndexReader reader, int base_Renamed)
 		{
 			docBase = base_Renamed;
 		}
 		
-		public override void  SetScorer(Scorer scorer)
+		public override void SetScorer(Scorer scorer)
 		{
 			this.scorer = scorer;
 		}

Modified: incubator/lucene.net/trunk/src/core/Search/Weight.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/Weight.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/Weight.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/Weight.cs Wed Mar 21 06:04:26 2012
@@ -105,7 +105,8 @@ namespace Lucene.Net.Search
 		public abstract Scorer Scorer(IndexReader reader, bool scoreDocsInOrder, bool topScorer);
 
 	    /// <summary>The sum of squared weights of contained query clauses. </summary>
-	    public abstract float SumOfSquaredWeights { get; }
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
+        public abstract float GetSumOfSquaredWeights();
 
 	    /// <summary> Returns true iff this implementation scores docs only out of order. This
 	    /// method is used in conjunction with <see cref="Collector" />'s 
@@ -117,9 +118,10 @@ namespace Lucene.Net.Search
 	    /// <b>NOTE:</b> the default implementation returns <c>false</c>, i.e.
 	    /// the <c>Scorer</c> scores documents in-order.
 	    /// </summary>
-	    public virtual bool ScoresDocsOutOfOrder
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
+        public virtual bool GetScoresDocsOutOfOrder()
 	    {
-	        get { return false; }
+	        return false;
 	    }
 	}
 }
\ No newline at end of file

Modified: incubator/lucene.net/trunk/src/core/Search/WildcardQuery.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Search/WildcardQuery.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Search/WildcardQuery.cs (original)
+++ incubator/lucene.net/trunk/src/core/Search/WildcardQuery.cs Wed Mar 21 06:04:26 2012
@@ -41,24 +41,24 @@ namespace Lucene.Net.Search
 	[Serializable]
 	public class WildcardQuery : MultiTermQuery
 	{
-		private bool termContainsWildcard;
-	    private bool termIsPrefix;
+		private readonly bool _termContainsWildcard;
+	    private readonly bool _termIsPrefix;
 		protected internal Term term;
 		
 		public WildcardQuery(Term term)
 		{ 
 			this.term = term;
 		    string text = term.Text;
-		    this.termContainsWildcard = (term.Text.IndexOf('*') != -1)
+		    _termContainsWildcard = (term.Text.IndexOf('*') != -1)
 		                                || (term.Text.IndexOf('?') != -1);
-		    this.termIsPrefix = termContainsWildcard
+		    _termIsPrefix = _termContainsWildcard
 		                        && (text.IndexOf('?') == -1)
 		                        && (text.IndexOf('*') == text.Length - 1);
 		}
 		
 		protected internal override FilteredTermEnum GetEnum(IndexReader reader)
 		{
-            if (termContainsWildcard)
+            if (_termContainsWildcard)
             {
                 return new WildcardTermEnum(reader, Term);
             }
@@ -76,7 +76,7 @@ namespace Lucene.Net.Search
 
 	    public override Query Rewrite(IndexReader reader)
 		{
-            if (termIsPrefix)
+            if (_termIsPrefix)
             {
                 MultiTermQuery rewritten =
                     new PrefixQuery(term.CreateTerm(term.text.Substring(0, term.text.IndexOf('*'))));

Modified: incubator/lucene.net/trunk/src/core/Store/BufferedIndexInput.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Store/BufferedIndexInput.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Store/BufferedIndexInput.cs (original)
+++ incubator/lucene.net/trunk/src/core/Store/BufferedIndexInput.cs Wed Mar 21 06:04:26 2012
@@ -94,14 +94,6 @@ namespace Lucene.Net.Store
 	    {
 	        get { return _bufferSize; }
 	    }
-        
-        /// <seealso cref="SetBufferSize">
-        /// </seealso>
-        [Obsolete("Use BufferSize property instead.")]
-        public virtual int GetBufferSize()
-        {
-            return BufferSize;
-        }
 
 	    private void  CheckBufferSize(int bufferSize)
 		{

Modified: incubator/lucene.net/trunk/src/core/Store/Directory.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Store/Directory.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Store/Directory.cs (original)
+++ incubator/lucene.net/trunk/src/core/Store/Directory.cs Wed Mar 21 06:04:26 2012
@@ -146,7 +146,7 @@ namespace Lucene.Net.Store
 		{
 		    System.Diagnostics.Debug.Assert(lockFactory != null);
 			this.lockFactory = lockFactory;
-			lockFactory.LockPrefix = this.LockId;
+			lockFactory.LockPrefix = this.GetLockId();
 		}
 
 	    /// <summary> Get the LockFactory that this Directory instance is
@@ -160,26 +160,14 @@ namespace Lucene.Net.Store
 	    }
 
 	    /// <summary> Return a string identifier that uniquely differentiates
-	    /// this Directory instance from other Directory instances.
-	    /// This ID should be the same if two Directory instances
-	    /// (even in different JVMs and/or on different machines)
-	    /// are considered "the same index".  This is how locking
-	    /// "scopes" to the right index.
-	    /// </summary>
-	    public virtual string LockId
-	    {
-	        get { return ToString(); }
-	    }
-
-        /// <summary> Return a string identifier that uniquely differentiates
         /// this Directory instance from other Directory instances.
         /// This ID should be the same if two Directory instances
         /// (even in different JVMs and/or on different machines)
         /// are considered "the same index".  This is how locking
         /// "scopes" to the right index.
         /// </summary>
-        [Obsolete("Use LockId property instead.")]
-        public virtual System.String GetLockID()
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
+        public virtual string GetLockId()
         {
             return ToString();
         }

Modified: incubator/lucene.net/trunk/src/core/Store/FSDirectory.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Store/FSDirectory.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Store/FSDirectory.cs (original)
+++ incubator/lucene.net/trunk/src/core/Store/FSDirectory.cs Wed Mar 21 06:04:26 2012
@@ -421,37 +421,34 @@ namespace Lucene.Net.Store
 		private static readonly char[] HEX_DIGITS = new char[]{'0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f'};
 
 
-	    public override string LockId
+	    public override string GetLockId()
 	    {
-	        get
+	        EnsureOpen();
+	        System.String dirName; // name to be hashed
+	        try
 	        {
-	            EnsureOpen();
-	            System.String dirName; // name to be hashed
-	            try
-	            {
-	                dirName = directory.FullName;
-	            }
-	            catch (System.IO.IOException e)
-	            {
-	                throw new System.SystemException(e.ToString(), e);
-	            }
-
-	            byte[] digest;
-	            lock (DIGESTER)
-	            {
-	                digest = DIGESTER.ComputeHash(System.Text.Encoding.UTF8.GetBytes(dirName));
-	            }
-	            System.Text.StringBuilder buf = new System.Text.StringBuilder();
-	            buf.Append("lucene-");
-	            for (int i = 0; i < digest.Length; i++)
-	            {
-	                int b = digest[i];
-	                buf.Append(HEX_DIGITS[(b >> 4) & 0xf]);
-	                buf.Append(HEX_DIGITS[b & 0xf]);
-	            }
+	            dirName = directory.FullName;
+	        }
+	        catch (System.IO.IOException e)
+	        {
+	            throw new System.SystemException(e.ToString(), e);
+	        }
 
-	            return buf.ToString();
+	        byte[] digest;
+	        lock (DIGESTER)
+	        {
+	            digest = DIGESTER.ComputeHash(System.Text.Encoding.UTF8.GetBytes(dirName));
+	        }
+	        System.Text.StringBuilder buf = new System.Text.StringBuilder();
+	        buf.Append("lucene-");
+	        for (int i = 0; i < digest.Length; i++)
+	        {
+	            int b = digest[i];
+	            buf.Append(HEX_DIGITS[(b >> 4) & 0xf]);
+	            buf.Append(HEX_DIGITS[b & 0xf]);
 	        }
+
+	        return buf.ToString();
 	    }
 
 	    protected override void Dispose(bool disposing)

Modified: incubator/lucene.net/trunk/src/core/Util/AttributeSource.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Util/AttributeSource.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Util/AttributeSource.cs (original)
+++ incubator/lucene.net/trunk/src/core/Util/AttributeSource.cs Wed Mar 21 06:04:26 2012
@@ -131,76 +131,41 @@ namespace Lucene.Net.Util
             this.factory = factory;
 		}
 
-        /// <summary> returns the used AttributeFactory.</summary>
-        public virtual AttributeFactory Factory
+	    /// <summary>Returns the used AttributeFactory.</summary>
+	    public virtual AttributeFactory Factory
 	    {
-            get { return this.factory; }
+	        get { return factory; }
 	    }
 
-		/// <summary> returns the used AttributeFactory.</summary>
-		[Obsolete("Use Factory property instead")]
-		public virtual AttributeFactory GetAttributeFactory()
-		{
-		    return Factory;
-		}
-		
-		/// <summary>Returns a new iterator that iterates the attribute classes
+	    /// <summary>Returns a new iterator that iterates the attribute classes
 		/// in the same order they were added in.
 		/// Signature for Java 1.5: <c>public Iterator&lt;Class&lt;? extends Attribute&gt;&gt; getAttributeClassesIterator()</c>
 		///
 		/// Note that this return value is different from Java in that it enumerates over the values
 		/// and not the keys
 		/// </summary>
-		[Obsolete("Use AttributeClasses property instead")]
-		public virtual IEnumerable<Type> GetAttributeClassesIterator()
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
+        public virtual IEnumerable<Type> GetAttributeTypesIterator()
 		{
-		    return AttributeClasses;
-		}
-
-        /// <summary>Returns a new iterator that iterates the attribute classes
-        /// in the same order they were added in.
-        /// Signature for Java 1.5: <c>public Iterator&lt;Class&lt;? extends Attribute&gt;&gt; getAttributeClassesIterator()</c>
-        ///
-        /// Note that this return value is different from Java in that it enumerates over the values
-        /// and not the keys
-        /// </summary>
-	    public virtual IEnumerable<Type> AttributeClasses
-	    {
-	        get 
-            {
-	            return this.attributes.Select(item => item.Key);
-	        }
-	    }
-		
-		/// <summary>Returns a new iterator that iterates all unique Attribute implementations.
-		/// This iterator may contain less entries that <see cref="GetAttributeClassesIterator" />,
-		/// if one instance implements more than one Attribute interface.
-		/// Signature for Java 1.5: <c>public Iterator&lt;AttributeImpl&gt; getAttributeImplsIterator()</c>
-		/// </summary>
-		[Obsolete("Use AttributeImpls property instead")]
-		public virtual IEnumerable<AttributeImpl> GetAttributeImplsIterator()
-		{
-		    return AttributeImpls;
+            return this.attributes.Select(item => item.Key);
 		}
 
         /// <summary>Returns a new iterator that iterates all unique Attribute implementations.
-        /// This iterator may contain less entries that <see cref="GetAttributeClassesIterator" />,
+        /// This iterator may contain less entries that <see cref="GetAttributeTypesIterator" />,
         /// if one instance implements more than one Attribute interface.
         /// Signature for Java 1.5: <c>public Iterator&lt;AttributeImpl&gt; getAttributeImplsIterator()</c>
         /// </summary>
-	    public virtual IEnumerable<AttributeImpl> AttributeImpls
-	    {
-	        get
-	        {
-                var initState = GetCurrentState();
-                while (initState != null)
-                {
-                    var att = initState.attribute;
-                    initState = initState.next;
-                    yield return att;
-                }
-	        }
-	    }
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Design", "CA1024:UsePropertiesWhereAppropriate")]
+        public virtual IEnumerable<AttributeImpl> GetAttributeImplsIterator()
+        {
+            var initState = GetCurrentState();
+            while (initState != null)
+            {
+                var att = initState.attribute;
+                initState = initState.next;
+                yield return att;
+            }
+        }
 
 	    /// <summary>a cache that stores all interfaces for known implementation classes for performance (slow reflection) </summary>
 	    private static readonly WeakDictionary<Type, System.Collections.Generic.LinkedList<WeakReference>>
@@ -293,14 +258,14 @@ namespace Lucene.Net.Util
 
             return (T)(Attribute)attributes[attClass].Value;
 		}
-		
-		/// <summary>Returns true, iff this AttributeSource has any attributes </summary>
-		public virtual bool HasAttributes()
-		{
-			return this.attributes.Count != 0;
-		}
-		
-		/// <summary> The caller must pass in a Class&lt;? extends Attribute&gt; value. 
+
+	    /// <summary>Returns true, iff this AttributeSource has any attributes </summary>
+	    public virtual bool HasAttributes
+	    {
+	        get { return this.attributes.Count != 0; }
+	    }
+
+	    /// <summary> The caller must pass in a Class&lt;? extends Attribute&gt; value. 
 		/// Returns true, iff this AttributeSource contains the passed-in Attribute.
         /// </summary>\
 		public virtual bool HasAttribute<T>() where T : Attribute
@@ -361,7 +326,7 @@ namespace Lucene.Net.Util
 		private State GetCurrentState()
 		{
 		    var s = currentState[0];
-            if (s != null || !HasAttributes())
+            if (s != null || !HasAttributes)
             {
                 return s;
             }
@@ -455,9 +420,9 @@ namespace Lucene.Net.Util
 			{
 				AttributeSource other = (AttributeSource) obj;
 				
-				if (HasAttributes())
+				if (HasAttributes)
 				{
-					if (!other.HasAttributes())
+					if (!other.HasAttributes)
 					{
 						return false;
 					}
@@ -483,7 +448,7 @@ namespace Lucene.Net.Util
 				}
 				else
 				{
-					return !other.HasAttributes();
+					return !other.HasAttributes;
 				}
 			}
 			else
@@ -494,7 +459,7 @@ namespace Lucene.Net.Util
 		{
             System.Text.StringBuilder sb = new System.Text.StringBuilder().Append('(');
 			
-			if (HasAttributes())
+			if (HasAttributes)
 			{
 				if (currentState[0] == null)
 				{
@@ -519,7 +484,7 @@ namespace Lucene.Net.Util
 			var clone = new AttributeSource(this.factory);
 			
 			// first clone the impls
-			if (HasAttributes())
+			if (HasAttributes)
 			{
                 for (var state = GetCurrentState(); state != null; state = state.next)
                 {

Modified: incubator/lucene.net/trunk/src/core/Util/FieldCacheSanityChecker.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Util/FieldCacheSanityChecker.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Util/FieldCacheSanityChecker.cs (original)
+++ incubator/lucene.net/trunk/src/core/Util/FieldCacheSanityChecker.cs Wed Mar 21 06:04:26 2012
@@ -281,7 +281,7 @@ namespace Lucene.Net.Util
 				System.Object obj = all[i];
 				if (obj is IndexReader)
 				{
-					IndexReader[] subs = ((IndexReader) obj).SequentialSubReaders;
+					IndexReader[] subs = ((IndexReader) obj).GetSequentialSubReaders();
 					for (int j = 0; (null != subs) && (j < subs.Length); j++)
 					{
 						all.Add(subs[j].FieldCacheKey);

Modified: incubator/lucene.net/trunk/src/core/Util/OpenBitSet.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Util/OpenBitSet.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Util/OpenBitSet.cs (original)
+++ incubator/lucene.net/trunk/src/core/Util/OpenBitSet.cs Wed Mar 21 06:04:26 2012
@@ -148,11 +148,12 @@ namespace Lucene.Net.Util
 			return Cardinality() == 0;
 		}
 
-	    /// <summary>Expert: gets or sets the long[] storing the bits </summary>
-	    public virtual long[] Bits
+	    /// <summary>Expert: Gets or sets the long[] storing the bits </summary>
+        [System.Diagnostics.CodeAnalysis.SuppressMessage("Microsoft.Performance", "CA1819:PropertiesShouldNotReturnArrays")]
+        public virtual long[] Bits
 	    {
-	        get { return bits; }
 	        set { this.bits = value; }
+	        get { return bits; }
 	    }
 
 	    /// <summary>Expert: gets or sets the number of longs in the array that are in use </summary>

Modified: incubator/lucene.net/trunk/src/core/Util/ReaderUtil.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/src/core/Util/ReaderUtil.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/src/core/Util/ReaderUtil.cs (original)
+++ incubator/lucene.net/trunk/src/core/Util/ReaderUtil.cs Wed Mar 21 06:04:26 2012
@@ -31,7 +31,7 @@ namespace Lucene.Net.Util
 		/// <param name="reader"></param>
 		public static void GatherSubReaders(System.Collections.Generic.IList<IndexReader> allSubReaders, IndexReader reader)
 		{
-			IndexReader[] subReaders = reader.SequentialSubReaders;
+			IndexReader[] subReaders = reader.GetSequentialSubReaders();
 			if (subReaders == null)
 			{
 				// Add the reader itself, and do not recurse

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Cn/TestChineseTokenizer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Cn/TestChineseTokenizer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Cn/TestChineseTokenizer.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Cn/TestChineseTokenizer.cs Wed Mar 21 06:04:26 2012
@@ -1,4 +1,4 @@
-/*
+/*
  *
  * Licensed to the Apache Software Foundation (ASF) under one
  * or more contributor license agreements.  See the NOTICE file
@@ -45,8 +45,8 @@ namespace Lucene.Net.Analyzers.Cn
             OffsetAttribute offsetAtt = tokenizer.GetAttribute<OffsetAttribute>();
             while (tokenizer.IncrementToken())
             {
-                Assert.AreEqual(correctStartOffset, offsetAtt.StartOffset());
-                Assert.AreEqual(correctEndOffset, offsetAtt.EndOffset());
+                Assert.AreEqual(correctStartOffset, offsetAtt.StartOffset);
+                Assert.AreEqual(correctEndOffset, offsetAtt.EndOffset);
                 correctStartOffset++;
                 correctEndOffset++;
             }

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/NumericPayloadTokenFilterTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/NumericPayloadTokenFilterTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/NumericPayloadTokenFilterTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/NumericPayloadTokenFilterTest.cs Wed Mar 21 06:04:26 2012
@@ -48,7 +48,7 @@ namespace Lucene.Net.Analyzers.Payloads
                 if (termAtt.Term().Equals("dogs"))
                 {
                     seenDogs = true;
-                    Assert.True(typeAtt.Type().Equals("D") == true, typeAtt.Type() + " is not equal to " + "D");
+                    Assert.True(typeAtt.Type.Equals("D") == true, typeAtt.Type + " is not equal to " + "D");
                     Assert.True(payloadAtt.Payload != null, "payloadAtt.GetPayload() is null and it shouldn't be");
                     byte[] bytes = payloadAtt.Payload.GetData();//safe here to just use the bytes, otherwise we should use offset, length
                     Assert.True(bytes.Length == payloadAtt.Payload.Length, bytes.Length + " does not equal: " + payloadAtt.Payload.Length);
@@ -58,7 +58,7 @@ namespace Lucene.Net.Analyzers.Payloads
                 }
                 else
                 {
-                    Assert.True(typeAtt.Type().Equals("word"), typeAtt.Type() + " is not null and it should be");
+                    Assert.True(typeAtt.Type.Equals("word"), typeAtt.Type + " is not null and it should be");
                 }
             }
             Assert.True(seenDogs == true, seenDogs + " does not equal: " + true);
@@ -81,7 +81,7 @@ namespace Lucene.Net.Analyzers.Payloads
                 if (input.IncrementToken())
                 {
                     if (termAtt.Term().Equals("dogs"))
-                        typeAtt.SetType("D");
+                        typeAtt.Type = "D";
                     return true;
                 }
                 else

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilterTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilterTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilterTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TokenOffsetPayloadTokenFilterTest.cs Wed Mar 21 06:04:26 2012
@@ -50,9 +50,9 @@ namespace Lucene.Net.Analyzers.Payloads
                 Assert.True(pay != null, "pay is null and it shouldn't be");
                 byte[] data = pay.GetData();
                 int start = PayloadHelper.DecodeInt(data, 0);
-                Assert.True(start == offsetAtt.StartOffset(), start + " does not equal: " + offsetAtt.StartOffset());
+                Assert.True(start == offsetAtt.StartOffset, start + " does not equal: " + offsetAtt.StartOffset);
                 int end = PayloadHelper.DecodeInt(data, 4);
-                Assert.True(end == offsetAtt.EndOffset(), end + " does not equal: " + offsetAtt.EndOffset());
+                Assert.True(end == offsetAtt.EndOffset, end + " does not equal: " + offsetAtt.EndOffset);
                 count++;
             }
             Assert.True(count == 10, count + " does not equal: " + 10);

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilterTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilterTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilterTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Payloads/TypeAsPayloadTokenFilterTest.cs Wed Mar 21 06:04:26 2012
@@ -46,11 +46,11 @@ namespace Lucene.Net.Analyzers.Payloads
 
             while (nptf.IncrementToken())
             {
-                Assert.True(typeAtt.Type().Equals(char.ToUpper(termAtt.TermBuffer()[0]).ToString()), typeAtt.Type() + " is not null and it should be");
+                Assert.True(typeAtt.Type.Equals(char.ToUpper(termAtt.TermBuffer()[0]).ToString()), typeAtt.Type + " is not null and it should be");
                 Assert.True(payloadAtt.Payload != null, "nextToken.getPayload() is null and it shouldn't be");
                 String type = Encoding.UTF8.GetString(payloadAtt.Payload.GetData()); ;
                 Assert.True(type != null, "type is null and it shouldn't be");
-                Assert.True(type.Equals(typeAtt.Type()) == true, type + " is not equal to " + typeAtt.Type());
+                Assert.True(type.Equals(typeAtt.Type) == true, type + " is not equal to " + typeAtt.Type);
                 count++;
             }
 
@@ -73,7 +73,7 @@ namespace Lucene.Net.Analyzers.Payloads
             {
                 if (input.IncrementToken())
                 {
-                    typeAtt.SetType(char.ToUpper(termAtt.TermBuffer()[0]).ToString());
+                    typeAtt.Type = char.ToUpper(termAtt.TermBuffer()[0]).ToString();
                     return true;
                 }
                 else

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/ShingleFilterTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/ShingleFilterTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/ShingleFilterTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/ShingleFilterTest.cs Wed Mar 21 06:04:26 2012
@@ -470,13 +470,13 @@ namespace Lucene.Net.Analyzers.Shingle
                 String goldText = tokensToCompare[i].Term();
 
                 Assert.AreEqual(goldText, termText, "Wrong termText");
-                Assert.AreEqual(tokensToCompare[i].StartOffset(), offsetAtt.StartOffset(),
+                Assert.AreEqual(tokensToCompare[i].StartOffset, offsetAtt.StartOffset,
                                 "Wrong startOffset for token \"" + termText + "\"");
-                Assert.AreEqual(tokensToCompare[i].EndOffset(), offsetAtt.EndOffset(),
+                Assert.AreEqual(tokensToCompare[i].EndOffset, offsetAtt.EndOffset,
                                 "Wrong endOffset for token \"" + termText + "\"");
                 Assert.AreEqual(positionIncrements[i], posIncrAtt.PositionIncrement,
                                 "Wrong positionIncrement for token \"" + termText + "\"");
-                Assert.AreEqual(types[i], typeAtt.Type(), "Wrong type for token \"" + termText + "\"");
+                Assert.AreEqual(types[i], typeAtt.Type, "Wrong type for token \"" + termText + "\"");
 
                 i++;
             }
@@ -517,9 +517,9 @@ namespace Lucene.Net.Analyzers.Shingle
                 Token t = _testToken[_index++];
 
                 _termAtt.SetTermBuffer(t.TermBuffer(), 0, t.TermLength());
-                _offsetAtt.SetOffset(t.StartOffset(), t.EndOffset());
+                _offsetAtt.SetOffset(t.StartOffset, t.EndOffset);
                 _posIncrAtt.PositionIncrement = t.PositionIncrement;
-                _typeAtt.SetType(TypeAttributeImpl.DEFAULT_TYPE);
+                _typeAtt.Type = TypeAttributeImpl.DEFAULT_TYPE;
 
                 return true;
             }

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/TestShingleMatrixFilter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/TestShingleMatrixFilter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/TestShingleMatrixFilter.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Shingle/TestShingleMatrixFilter.cs Wed Mar 21 06:04:26 2012
@@ -499,8 +499,8 @@ namespace Lucene.Net.Analyzers.Shingle
                             payloadAtt.Payload == null
                                 ? 1f
                                 : PayloadHelper.DecodeFloat(payloadAtt.Payload.GetData()), 0);
-            Assert.AreEqual(startOffset, offsetAtt.StartOffset());
-            Assert.AreEqual(endOffset, offsetAtt.EndOffset());
+            Assert.AreEqual(startOffset, offsetAtt.StartOffset);
+            Assert.AreEqual(endOffset, offsetAtt.EndOffset);
         }
 
         private static void AssertNext(TokenStream ts, String text, int startOffset, int endOffset)
@@ -510,8 +510,8 @@ namespace Lucene.Net.Analyzers.Shingle
 
             Assert.IsTrue(ts.IncrementToken());
             Assert.AreEqual(text, termAtt.Term());
-            Assert.AreEqual(startOffset, offsetAtt.StartOffset());
-            Assert.AreEqual(endOffset, offsetAtt.EndOffset());
+            Assert.AreEqual(startOffset, offsetAtt.StartOffset);
+            Assert.AreEqual(endOffset, offsetAtt.EndOffset);
         }
 
         private static Token CreateToken(String term, int start, int offset)
@@ -561,8 +561,8 @@ namespace Lucene.Net.Analyzers.Shingle
                 _termAtt.SetTermBuffer(prototype.TermBuffer(), 0, prototype.TermLength());
                 _posIncrAtt.PositionIncrement = prototype.PositionIncrement;
                 _flagsAtt.Flags = prototype.Flags;
-                _offsetAtt.SetOffset(prototype.StartOffset(), prototype.EndOffset());
-                _typeAtt.SetType(prototype.Type());
+                _offsetAtt.SetOffset(prototype.StartOffset, prototype.EndOffset);
+                _typeAtt.Type = prototype.Type;
                 _payloadAtt.Payload = prototype.Payload;
 
                 return true;

Modified: incubator/lucene.net/trunk/test/contrib/Analyzers/Sinks/TokenTypeSinkTokenizerTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Analyzers/Sinks/TokenTypeSinkTokenizerTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Analyzers/Sinks/TokenTypeSinkTokenizerTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Analyzers/Sinks/TokenTypeSinkTokenizerTest.cs Wed Mar 21 06:04:26 2012
@@ -53,11 +53,11 @@ namespace Lucene.Net.Analyzers.Sinks
                 if (termAtt.Term().Equals("dogs"))
                 {
                     seenDogs = true;
-                    Assert.True(typeAtt.Type().Equals("D") == true, typeAtt.Type() + " is not equal to " + "D");
+                    Assert.True(typeAtt.Type.Equals("D") == true, typeAtt.Type + " is not equal to " + "D");
                 }
                 else
                 {
-                    Assert.True(typeAtt.Type().Equals("word"), typeAtt.Type() + " is not null and it should be");
+                    Assert.True(typeAtt.Type.Equals("word"), typeAtt.Type + " is not null and it should be");
                 }
             }
             Assert.True(seenDogs == true, seenDogs + " does not equal: " + true);
@@ -90,7 +90,7 @@ namespace Lucene.Net.Analyzers.Sinks
 
                 if (termAtt.Term().Equals("dogs"))
                 {
-                    typeAtt.SetType("D");
+                    typeAtt.Type = "D";
                 }
                 return true;
             }

Modified: incubator/lucene.net/trunk/test/contrib/Core/Index/SegmentsGenCommitTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Core/Index/SegmentsGenCommitTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Core/Index/SegmentsGenCommitTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Core/Index/SegmentsGenCommitTest.cs Wed Mar 21 06:04:26 2012
@@ -92,7 +92,7 @@ namespace Lucene.Net.Test.Index
         {
             SegmentsGenCommit sgCommit = new SegmentsGenCommit(this.directory);
             IndexWriter writer = new IndexWriter(this.directory, new WhitespaceAnalyzer(), null, IndexWriter.MaxFieldLength.UNLIMITED, sgCommit);
-            Assert.AreEqual(10, writer.MaxDoc);
+            Assert.AreEqual(10, writer.MaxDoc());
             IndexReader reader = writer.GetReader();
 
             IndexCommit commit = reader.IndexCommit;

Modified: incubator/lucene.net/trunk/test/contrib/Core/Index/TermVectorEnumeratorTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Core/Index/TermVectorEnumeratorTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Core/Index/TermVectorEnumeratorTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Core/Index/TermVectorEnumeratorTest.cs Wed Mar 21 06:04:26 2012
@@ -86,7 +86,7 @@ namespace Lucene.Net.Test.Index
         public void NoPosNoOffsetTest()
         {
             int doc = 0;
-            foreach (TermFreqVector vector in new TermVectorEnumerator(this.reader, "test"))
+            foreach (ITermFreqVector vector in new TermVectorEnumerator(this.reader, "test"))
             {
                 if (doc == 0)
                 {
@@ -114,7 +114,7 @@ namespace Lucene.Net.Test.Index
             this.reader.Flush();
 
             int doc = 0;
-            foreach (TermFreqVector vector in new TermVectorEnumerator(this.reader, "test"))
+            foreach (ITermFreqVector vector in new TermVectorEnumerator(this.reader, "test"))
             {
                 if (doc == 0)
                 {

Modified: incubator/lucene.net/trunk/test/contrib/Core/Util/Cache/SegmentCacheTest.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Core/Util/Cache/SegmentCacheTest.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Core/Util/Cache/SegmentCacheTest.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Core/Util/Cache/SegmentCacheTest.cs Wed Mar 21 06:04:26 2012
@@ -116,7 +116,7 @@ namespace Lucene.Net.Test.Util.Cache
             // add items to the existing index
             this.AddItemsToIndex(this.directory);
             IndexReader newReader = IndexReader.Open(this.directory, true);
-            Assert.AreEqual(20, newReader.NumDocs);
+            Assert.AreEqual(20, newReader.GetNumDocs());
 
             // test the cache, the old item from the class reader should still be there
             t = MockCache.Instance.Get(newReader, "Cached");

Modified: incubator/lucene.net/trunk/test/contrib/Snowball/Analysis/Snowball/TestSnowball.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/Snowball/Analysis/Snowball/TestSnowball.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/Snowball/Analysis/Snowball/TestSnowball.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/Snowball/Analysis/Snowball/TestSnowball.cs Wed Mar 21 06:04:26 2012
@@ -100,9 +100,9 @@ namespace Lucene.Net.Analysis.Snowball
             filter.IncrementToken();
 
             Assert.AreEqual("accent", termAtt.Term());
-            Assert.AreEqual(2, offsetAtt.StartOffset());
-            Assert.AreEqual(7, offsetAtt.EndOffset());
-            Assert.AreEqual("wrd", typeAtt.Type());
+            Assert.AreEqual(2, offsetAtt.StartOffset);
+            Assert.AreEqual(7, offsetAtt.EndOffset);
+            Assert.AreEqual("wrd", typeAtt.Type);
             Assert.AreEqual(3, posIncAtt.PositionIncrement);
             Assert.AreEqual(77, flagsAtt.Flags);
             Assert.AreEqual(new Payload(new byte[] { 0, 1, 2, 3 }), payloadAtt.Payload);
@@ -143,7 +143,7 @@ namespace Lucene.Net.Analysis.Snowball
                 ClearAttributes();
                 termAtt.SetTermBuffer("accents");
                 offsetAtt.SetOffset(2, 7);
-                typeAtt.SetType("wrd");
+                typeAtt.Type = "wrd";
                 posIncAtt.PositionIncrement = 3;
                 payloadAtt.Payload = new Payload(new byte[] { 0, 1, 2, 3 });
                 flagsAtt.Flags = 77;

Modified: incubator/lucene.net/trunk/test/contrib/SpellChecker/Test/TestSpellChecker.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/contrib/SpellChecker/Test/TestSpellChecker.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/contrib/SpellChecker/Test/TestSpellChecker.cs (original)
+++ incubator/lucene.net/trunk/test/contrib/SpellChecker/Test/TestSpellChecker.cs Wed Mar 21 06:04:26 2012
@@ -219,7 +219,7 @@ namespace SpellChecker.Net.Test.Search.S
         private int Numdoc()
         {
             var rs = IndexReader.Open(spellindex);
-            int num = rs.NumDocs;
+            int num = rs.GetNumDocs();
             Assert.IsTrue(num != 0);
             
             rs.Close();

Modified: incubator/lucene.net/trunk/test/core/Analysis/BaseTokenStreamTestCase.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Analysis/BaseTokenStreamTestCase.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Analysis/BaseTokenStreamTestCase.cs (original)
+++ incubator/lucene.net/trunk/test/core/Analysis/BaseTokenStreamTestCase.cs Wed Mar 21 06:04:26 2012
@@ -117,7 +117,7 @@ namespace Lucene.Net.Analysis
                 ts.ClearAttributes();
                 termAtt.SetTermBuffer("bogusTerm");
                 if (offsetAtt != null) offsetAtt.SetOffset(14584724, 24683243);
-                if (typeAtt != null) typeAtt.SetType("bogusType");
+                if (typeAtt != null) typeAtt.Type = "bogusType";
                 if (posIncrAtt != null) posIncrAtt.PositionIncrement = 45987657;
 
                 checkClearAtt.GetAndResetClearCalled(); // reset it, because we called clearAttribute() before
@@ -126,18 +126,18 @@ namespace Lucene.Net.Analysis
 
                 Assert.AreEqual(output[i], termAtt.Term(), "term " + i);
                 if (startOffsets != null)
-                    Assert.AreEqual(startOffsets[i], offsetAtt.StartOffset(), "startOffset " + i);
+                    Assert.AreEqual(startOffsets[i], offsetAtt.StartOffset, "startOffset " + i);
                 if (endOffsets != null)
-                    Assert.AreEqual(endOffsets[i], offsetAtt.EndOffset(), "endOffset " + i);
+                    Assert.AreEqual(endOffsets[i], offsetAtt.EndOffset, "endOffset " + i);
                 if (types != null)
-                    Assert.AreEqual(types[i], typeAtt.Type(), "type " + i);
+                    Assert.AreEqual(types[i], typeAtt.Type, "type " + i);
                 if (posIncrements != null)
                     Assert.AreEqual(posIncrements[i], posIncrAtt.PositionIncrement, "posIncrement " + i);
             }
             Assert.IsFalse(ts.IncrementToken(), "end of stream");
             ts.End();
             if (finalOffset.HasValue)
-                Assert.AreEqual(finalOffset, offsetAtt.EndOffset(), "finalOffset ");
+                Assert.AreEqual(finalOffset, offsetAtt.EndOffset, "finalOffset ");
             ts.Close();
         }
 

Modified: incubator/lucene.net/trunk/test/core/Analysis/TestKeywordAnalyzer.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Analysis/TestKeywordAnalyzer.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Analysis/TestKeywordAnalyzer.cs (original)
+++ incubator/lucene.net/trunk/test/core/Analysis/TestKeywordAnalyzer.cs Wed Mar 21 06:04:26 2012
@@ -101,8 +101,8 @@ namespace Lucene.Net.Analysis
 			TokenStream stream = new KeywordAnalyzer().TokenStream("field", new System.IO.StringReader("abcd"));
             OffsetAttribute offsetAtt = stream.AddAttribute<OffsetAttribute>();
 			Assert.IsTrue(stream.IncrementToken());
-			Assert.AreEqual(0, offsetAtt.StartOffset());
-			Assert.AreEqual(4, offsetAtt.EndOffset());
+			Assert.AreEqual(0, offsetAtt.StartOffset);
+			Assert.AreEqual(4, offsetAtt.EndOffset);
 		}
 	}
 }
\ No newline at end of file

Modified: incubator/lucene.net/trunk/test/core/Analysis/TestNumericTokenStream.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Analysis/TestNumericTokenStream.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Analysis/TestNumericTokenStream.cs (original)
+++ incubator/lucene.net/trunk/test/core/Analysis/TestNumericTokenStream.cs Wed Mar 21 06:04:26 2012
@@ -44,7 +44,7 @@ namespace Lucene.Net.Analysis
 			{
 				Assert.IsTrue(stream.IncrementToken(), "New token is available");
 				Assert.AreEqual(NumericUtils.LongToPrefixCoded(lvalue, shift), termAtt.Term(), "Term is correctly encoded");
-				Assert.AreEqual((shift == 0)?NumericTokenStream.TOKEN_TYPE_FULL_PREC:NumericTokenStream.TOKEN_TYPE_LOWER_PREC, typeAtt.Type(), "Type correct");
+				Assert.AreEqual((shift == 0)?NumericTokenStream.TOKEN_TYPE_FULL_PREC:NumericTokenStream.TOKEN_TYPE_LOWER_PREC, typeAtt.Type, "Type correct");
 			}
 			Assert.IsFalse(stream.IncrementToken(), "No more tokens available");
 		}
@@ -60,7 +60,7 @@ namespace Lucene.Net.Analysis
 			{
 				Assert.IsTrue(stream.IncrementToken(), "New token is available");
 				Assert.AreEqual(NumericUtils.IntToPrefixCoded(ivalue, shift), termAtt.Term(), "Term is correctly encoded");
-				Assert.AreEqual((shift == 0)?NumericTokenStream.TOKEN_TYPE_FULL_PREC:NumericTokenStream.TOKEN_TYPE_LOWER_PREC, typeAtt.Type(), "Type correct");
+				Assert.AreEqual((shift == 0)?NumericTokenStream.TOKEN_TYPE_FULL_PREC:NumericTokenStream.TOKEN_TYPE_LOWER_PREC, typeAtt.Type, "Type correct");
 			}
 			Assert.IsFalse(stream.IncrementToken(), "No more tokens available");
 		}

Modified: incubator/lucene.net/trunk/test/core/Analysis/TestToken.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Analysis/TestToken.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Analysis/TestToken.cs (original)
+++ incubator/lucene.net/trunk/test/core/Analysis/TestToken.cs Wed Mar 21 06:04:26 2012
@@ -49,14 +49,14 @@ namespace Lucene.Net.Analysis
 			char[] buf = t.TermBuffer();
 			Assert.AreNotEqual(t.TermBuffer(), content);
 			Assert.AreEqual("hello", t.Term());
-			Assert.AreEqual("word", t.Type());
+			Assert.AreEqual("word", t.Type);
 			Assert.AreEqual(0, t.Flags);
 			
 			t = new Token(6, 22);
 			t.SetTermBuffer(content, 0, content.Length);
 			Assert.AreEqual("hello", t.Term());
 			Assert.AreEqual("(hello,6,22)", t.ToString());
-			Assert.AreEqual("word", t.Type());
+			Assert.AreEqual("word", t.Type);
 			Assert.AreEqual(0, t.Flags);
 			
 			t = new Token(6, 22, 7);

Modified: incubator/lucene.net/trunk/test/core/Analysis/Tokenattributes/TestSimpleAttributeImpls.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Analysis/Tokenattributes/TestSimpleAttributeImpls.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Analysis/Tokenattributes/TestSimpleAttributeImpls.cs (original)
+++ incubator/lucene.net/trunk/test/core/Analysis/Tokenattributes/TestSimpleAttributeImpls.cs Wed Mar 21 06:04:26 2012
@@ -76,19 +76,19 @@ namespace Lucene.Net.Analysis.Tokenattri
 		public virtual void  TestTypeAttribute()
 		{
 			TypeAttributeImpl att = new TypeAttributeImpl();
-			Assert.AreEqual(TypeAttributeImpl.DEFAULT_TYPE, att.Type());
+			Assert.AreEqual(TypeAttributeImpl.DEFAULT_TYPE, att.Type);
 			
-			att.SetType("hallo");
+			att.Type = "hallo";
 			Assert.AreEqual("type=hallo", att.ToString());
 			
 			TypeAttributeImpl att2 = (TypeAttributeImpl) AssertCloneIsEqual(att);
-			Assert.AreEqual("hallo", att2.Type());
+			Assert.AreEqual("hallo", att2.Type);
 			
 			att2 = (TypeAttributeImpl) AssertCopyIsEqual(att);
-			Assert.AreEqual("hallo", att2.Type());
+			Assert.AreEqual("hallo", att2.Type);
 			
 			att.Clear();
-			Assert.AreEqual(TypeAttributeImpl.DEFAULT_TYPE, att.Type());
+			Assert.AreEqual(TypeAttributeImpl.DEFAULT_TYPE, att.Type);
 		}
 		
         [Test]
@@ -116,23 +116,23 @@ namespace Lucene.Net.Analysis.Tokenattri
 		public virtual void  TestOffsetAttribute()
 		{
 			OffsetAttributeImpl att = new OffsetAttributeImpl();
-			Assert.AreEqual(0, att.StartOffset());
-			Assert.AreEqual(0, att.EndOffset());
+			Assert.AreEqual(0, att.StartOffset);
+			Assert.AreEqual(0, att.EndOffset);
 			
 			att.SetOffset(12, 34);
 			// no string test here, because order unknown
 			
 			OffsetAttributeImpl att2 = (OffsetAttributeImpl) AssertCloneIsEqual(att);
-			Assert.AreEqual(12, att2.StartOffset());
-			Assert.AreEqual(34, att2.EndOffset());
+			Assert.AreEqual(12, att2.StartOffset);
+			Assert.AreEqual(34, att2.EndOffset);
 			
 			att2 = (OffsetAttributeImpl) AssertCopyIsEqual(att);
-			Assert.AreEqual(12, att2.StartOffset());
-			Assert.AreEqual(34, att2.EndOffset());
+			Assert.AreEqual(12, att2.StartOffset);
+			Assert.AreEqual(34, att2.EndOffset);
 			
 			att.Clear();
-			Assert.AreEqual(0, att.StartOffset());
-			Assert.AreEqual(0, att.EndOffset());
+			Assert.AreEqual(0, att.StartOffset);
+			Assert.AreEqual(0, att.EndOffset);
 		}
 		
 		public static AttributeImpl AssertCloneIsEqual(AttributeImpl att)

Modified: incubator/lucene.net/trunk/test/core/Document/TestBinaryDocument.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Document/TestBinaryDocument.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Document/TestBinaryDocument.cs (original)
+++ incubator/lucene.net/trunk/test/core/Document/TestBinaryDocument.cs Wed Mar 21 06:04:26 2012
@@ -83,7 +83,7 @@ namespace Lucene.Net.Documents
 			
 			/** delete the document from index */
 			reader.DeleteDocument(0);
-			Assert.AreEqual(0, reader.NumDocs);
+			Assert.AreEqual(0, reader.GetNumDocs());
 			
 			reader.Close();
 			dir.Close();

Modified: incubator/lucene.net/trunk/test/core/Index/TestAddIndexesNoOptimize.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestAddIndexesNoOptimize.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestAddIndexesNoOptimize.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestAddIndexesNoOptimize.cs Wed Mar 21 06:04:26 2012
@@ -48,27 +48,27 @@ namespace Lucene.Net.Index
 			writer = NewWriter(dir, true);
 			// add 100 documents
 			AddDocs(writer, 100);
-			Assert.AreEqual(100, writer.MaxDoc);
+			Assert.AreEqual(100, writer.MaxDoc());
 			writer.Close();
 			
 			writer = NewWriter(aux, true);
 			writer.UseCompoundFile = false; // use one without a compound file
 			// add 40 documents in separate files
 			AddDocs(writer, 40);
-            Assert.AreEqual(40, writer.MaxDoc);
+            Assert.AreEqual(40, writer.MaxDoc());
 			writer.Close();
 			
 			writer = NewWriter(aux2, true);
 			// add 40 documents in compound files
 			AddDocs2(writer, 50);
-            Assert.AreEqual(50, writer.MaxDoc);
+            Assert.AreEqual(50, writer.MaxDoc());
 			writer.Close();
 			
 			// test doc count before segments are merged
 			writer = NewWriter(dir, false);
-            Assert.AreEqual(100, writer.MaxDoc);
+            Assert.AreEqual(100, writer.MaxDoc());
 			writer.AddIndexesNoOptimize(new Directory[]{aux, aux2});
-            Assert.AreEqual(190, writer.MaxDoc);
+            Assert.AreEqual(190, writer.MaxDoc());
 			writer.Close();
 			
 			// make sure the old index is correct
@@ -82,14 +82,14 @@ namespace Lucene.Net.Index
 			writer = NewWriter(aux3, true);
 			// add 40 documents
 			AddDocs(writer, 40);
-            Assert.AreEqual(40, writer.MaxDoc);
+            Assert.AreEqual(40, writer.MaxDoc());
 			writer.Close();
 			
 			// test doc count before segments are merged/index is optimized
 			writer = NewWriter(dir, false);
-            Assert.AreEqual(190, writer.MaxDoc);
+            Assert.AreEqual(190, writer.MaxDoc());
 			writer.AddIndexesNoOptimize(new Directory[]{aux3});
-            Assert.AreEqual(230, writer.MaxDoc);
+            Assert.AreEqual(230, writer.MaxDoc());
 			writer.Close();
 			
 			// make sure the new index is correct
@@ -118,9 +118,9 @@ namespace Lucene.Net.Index
 			writer.Close();
 			
 			writer = NewWriter(dir, false);
-            Assert.AreEqual(230, writer.MaxDoc);
+            Assert.AreEqual(230, writer.MaxDoc());
 			writer.AddIndexesNoOptimize(new Directory[]{aux4});
-            Assert.AreEqual(231, writer.MaxDoc);
+            Assert.AreEqual(231, writer.MaxDoc());
 			writer.Close();
 			
 			VerifyNumDocs(dir, 231);
@@ -263,7 +263,7 @@ namespace Lucene.Net.Index
 			writer = NewWriter(dir, true);
 			// add 100 documents
 			AddDocs(writer, 100);
-            Assert.AreEqual(100, writer.MaxDoc);
+            Assert.AreEqual(100, writer.MaxDoc());
 			writer.Close();
 			
 			writer = NewWriter(aux, true);
@@ -287,7 +287,7 @@ namespace Lucene.Net.Index
 			}
 			catch (System.ArgumentException e)
 			{
-                Assert.AreEqual(100, writer.MaxDoc);
+                Assert.AreEqual(100, writer.MaxDoc());
 			}
 			writer.Close();
 			
@@ -310,11 +310,11 @@ namespace Lucene.Net.Index
 			
 			IndexWriter writer = NewWriter(dir, false);
 			writer.SetMaxBufferedDocs(10);
-			writer.SetMergeFactor(4);
+			writer.MergeFactor = 4;
 			AddDocs(writer, 10);
 			
 			writer.AddIndexesNoOptimize(new Directory[]{aux});
-            Assert.AreEqual(1040, writer.MaxDoc);
+            Assert.AreEqual(1040, writer.MaxDoc());
 			Assert.AreEqual(2, writer.GetSegmentCount());
 			Assert.AreEqual(1000, writer.GetDocCount(0));
 			writer.Close();
@@ -336,11 +336,11 @@ namespace Lucene.Net.Index
 			
 			IndexWriter writer = NewWriter(dir, false);
 			writer.SetMaxBufferedDocs(9);
-			writer.SetMergeFactor(4);
+			writer.MergeFactor = 4;
 			AddDocs(writer, 2);
 			
 			writer.AddIndexesNoOptimize(new Directory[]{aux});
-            Assert.AreEqual(1032, writer.MaxDoc);
+            Assert.AreEqual(1032, writer.MaxDoc());
 			Assert.AreEqual(2, writer.GetSegmentCount());
 			Assert.AreEqual(1000, writer.GetDocCount(0));
 			writer.Close();
@@ -362,10 +362,10 @@ namespace Lucene.Net.Index
 			
 			IndexWriter writer = NewWriter(dir, false);
 			writer.SetMaxBufferedDocs(10);
-			writer.SetMergeFactor(4);
+			writer.MergeFactor = 4;
 			
 			writer.AddIndexesNoOptimize(new Directory[]{aux, new RAMDirectory(aux)});
-            Assert.AreEqual(1060, writer.MaxDoc);
+            Assert.AreEqual(1060, writer.MaxDoc());
 			Assert.AreEqual(1000, writer.GetDocCount(0));
 			writer.Close();
 			
@@ -389,15 +389,15 @@ namespace Lucene.Net.Index
 			{
 				reader.DeleteDocument(i);
 			}
-			Assert.AreEqual(10, reader.NumDocs);
+			Assert.AreEqual(10, reader.GetNumDocs());
 			reader.Close();
 			
 			IndexWriter writer = NewWriter(dir, false);
 			writer.SetMaxBufferedDocs(4);
-			writer.SetMergeFactor(4);
+			writer.MergeFactor = 4;
 			
 			writer.AddIndexesNoOptimize(new Directory[]{aux, new RAMDirectory(aux)});
-            Assert.AreEqual(1020, writer.MaxDoc);
+            Assert.AreEqual(1020, writer.MaxDoc());
 			Assert.AreEqual(1000, writer.GetDocCount(0));
 			writer.Close();
 			
@@ -419,9 +419,9 @@ namespace Lucene.Net.Index
 			
 			IndexWriter writer = NewWriter(aux2, true);
 			writer.SetMaxBufferedDocs(100);
-			writer.SetMergeFactor(10);
+			writer.MergeFactor = 10;
 			writer.AddIndexesNoOptimize(new Directory[]{aux});
-            Assert.AreEqual(30, writer.MaxDoc);
+            Assert.AreEqual(30, writer.MaxDoc());
 			Assert.AreEqual(3, writer.GetSegmentCount());
 			writer.Close();
 			
@@ -430,7 +430,7 @@ namespace Lucene.Net.Index
 			{
 				reader.DeleteDocument(i);
 			}
-			Assert.AreEqual(3, reader.NumDocs);
+			Assert.AreEqual(3, reader.GetNumDocs());
 			reader.Close();
 			
 			reader = IndexReader.Open(aux2, false);
@@ -438,15 +438,15 @@ namespace Lucene.Net.Index
 			{
 				reader.DeleteDocument(i);
 			}
-			Assert.AreEqual(22, reader.NumDocs);
+			Assert.AreEqual(22, reader.GetNumDocs());
 			reader.Close();
 			
 			writer = NewWriter(dir, false);
 			writer.SetMaxBufferedDocs(6);
-			writer.SetMergeFactor(4);
+			writer.MergeFactor = 4;
 			
 			writer.AddIndexesNoOptimize(new Directory[]{aux, aux2});
-            Assert.AreEqual(1025, writer.MaxDoc);
+            Assert.AreEqual(1025, writer.MaxDoc());
 			Assert.AreEqual(1000, writer.GetDocCount(0));
 			writer.Close();
 			
@@ -485,7 +485,7 @@ namespace Lucene.Net.Index
 		{
 			IndexReader reader = IndexReader.Open(dir, true);
 			Assert.AreEqual(numDocs, reader.MaxDoc);
-			Assert.AreEqual(numDocs, reader.NumDocs);
+			Assert.AreEqual(numDocs, reader.GetNumDocs());
 			reader.Close();
 		}
 		
@@ -508,14 +508,14 @@ namespace Lucene.Net.Index
 			writer.SetMaxBufferedDocs(1000);
 			// add 1000 documents in 1 segment
 			AddDocs(writer, 1000);
-            Assert.AreEqual(1000, writer.MaxDoc);
+            Assert.AreEqual(1000, writer.MaxDoc());
 			Assert.AreEqual(1, writer.GetSegmentCount());
 			writer.Close();
 			
 			writer = NewWriter(aux, true);
 			writer.UseCompoundFile = false; // use one without a compound file
 			writer.SetMaxBufferedDocs(100);
-			writer.SetMergeFactor(10);
+			writer.MergeFactor = 10;
 			// add 30 documents in 3 segments
 			for (int i = 0; i < 3; i++)
 			{
@@ -524,9 +524,9 @@ namespace Lucene.Net.Index
 				writer = NewWriter(aux, false);
 				writer.UseCompoundFile = false; // use one without a compound file
 				writer.SetMaxBufferedDocs(100);
-				writer.SetMergeFactor(10);
+				writer.MergeFactor = 10;
 			}
-            Assert.AreEqual(30, writer.MaxDoc);
+            Assert.AreEqual(30, writer.MaxDoc());
 			Assert.AreEqual(3, writer.GetSegmentCount());
 			writer.Close();
 		}
@@ -541,7 +541,7 @@ namespace Lucene.Net.Index
 			writer.SetMergePolicy(new LogByteSizeMergePolicy(writer));
 			writer.SetMaxBufferedDocs(5);
 			writer.UseCompoundFile = false;
-			writer.SetMergeFactor(100);
+			writer.MergeFactor = 100;
 			
 			Document doc = new Document();
 			doc.Add(new Field("content", "aaa bbb ccc ddd eee fff ggg hhh iii", Field.Store.YES, Field.Index.ANALYZED, Field.TermVector.WITH_POSITIONS_OFFSETS));
@@ -562,7 +562,7 @@ namespace Lucene.Net.Index
 			LogByteSizeMergePolicy lmp = new LogByteSizeMergePolicy(writer);
 			lmp.MinMergeMB = 0.0001;
 			writer.SetMergePolicy(lmp);
-			writer.SetMergeFactor(4);
+			writer.MergeFactor = 4;
 			writer.UseCompoundFile = false;
 			writer.SetMergeScheduler(new SerialMergeScheduler());
 			writer.AddIndexesNoOptimize(new Directory[]{dir});
@@ -586,7 +586,7 @@ namespace Lucene.Net.Index
 			writer = NewWriter(other, true);
 			writer.UseCompoundFile = true;
 			writer.AddIndexesNoOptimize(new Directory[]{dir});
-			Assert.IsTrue(writer.NewestSegment().UseCompoundFile);
+			Assert.IsTrue(writer.NewestSegment().GetUseCompoundFile());
 			writer.Close();
 		}
 	}

Modified: incubator/lucene.net/trunk/test/core/Index/TestAtomicUpdate.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestAtomicUpdate.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestAtomicUpdate.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestAtomicUpdate.cs Wed Mar 21 06:04:26 2012
@@ -149,7 +149,7 @@ namespace Lucene.Net.Index
 			public override void  DoWork()
 			{
 				IndexReader r = IndexReader.Open(directory, true);
-				Assert.AreEqual(100, r.NumDocs);
+				Assert.AreEqual(100, r.GetNumDocs());
 				r.Close();
 			}
 		}
@@ -165,7 +165,7 @@ namespace Lucene.Net.Index
 
             IndexWriter writer = new MockIndexWriter(this, directory, ANALYZER, true, IndexWriter.MaxFieldLength.UNLIMITED);
 			writer.SetMaxBufferedDocs(7);
-			writer.SetMergeFactor(3);
+			writer.MergeFactor = 3;
 			
 			// Establish a base index of 100 docs:
 			for (int i = 0; i < 100; i++)
@@ -182,7 +182,7 @@ namespace Lucene.Net.Index
 			writer.Commit();
 			
 			IndexReader r = IndexReader.Open(directory, true);
-			Assert.AreEqual(100, r.NumDocs);
+			Assert.AreEqual(100, r.GetNumDocs());
 			r.Close();
 			
 			IndexerThread indexerThread = new IndexerThread(writer, threads);

Modified: incubator/lucene.net/trunk/test/core/Index/TestBackwardsCompatibility.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestBackwardsCompatibility.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestBackwardsCompatibility.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestBackwardsCompatibility.cs Wed Mar 21 06:04:26 2012
@@ -175,7 +175,7 @@ namespace Lucene.Net.Index
                         else
                         {
                             Assert.IsTrue(compressed.IsBinary);
-                            Assert.IsTrue(BINARY_TO_COMPRESS.SequenceEqual(compressed.BinaryValue),
+                            Assert.IsTrue(BINARY_TO_COMPRESS.SequenceEqual(compressed.GetBinaryValue()),
                                           "incorrectly decompressed binary");
                         }
                     }
@@ -191,7 +191,7 @@ namespace Lucene.Net.Index
                         count++;
                         // read the size from the binary value using BinaryReader (this prevents us from doing the shift ops ourselves):
                         // ugh, Java uses Big-Endian streams, so we need to do it manually.
-                        byte[] encodedSize = d.GetFieldable("compressed").BinaryValue.Take(4).Reverse().ToArray();
+                        byte[] encodedSize = d.GetFieldable("compressed").GetBinaryValue().Take(4).Reverse().ToArray();
                         int actualSize = BitConverter.ToInt32(encodedSize, 0);
                         int compressedSize = int.Parse(d.Get("compressedSize"));
                         bool binary = int.Parse(d.Get("id"))%2 > 0;
@@ -402,7 +402,7 @@ namespace Lucene.Net.Index
 			{
 				expected = 46;
 			}
-			Assert.AreEqual(expected, writer.MaxDoc, "wrong doc count");
+			Assert.AreEqual(expected, writer.MaxDoc(), "wrong doc count");
 			writer.Close();
 			
 			// make sure searching sees right # hits
@@ -513,7 +513,7 @@ namespace Lucene.Net.Index
 			{
 				AddDoc(writer, i);
 			}
-			Assert.AreEqual(35, writer.MaxDoc, "wrong doc count");
+			Assert.AreEqual(35, writer.MaxDoc(), "wrong doc count");
 			writer.Close();
 			
 			// open fresh writer so we get no prx file in the added segment
@@ -553,7 +553,7 @@ namespace Lucene.Net.Index
                 {
                     AddDoc(writer, i);
                 }
-                Assert.AreEqual(35, writer.MaxDoc, "wrong doc count");
+                Assert.AreEqual(35, writer.MaxDoc(), "wrong doc count");
                 writer.Close();
 
                 // Delete one doc so we get a .del file:

Modified: incubator/lucene.net/trunk/test/core/Index/TestConcurrentMergeScheduler.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestConcurrentMergeScheduler.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestConcurrentMergeScheduler.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestConcurrentMergeScheduler.cs Wed Mar 21 06:04:26 2012
@@ -121,7 +121,7 @@ namespace Lucene.Net.Index
 			
 			writer.Close();
 			IndexReader reader = IndexReader.Open(directory, true);
-			Assert.AreEqual(200+extraCount, reader.NumDocs);
+			Assert.AreEqual(200+extraCount, reader.GetNumDocs());
 			reader.Close();
 			directory.Close();
 		}
@@ -170,7 +170,7 @@ namespace Lucene.Net.Index
 			writer.Close();
 			IndexReader reader = IndexReader.Open(directory, true);
 			// Verify that we did not lose any deletes...
-			Assert.AreEqual(450, reader.NumDocs);
+			Assert.AreEqual(450, reader.GetNumDocs());
 			reader.Close();
 			directory.Close();
 		}
@@ -219,7 +219,7 @@ namespace Lucene.Net.Index
                 ConcurrentMergeScheduler cms = new ConcurrentMergeScheduler();
                 writer.SetMergeScheduler(cms);
                 writer.SetMaxBufferedDocs(2);
-                writer.SetMergeFactor(100);
+                writer.MergeFactor = 100;
 
                 for (int j = 0; j < 201; j++)
                 {
@@ -236,14 +236,14 @@ namespace Lucene.Net.Index
 
                 // Force a bunch of merge threads to kick off so we
                 // stress out aborting them on close:
-                writer.SetMergeFactor(3);
+                writer.MergeFactor = 3;
                 writer.AddDocument(doc);
                 writer.Commit();
 
                 writer.Close(false);
 
                 IndexReader reader = IndexReader.Open(directory, true);
-                Assert.AreEqual((1 + iter)*182, reader.NumDocs);
+                Assert.AreEqual((1 + iter)*182, reader.GetNumDocs());
                 reader.Close();
 
                 // Reopen

Modified: incubator/lucene.net/trunk/test/core/Index/TestCrash.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestCrash.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestCrash.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestCrash.cs Wed Mar 21 06:04:26 2012
@@ -58,7 +58,7 @@ namespace Lucene.Net.Index
 		
 		private void  Crash(IndexWriter writer)
 		{
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			ConcurrentMergeScheduler cms = (ConcurrentMergeScheduler) writer.MergeScheduler;
 			dir.Crash();
 			cms.Sync();
@@ -69,34 +69,34 @@ namespace Lucene.Net.Index
 		public virtual void  TestCrashWhileIndexing()
 		{
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			Crash(writer);
 			IndexReader reader = IndexReader.Open(dir, true);
-			Assert.IsTrue(reader.NumDocs < 157);
+			Assert.IsTrue(reader.GetNumDocs() < 157);
 		}
 		
 		[Test]
 		public virtual void  TestWriterAfterCrash()
 		{
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			dir.SetPreventDoubleWrite(false);
 			Crash(writer);
 			writer = InitIndex(dir);
 			writer.Close();
 			
 			IndexReader reader = IndexReader.Open(dir, false);
-			Assert.IsTrue(reader.NumDocs < 314);
+			Assert.IsTrue(reader.GetNumDocs() < 314);
 		}
 		
 		[Test]
 		public virtual void  TestCrashAfterReopen()
 		{
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			writer.Close();
 			writer = InitIndex(dir);
-			Assert.AreEqual(314, writer.MaxDoc);
+			Assert.AreEqual(314, writer.MaxDoc());
 			Crash(writer);
 			
 			/*
@@ -109,7 +109,7 @@ namespace Lucene.Net.Index
 			*/
 			
 			IndexReader reader = IndexReader.Open(dir, false);
-			Assert.IsTrue(reader.NumDocs >= 157);
+			Assert.IsTrue(reader.GetNumDocs() >= 157);
 		}
 		
 		[Test]
@@ -117,7 +117,7 @@ namespace Lucene.Net.Index
 		{
 			
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			
 			writer.Close();
 			dir.Crash();
@@ -130,7 +130,7 @@ namespace Lucene.Net.Index
 			*/
 			
 			IndexReader reader = IndexReader.Open(dir, false);
-			Assert.AreEqual(157, reader.NumDocs);
+			Assert.AreEqual(157, reader.GetNumDocs());
 		}
 		
 		[Test]
@@ -138,7 +138,7 @@ namespace Lucene.Net.Index
 		{
 			
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			
 			writer.Close(false);
 			
@@ -151,7 +151,7 @@ namespace Lucene.Net.Index
 			System.out.println("file " + i + " = " + l[i] + " " + dir.fileLength(l[i]) + " bytes");
 			*/
 			IndexReader reader = IndexReader.Open(dir, false);
-			Assert.AreEqual(157, reader.NumDocs);
+			Assert.AreEqual(157, reader.GetNumDocs());
 		}
 		
 		[Test]
@@ -159,7 +159,7 @@ namespace Lucene.Net.Index
 		{
 			
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			
 			writer.Close(false);
 			IndexReader reader = IndexReader.Open(dir, false);
@@ -174,7 +174,7 @@ namespace Lucene.Net.Index
 			System.out.println("file " + i + " = " + l[i] + " " + dir.fileLength(l[i]) + " bytes");
 			*/
 			reader = IndexReader.Open(dir, false);
-			Assert.AreEqual(157, reader.NumDocs);
+			Assert.AreEqual(157, reader.GetNumDocs());
 		}
 		
 		[Test]
@@ -182,7 +182,7 @@ namespace Lucene.Net.Index
 		{
 			
 			IndexWriter writer = InitIndex();
-			MockRAMDirectory dir = (MockRAMDirectory) writer.GetDirectory();
+			MockRAMDirectory dir = (MockRAMDirectory) writer.Directory;
 			
 			writer.Close(false);
 			IndexReader reader = IndexReader.Open(dir, false);
@@ -198,7 +198,7 @@ namespace Lucene.Net.Index
 			System.out.println("file " + i + " = " + l[i] + " " + dir.fileLength(l[i]) + " bytes");
 			*/
 			reader = IndexReader.Open(dir, false);
-			Assert.AreEqual(156, reader.NumDocs);
+			Assert.AreEqual(156, reader.GetNumDocs());
 		}
 	}
 }
\ No newline at end of file

Modified: incubator/lucene.net/trunk/test/core/Index/TestDeletionPolicy.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestDeletionPolicy.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestDeletionPolicy.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestDeletionPolicy.cs Wed Mar 21 06:04:26 2012
@@ -95,7 +95,7 @@ namespace Lucene.Net.Index
 			{
 				IndexCommit lastCommit = (IndexCommit) commits[commits.Count - 1];
 				IndexReader r = IndexReader.Open(dir, true);
-				Assert.AreEqual(r.IsOptimized, lastCommit.IsOptimized, "lastCommit.isOptimized()=" + lastCommit.IsOptimized + " vs IndexReader.isOptimized=" + r.IsOptimized);
+				Assert.AreEqual(r.IsOptimized(), lastCommit.IsOptimized, "lastCommit.isOptimized()=" + lastCommit.IsOptimized + " vs IndexReader.isOptimized=" + r.IsOptimized());
 				r.Close();
 				Enclosing_Instance.VerifyCommitOrder(commits);
 				numOnCommit++;
@@ -481,8 +481,8 @@ namespace Lucene.Net.Index
 			
 			IndexReader r = IndexReader.Open(dir, true);
 			// Still optimized, still 11 docs
-			Assert.IsTrue(r.IsOptimized);
-			Assert.AreEqual(11, r.NumDocs);
+			Assert.IsTrue(r.IsOptimized());
+			Assert.AreEqual(11, r.GetNumDocs());
 			r.Close();
 			
 			writer = new IndexWriter(dir, new WhitespaceAnalyzer(), policy, IndexWriter.MaxFieldLength.LIMITED, lastCommit);
@@ -496,8 +496,8 @@ namespace Lucene.Net.Index
 			r = IndexReader.Open(dir, true);
 			// Not optimized because we rolled it back, and now only
 			// 10 docs
-			Assert.IsTrue(!r.IsOptimized);
-			Assert.AreEqual(10, r.NumDocs);
+			Assert.IsTrue(!r.IsOptimized());
+			Assert.AreEqual(10, r.GetNumDocs());
 			r.Close();
 			
 			// Reoptimize
@@ -506,8 +506,8 @@ namespace Lucene.Net.Index
 			writer.Close();
 			
 			r = IndexReader.Open(dir, true);
-			Assert.IsTrue(r.IsOptimized);
-			Assert.AreEqual(10, r.NumDocs);
+			Assert.IsTrue(r.IsOptimized());
+			Assert.AreEqual(10, r.GetNumDocs());
 			r.Close();
 			
 			// Now open writer on the commit just before optimize,
@@ -518,16 +518,16 @@ namespace Lucene.Net.Index
 			// Reader still sees optimized index, because writer
 			// opened on the prior commit has not yet committed:
 			r = IndexReader.Open(dir, true);
-			Assert.IsTrue(r.IsOptimized);
-			Assert.AreEqual(10, r.NumDocs);
+			Assert.IsTrue(r.IsOptimized());
+			Assert.AreEqual(10, r.GetNumDocs());
 			r.Close();
 			
 			writer.Close();
 			
 			// Now reader sees unoptimized index:
 			r = IndexReader.Open(dir, true);
-			Assert.IsTrue(!r.IsOptimized);
-			Assert.AreEqual(10, r.NumDocs);
+			Assert.IsTrue(!r.IsOptimized());
+			Assert.AreEqual(10, r.GetNumDocs());
 			r.Close();
 			
 			dir.Close();

Modified: incubator/lucene.net/trunk/test/core/Index/TestDirectoryReader.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestDirectoryReader.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestDirectoryReader.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestDirectoryReader.cs Wed Mar 21 06:04:26 2012
@@ -94,7 +94,7 @@ namespace Lucene.Net.Index
 			Document newDoc2 = reader.Document(1);
 			Assert.IsTrue(newDoc2 != null);
 			Assert.IsTrue(DocHelper.NumFields(newDoc2) == DocHelper.NumFields(doc2) - DocHelper.unstored.Count);
-			TermFreqVector vector = reader.GetTermFreqVector(0, DocHelper.TEXT_FIELD_2_KEY);
+			ITermFreqVector vector = reader.GetTermFreqVector(0, DocHelper.TEXT_FIELD_2_KEY);
 			Assert.IsTrue(vector != null);
 			TestSegmentReader.CheckNorms(reader);
 		}
@@ -104,11 +104,11 @@ namespace Lucene.Net.Index
 			sis.Read(dir);
 			IndexReader reader = OpenReader();
 			Assert.IsTrue(reader != null);
-			Assert.AreEqual(2, reader.NumDocs);
+			Assert.AreEqual(2, reader.GetNumDocs());
 			reader.DeleteDocument(0);
-			Assert.AreEqual(1, reader.NumDocs);
+			Assert.AreEqual(1, reader.GetNumDocs());
 			reader.UndeleteAll();
-			Assert.AreEqual(2, reader.NumDocs);
+			Assert.AreEqual(2, reader.GetNumDocs());
 			
 			// Ensure undeleteAll survives commit/close/reopen:
 			reader.Commit();
@@ -121,10 +121,10 @@ namespace Lucene.Net.Index
 			
 			sis.Read(dir);
 			reader = OpenReader();
-			Assert.AreEqual(2, reader.NumDocs);
+			Assert.AreEqual(2, reader.GetNumDocs());
 			
 			reader.DeleteDocument(0);
-			Assert.AreEqual(1, reader.NumDocs);
+			Assert.AreEqual(1, reader.GetNumDocs());
 			reader.Commit();
 			reader.Close();
 			if (reader is MultiReader)
@@ -133,7 +133,7 @@ namespace Lucene.Net.Index
 				sis.Commit(dir);
 			sis.Read(dir);
 			reader = OpenReader();
-			Assert.AreEqual(1, reader.NumDocs);
+			Assert.AreEqual(1, reader.GetNumDocs());
 		}
 		
 		
@@ -153,11 +153,11 @@ namespace Lucene.Net.Index
 			AddDoc(ramDir2, "test blah", true);
 			IndexReader[] readers = new IndexReader[]{IndexReader.Open(ramDir1, false), IndexReader.Open(ramDir2, false)};
 			MultiReader mr = new MultiReader(readers);
-			Assert.IsTrue(mr.IsCurrent); // just opened, must be current
+			Assert.IsTrue(mr.IsCurrent()); // just opened, must be current
 			AddDoc(ramDir1, "more text", false);
-			Assert.IsFalse(mr.IsCurrent); // has been modified, not current anymore
+			Assert.IsFalse(mr.IsCurrent()); // has been modified, not current anymore
 			AddDoc(ramDir2, "even more text", false);
-			Assert.IsFalse(mr.IsCurrent); // has been modified even more, not current anymore
+			Assert.IsFalse(mr.IsCurrent()); // has been modified even more, not current anymore
 			try
 			{
 				var ver = mr.Version;

Modified: incubator/lucene.net/trunk/test/core/Index/TestDoc.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestDoc.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestDoc.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestDoc.cs Wed Mar 21 06:04:26 2012
@@ -226,7 +226,7 @@ namespace Lucene.Net.Index
 		{
 			SegmentReader reader = SegmentReader.Get(true, si, IndexReader.DEFAULT_TERMS_INDEX_DIVISOR);
 			
-			for (int i = 0; i < reader.NumDocs; i++)
+			for (int i = 0; i < reader.GetNumDocs(); i++)
 			{
 				out_Renamed.WriteLine(reader.Document(i));
 			}

Modified: incubator/lucene.net/trunk/test/core/Index/TestDocumentWriter.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestDocumentWriter.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestDocumentWriter.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestDocumentWriter.cs Wed Mar 21 06:04:26 2012
@@ -386,11 +386,11 @@ namespace Lucene.Net.Index
 			
 			IndexReader reader = IndexReader.Open(dir, true);
 			// f1
-			TermFreqVector tfv1 = reader.GetTermFreqVector(0, "f1");
+			ITermFreqVector tfv1 = reader.GetTermFreqVector(0, "f1");
 			Assert.IsNotNull(tfv1);
 			Assert.AreEqual(2, tfv1.GetTerms().Length, "the 'with_tv' setting should rule!");
 			// f2
-			TermFreqVector tfv2 = reader.GetTermFreqVector(0, "f2");
+			ITermFreqVector tfv2 = reader.GetTermFreqVector(0, "f2");
 			Assert.IsNotNull(tfv2);
 			Assert.AreEqual(2, tfv2.GetTerms().Length, "the 'with_tv' setting should rule!");
 		}

Modified: incubator/lucene.net/trunk/test/core/Index/TestFieldsReader.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestFieldsReader.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestFieldsReader.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestFieldsReader.cs Wed Mar 21 06:04:26 2012
@@ -186,7 +186,7 @@ namespace Lucene.Net.Index
 			Assert.IsTrue(field != null, "field is null and it shouldn't be");
 			Assert.IsTrue(field.StringValue == null, "stringValue isn't null for lazy binary field");
 			
-			byte[] bytes = field.BinaryValue;
+			byte[] bytes = field.GetBinaryValue();
 			Assert.IsTrue(bytes != null, "bytes is null and it shouldn't be");
 			Assert.IsTrue(DocHelper.LAZY_FIELD_BINARY_BYTES.Length == bytes.Length, "");
 			for (int i = 0; i < bytes.Length; i++)
@@ -339,9 +339,9 @@ namespace Lucene.Net.Index
 			Assert.IsTrue(f1.IsBinary);
 			Assert.IsTrue(!f3.IsBinary);
 			Assert.IsTrue(fb.IsBinary);
-			AssertSizeEquals(2 * DocHelper.FIELD_1_TEXT.Length, f1.BinaryValue);
+			AssertSizeEquals(2 * DocHelper.FIELD_1_TEXT.Length, f1.GetBinaryValue());
 			Assert.AreEqual(DocHelper.FIELD_3_TEXT, f3.StringValue);
-            AssertSizeEquals(DocHelper.LAZY_FIELD_BINARY_BYTES.Length, fb.BinaryValue);
+            AssertSizeEquals(DocHelper.LAZY_FIELD_BINARY_BYTES.Length, fb.GetBinaryValue());
 
             reader.Dispose();
 		}

Modified: incubator/lucene.net/trunk/test/core/Index/TestFilterIndexReader.cs
URL: http://svn.apache.org/viewvc/incubator/lucene.net/trunk/test/core/Index/TestFilterIndexReader.cs?rev=1303294&r1=1303293&r2=1303294&view=diff
==============================================================================
--- incubator/lucene.net/trunk/test/core/Index/TestFilterIndexReader.cs (original)
+++ incubator/lucene.net/trunk/test/core/Index/TestFilterIndexReader.cs Wed Mar 21 06:04:26 2012
@@ -123,7 +123,7 @@ namespace Lucene.Net.Index
 			
 			IndexReader reader = new TestReader(IndexReader.Open(directory, true));
 			
-			Assert.IsTrue(reader.IsOptimized);
+			Assert.IsTrue(reader.IsOptimized());
 			
 			TermEnum terms = reader.Terms();
 			while (terms.Next())



Mime
View raw message