lucene-solr-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Norberto Meijome <>
Subject Re: NGramTokenizer issue
Date Thu, 26 Jun 2008 00:44:32 GMT
On Wed, 25 Jun 2008 15:37:09 -0300
"Jonathan Ariel" <> wrote:

> I've been trying to use the NGramTokenizer and I ran into a problem.
> It seems like solr is trying to match documents with all the tokens that the
> analyzer returns from the query term. So if I index a document with a title
> field with the value "nice dog" and search for "dog" (where the
> NGramtokenizer is defined to generate tokens of min 2 and max 2) I won't get
> any results.

Hi Jonathan,
I don't have the expertise yet to have gone straight into testing code with
lucene, but my 'black box' testing with ngramtokenizer seems to agree with what
you found - see my latest posts over the last couple of days about this.

Have you tried searching for 'do' or 'ni' or any search term with size =
minGramSize ? I've found that Solr matches results just fine then.

> I can see in the Analysis tool that the tokenizer generates the right
> tokens, but then when solr searches it tries to match the exact Phrase
> instead of the tokens.



{Beto|Norberto|Numard} Meijome

"Some cause happiness wherever they go; others, whenever they go."
  Oscar Wilde

I speak for myself, not my employer. Contents may be hot. Slippery when wet.
Reading disclaimers makes you go blind. Writing them is worse. You have been

View raw message