lucene-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Robert Muir (JIRA)" <>
Subject [jira] Commented: (LUCENE-2090) convert automaton to char[] based processing and TermRef / TermsEnum api
Date Tue, 24 Nov 2009 18:14:39 GMT


Robert Muir commented on LUCENE-2090:

Mike, I implemented this common suffix, but only for dumb mode, it does not help smart mode.
so i got rid of common prefix entirely, as its useless, and just replaced it.
I also take measures to ensure the suffix is well-formed UTF-8 :)

on my *N trunk tests its now 5700/5800ms on average versus 6000ms, just using String.endsWith()
before checking the DFA.
its a consistent gain, so I think for really crappy worst-case wildcards and regular expressions,

we have a lot to gain by doing this with bytes, before converting to char[] and running against
the DFA.

I guess since TermRef exposes all the bytes, I could implement endsWith myself in AutomatonTermsEnum
in the future, 
but it seems like it would be a nice complement to startsWith() ?

> convert automaton to char[] based processing and TermRef / TermsEnum api
> ------------------------------------------------------------------------
>                 Key: LUCENE-2090
>                 URL:
>             Project: Lucene - Java
>          Issue Type: Improvement
>          Components: Search
>            Reporter: Robert Muir
>            Priority: Minor
>             Fix For: 3.1
> The automaton processing is currently done with String, mostly because TermEnum is based
on String.
> it is easy to change the processing to work with char[], since behind the scenes this
is used anyway.
> in general I think we should make sure char[] based processing is exposed in the automaton
pkg anyway, for things like pattern-based tokenizers and such.

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message