lucene-solr-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Matteo Grolla <>
Subject Re: optimize requests that fetch 1000 rows
Date Thu, 11 Feb 2016 15:05:29 GMT
I see a lot of time spent in splitOnTokens

which is called by (last part of stack trace)


2016-02-11 15:42 GMT+01:00 Matteo Grolla <>:

> Hi Yonic,
>      after the first query I find 1000 docs in the document cache.
> I'm using curl to send the request and requesting javabin format to mimic
> the application.
> gc activity is low
> I managed to load the entire 50GB index in the filesystem cache, after
> that queries don't cause disk activity anymore.
> Time improves now queries that took ~30s take <10s. But I hoped better
> I'm going to use jvisualvm's sampler to analyze where time is spent
> 2016-02-11 15:25 GMT+01:00 Yonik Seeley <>:
>> On Thu, Feb 11, 2016 at 7:45 AM, Matteo Grolla <>
>> wrote:
>> > Thanks Toke, yes, they are long times, and solr qtime (to execute the
>> > query) is a fraction of a second.
>> > The response in javabin format is around 300k.
>> OK, That tells us a lot.
>> And if you actually tested so that all the docs would be in the cache
>> (can you verify this by looking at the cache stats after you
>> re-execute?) then it seems like the slowness is down to any of:
>> a) serializing the response (it doesn't seem like a 300K response
>> should take *that* long to serialize)
>> b) reading/processing the response (how fast the client can do
>> something with each doc is also a factor...)
>> c) other (GC, network, etc)
>> You can try taking client processing out of the equation by trying a
>> curl request.
>> -Yonik

  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message