lucene-solr-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From neosky <neosk...@yahoo.com>
Subject Re: Fail to huge collection extraction
Date Sun, 09 Sep 2012 15:24:29 GMT
Thanks Alex!
Yes, you hit my key points.
Actually I have to implement both of the requirements.
The first one works very well as the reason you state. Now I have a website
client which is 20 records per page. It is fast.
However, my customer also wants to use Servlet to  download the whole query
set.(1 millions records maximum possible)
So at this time, I tried to use Solr pull out 10000 or 5000 records for each
page(Divided to 100 times or 200 times queries) . Then just print out these
records to the client browser. 
I am not sure how the exception was generated?
Is my client program(the Servlet program)  out of memory?or Connect timeout
for some reason?
This exception doesn't always happen. Sometimes it works well even I query
10000 records each time and works for many times , but sometimes it crashes
only 5000 records without an explicit reason.   
You suggestion is great! But the implementation is a little complicated for
us.
Is Lucene better than Solr for this requirement? But the paging in Lucene
seems not very intuitively.  



--
View this message in context: http://lucene.472066.n3.nabble.com/Fail-to-huge-collection-extraction-tp4003559p4006450.html
Sent from the Solr - User mailing list archive at Nabble.com.

Mime
View raw message