lucene-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hoss Man (JIRA)" <>
Subject [jira] Commented: (SOLR-1911) File descriptor leak while indexing, may cause index corruption
Date Sat, 15 May 2010 00:35:43 GMT


Hoss Man commented on SOLR-1911:

Simon: can you please provide a few more details:

1) what mechanism are you using to index content?  ie: POSTing XML from a remote client? using
the stream.url or stream.file params? Using SolrCell? using DIH? (and if you are using DIH,
from what source? DB? HTTP? File? .. and with what transformers?)

2) what files does lsof show are open after each successive commit until the limit is reached?
 seeing how the file list grows -- specifically which files are never getting closed -- over
time is really the only way to track down what code isn't closing files

(The stack trace you posted shows what it's doing when it runs out of file handles, but there
isn't necessarily any correlation between that and what code should be closing files but isn't

> File descriptor leak while indexing, may cause index corruption
> ---------------------------------------------------------------
>                 Key: SOLR-1911
>                 URL:
>             Project: Solr
>          Issue Type: Bug
>          Components: update
>    Affects Versions: 1.5
>         Environment:         Ubuntu Linux, Java build 1.6.0_16-b01
>         Solr Specification Version:
> 	Solr Implementation Version: 4.0-dev exported - simon - 2010-05-12 16:17:46  -- bult
from updated trunk
> 	Lucene Specification Version: 4.0-dev
> 	Lucene Implementation Version: 4.0-dev exported - 2010-05-12 16:18:26
> 	Current Time: Thu May 13 12:21:12 EDT 2010
> 	Server Start Time:Thu May 13 11:45:41 EDT 2010
>            Reporter: Simon Rosenthal
>            Priority: Critical
> While adding documents to an already existing index using this build, the number of open
file descriptors increases dramatically until the open file per-process limit is reached (1024)
, at which point there are error messages in the log to that effect. If the server is restarted
the index may be corrupt
> commits are handled by autocommit every 60 seconds or 500 documents (usually the time
limit is reached first). 
> mergeFactor is 10.
> It looks as though each time a commit takes place, the number of open files  (obtained
from " lsof -p `cat` | egrep ' [0-9]+r ' ") increases by 40, There are several open
file descriptors associated with each file in the index.
> Rerunning the same index updates with an older Solr (built from trunk in Feb 2010) doesn't
show this problem - the number of open files fluctuates up and down as segments are created
and merged, but stays basically constant.

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message