lucene-solr-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Chetas Joshi <chetas.jo...@gmail.com>
Subject Re: index dir of core xxx is already locked.
Date Wed, 16 Nov 2016 23:01:18 GMT
I don't kill the solr instance forcefully using "kill -9".

I checked the core.properties file for that shard. The content is different
from the core.properties file for all the other shards.
It has the following two lines which are different

config=solrconfig.xml

schema=schema.xml

In other shards, it is

collection.configName=v4 (name I have given to the config)

name=collectionName_shardNumber_replica1

Should I modify this file before restarting the Cloud?

There is a strange thing I just observed about the data dir of the shard
that is not coming up. There is an addition index dir that has been created

hdfs://Ingest/solr53/collection/core_node32/data/index/index/

The size and content is same as of

hdfs://Ingest/solr53/collection/core_node32/data/index/


What could be the reason of this extra dir? Should I delete it?


Thanks!


On Wed, Nov 16, 2016 at 1:51 PM, Erick Erickson <erickerickson@gmail.com>
wrote:

> bq: Before restarting, I delete all the write.lock files from the data
> dir. But
> every time I restart I get the same exception.
>
> First, this shouldn't be necessary. Are you by any chance killing the
> Solr instances with
> the equivalent of "kill -9"? Allow them to shut down gracefully. That
> said, until recently
> the bin/solr script would kill them forcefully after 5 seconds which
> is too short an interval.
>
> But the error really is telling you that somehow two or more Solr
> cores are pointing at the
> same data directory. Whichever one gets there first will block any
> later cores with the
> message you see. So check your core.properties files and your HDFS magic
> to see
> how this is occurring would be my first guess.
>
> Best,
> Erick
>
> On Wed, Nov 16, 2016 at 1:38 PM, Chetas Joshi <chetas.joshi@gmail.com>
> wrote:
> > Hi,
> >
> > I have a SolrCloud (on HDFS) of 52 nodes. I have 3 collections each with
> 50
> > shards and maxShards per node for every collection is 1.
> >
> > I am having problem restarting a solr shard for a collection.
> >
> > When I restart, there is always a particular shard of a particular
> > collection that remains down. The 2 shards on the same host for the rest
> of
> > the collections are up and running.
> >
> > Before restarting, I delete all the write.lock files from the data dir.
> But
> > every time I restart I get the same exception.
> >
> > index dir yyy of core xxx is already locked. The most likely cause is
> > another Solr server (or another solr core in this server) also configured
> > to use this directory; other possible causes may be specific to lockType:
> > hdfs
> >
> > Thanks!
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message