lucene-solr-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Mark Miller (JIRA)" <>
Subject [jira] Commented: (SOLR-1277) Implement a Solr specific naming service (using Zookeeper)
Date Wed, 16 Dec 2009 20:04:18 GMT


Mark Miller commented on SOLR-1277:

I think the timeouts are going to have to be different depending on the role of the particular
node. In a really distributed setup, indexing nodes are generally more likely to have long
GC pauses than searcher nodes, and a lengthy GC pause on an indexer is usually not a problem.
However, if a searcher node goes out on a long GC pause then you need to find out fast and
bypass the box before too many queries back up and need to be retried (though even this depends
on throughput, response time, and number of other available nodes.){quote}

Currently, I've got a default timeout, with the ability to override it at any node in solr.xml.
Do you think thats enough?

I can imagine putting the timeout for different roles in ZooKeeper, and then a node gets its
timeout there based on its role - but then it would have to make multiple connections - one
with a default timeout to get its timeout, and then another with the correct timeout.

> Implement a Solr specific naming service (using Zookeeper)
> ----------------------------------------------------------
>                 Key: SOLR-1277
>                 URL:
>             Project: Solr
>          Issue Type: New Feature
>    Affects Versions: 1.4
>            Reporter: Jason Rutherglen
>            Assignee: Grant Ingersoll
>            Priority: Minor
>             Fix For: 1.5
>         Attachments: log4j-1.2.15.jar, SOLR-1277.patch, SOLR-1277.patch, SOLR-1277.patch,
SOLR-1277.patch, zookeeper-3.2.1.jar
>   Original Estimate: 672h
>  Remaining Estimate: 672h
> The goal is to give Solr server clusters self-healing attributes
> where if a server fails, indexing and searching don't stop and
> all of the partitions remain searchable. For configuration, the
> ability to centrally deploy a new configuration without servers
> going offline.
> We can start with basic failover and start from there?
> Features:
> * Automatic failover (i.e. when a server fails, clients stop
> trying to index to or search it)
> * Centralized configuration management (i.e. new solrconfig.xml
> or schema.xml propagates to a live Solr cluster)
> * Optionally allow shards of a partition to be moved to another
> server (i.e. if a server gets hot, move the hot segments out to
> cooler servers). Ideally we'd have a way to detect hot segments
> and move them seamlessly. With NRT this becomes somewhat more
> difficult but not impossible?

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message