lucene-solr-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Noble Paul (JIRA)" <>
Subject [jira] Commented: (SOLR-1277) Implement a Solr specific naming service (using Zookeeper)
Date Tue, 08 Sep 2009 05:16:57 GMT


Noble Paul commented on SOLR-1277:

Grant a couple of points.
Let the zookeeper configuration be put into a zookeeper plugin itself.  Let us use the standard
configuration syntax. Custom syntax for each of the plugins is leading to too much of xml
<requestHandler name="/zoo" class="solr.ZooKeeperRequestHandler">
    <!-- See the ZooKeeper docs -->
    <str name="hostPorts">localhost:2181</str>
    <!-- TODO: figure out how to do this programmatically -->
    <str name="me">localhost:8983/solr</str>
    <!-- Timeout for the ZooKeeper.  Optional.  Default 10000 -->
    <!-- Timeout in ms -->
    <str name="timeout">5000</str>
    <str name="shardsNodeName">/solr_shards</str>
    <str name="mastersNodeName">/solr_masters</str>  
    <bool name="shard">true</bool>
    <str name="master">master_group_1</str>

users can drive the values of each of these from an external properties file anyway (using SOLR-1335)

Another important design aspect is that , do we want to have separate zookeeper instances
for each core in a multicore environment? or is it possible to have a zookeeper component
at the CoreContainer level.

> Implement a Solr specific naming service (using Zookeeper)
> ----------------------------------------------------------
>                 Key: SOLR-1277
>                 URL:
>             Project: Solr
>          Issue Type: New Feature
>    Affects Versions: 1.4
>            Reporter: Jason Rutherglen
>            Assignee: Grant Ingersoll
>            Priority: Minor
>             Fix For: 1.5
>         Attachments: log4j-1.2.15.jar, SOLR-1277.patch, zookeeper-3.2.0.jar
>   Original Estimate: 672h
>  Remaining Estimate: 672h
> The goal is to give Solr server clusters self-healing attributes
> where if a server fails, indexing and searching don't stop and
> all of the partitions remain searchable. For configuration, the
> ability to centrally deploy a new configuration without servers
> going offline.
> We can start with basic failover and start from there?
> Features:
> * Automatic failover (i.e. when a server fails, clients stop
> trying to index to or search it)
> * Centralized configuration management (i.e. new solrconfig.xml
> or schema.xml propagates to a live Solr cluster)
> * Optionally allow shards of a partition to be moved to another
> server (i.e. if a server gets hot, move the hot segments out to
> cooler servers). Ideally we'd have a way to detect hot segments
> and move them seamlessly. With NRT this becomes somewhat more
> difficult but not impossible?

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message