incubator-cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Шамим <sre...@yandex.ru>
Subject Re:Data not replicating to all datacenters
Date Mon, 03 Dec 2012 17:45:15 GMT
Yes, it's should be. dc1:3 means u have 3 copy of every piece of row, with local quorum you
always get a good consistency from 3 nodes. 
First you have to calculate token for data center dc1 and add offset 100 to token for the
second data center which will resolve your problem. After creating the keyspace you can run
the nodetool command with ring KEYSPACE NAME which should show u the load for every node as
33%
Hope it will help u
Shamim
> Hi Shamim
> 
> I have read a bit about the Tokens. I understand how that could effect
> the data distribution at first, but I don't understand if we have
> specified Options: [dc1:3, dc2:3], surely after a while all the data
> will be on every server?
> 
> Thanks,
> 
> Owen
> 
> On 3 December 2012 14:06, Шамим <srecon@yandex.ru> wrote:
> 
>> Hello Owen,
>> Seems you did not configure token for all nodes correctly. See the section Calculating
Tokens for multiple data centers here http://www.datastax.com/docs/0.8/install/cluster_init
>>
>> Best regards
>> Shamim
>> ---
>> On Mon, Dec 3, 2012 at 4:42 PM, Owen Davies <cassandra@obduk.com> wrote:
>>
>> We have a 2 data center test cassandra setup running, and are writing
>> to it using LOCAL_QUORUM. When reading, sometimes the data is there,
>> sometimes it is not, which we think is a replication issue, even
>> though we have left it plenty of time after the writes.
>>
>> We have the following setup:
>>
>> cassandra -v: 1.1.6
>>
>> cassandra.yaml
>> -----------------------
>>
>> cluster_name: something
>>
>> endpoint_snitch: PropertyFileSnitch
>>
>> cassandra-topology.properties
>> --------------------------------------------
>> 192.168.1.1=dc1:rack1
>> 192.168.1.2=dc1:rack1
>> 192.168.1.3=dc1:rack1
>>
>> 192.168.2.1=dc2:rack1
>> 192.168.2.2=dc2:rack1
>> 192.168.2.3=dc3:rack1
>>
>> default=nodc:norack
>>
>> cassandra-cli
>> --------------------
>> Keyspace: example:
>> Replication Strategy: org.apache.cassandra.locator.NetworkTopologyStrategy
>> Durable Writes: true
>> Options: [dc1:3, dc2:3]
>>
>> nodetool ring
>> -------------------
>>
>> Address DC Rack Status State Load
>> Effective-Ownership Token
>>
>> 159447687142037741049740936276011715300
>> 192.168.1.2 dc1 rack1 Up Normal 111.17 GB
>> 100.00% 67165620003619490909052924699950283577
>> 192.168.1.1 dc1 rack1 Up Normal 204.57 GB
>> 100.00% 71045951808949151217931264995073558408
>> 192.168.2.1 dc2 rack1 Up Normal 209.92 GB
>> 100.00% 107165019770579893816561717940612111506
>> 192.168.1.3 dc1 rack1 Up Normal 209.92 GB
>> 100.00% 114165363957966360026729000065495595953
>> 192.168.2.3 dc2 rack1 Up Normal 198.22 GB
>> 100.00% 147717787092318068320268200174271353451
>> 192.168.2.2 dc2 rack1 Up Normal 179.31 GB
>> 100.00% 159447687142037741049740936276011715300
>>
>> Does anyone have any ideas why every server does not have the same
>> amount of data on?
>>
>> Thanks,
>>
>> Owen


 -- 

Mime
View raw message