cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Wei Zhu <>
Subject Re: How many BATCH inserts in to many?
Date Mon, 14 Jan 2013 04:31:29 GMT
Another potential issue is when some failure happens to some of the mutations. Is atomic batches
in 1.2 designed to resolve this?


----- Original Message -----
From: "aaron morton" <>
Sent: Sunday, January 13, 2013 7:57:56 PM
Subject: Re: How many BATCH inserts in to many?

With regard to a large number of records in a batch mutation there are some potential issues.

Each row becomes a task in the write thread pool on each replica. If a single client sends
1,000 rows in a mutation it will take time for the (default) 32 threads in the write pool
to work through the mutations. While they are doing this other clients / requests will appear
to be starved / stalled. 

There are also issues with the max message size in thrift and cql over thrift. 

IMHO as a rule of thumb dont go over a few hundred if you have a high number of concurrent


Aaron Morton 
Freelance Cassandra Developer 
New Zealand 


On 14/01/2013, at 12:56 AM, Radim Kolar < > wrote: 

do not use cassandra for implementing queueing system with high throughput. It does not scale
because of tombstone management. Use hornetQ, its amazingly fast broker but it has quite slow
persistence if you want to create queues significantly larger then your memory and use selectors
for searching for specific messages in them. 

My point is for implementing queue message broker is what you want. 

View raw message