cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From charles THIBAULT <>
Subject low performance inserting
Date Tue, 03 May 2011 10:30:24 GMT
Hello everybody,

first: sorry for my english in advance!!

I'm getting started with Cassandra on a 5 nodes cluster inserting data
with the pycassa API.

I've read everywere on internet that cassandra's performance are better than
because of the writes append's only into commit logs files.

When i'm trying to insert 100 000 rows with 10 columns per row with batch
insert, I'v this result: 27 seconds
But with MySQL (load data infile) this take only 2 seconds (using indexes)

Here my configuration

cassandra version: 0.7.5
nodes :,,,,

My script
#!/usr/bin/env python

import pycassa
import time
import random
from cassandra import ttypes

pool = pycassa.connect('test', [''])
cf = pycassa.ColumnFamily(pool, 'test')
b = cf.batch(queue_size=50,

tps1 = time.time()
for i in range(100000):
    columns = dict()
    for j in range(10):
        columns[str(j)] = str(random.randint(0,100))
    b.insert(str(i), columns)
tps2 = time.time()

print("execution time: " + str(tps2 - tps1) + " seconds")

what I'm doing rong ?

View raw message