On Wed, Feb 3, 2010 at 4:19 AM, envio user <firstname.lastname@example.org> wrote:
After this I tried with 1 million keys:
/home/sun>python stress.py -n 1000000 -t 100 -c 25 -r -o read -i 10
WARNING: multiprocessing not present, threading will be used........................
Benchmark may not be accurate!
1M rows means you've stored 600M columns, which is around 32G of data after compaction. With 8G of memory for disk cache from the OS, minus at least 1G going to Cassandra's JVM, your machine is going to have to do a disk seek on at least nearly 80% of your reads. You can improve this with either more memory per node, or more nodes, but it's worth noting that ~1875 columns/sec isn't too bad for this situation, and you can probably read all 600 columns per row at nearly the same speed.