cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Jonathan Ellis <>
Subject Re: Cassandra OOM - 1.0.2
Date Sat, 04 Feb 2012 15:03:39 GMT
Sounds like you need to increase sysctl vm.max_map_count

On Fri, Feb 3, 2012 at 7:27 PM, Ajeet Grewal <> wrote:
> Hey guys,
> I am getting an out of memory (mmap failed) error with Cassandra
> 1.0.2. The relevant log lines are pasted at
> Cassandra works fine until it reaches about 300-400GB of load (on one
> instance, I have 12 nodes RF=2). Then nodes start failing with such
> errors. The nodes are pretty beefy, 32GB of ram, 8 cores. Increasing
> the JVM heap size does not help.
> I am running on a 64bit jvm. I am using jna. I have memlock unlimited
> for the user. (I confirmed this by looking at /proc/<pid>/limits).
> I also tried restarting the process as root, but it crashes with the same error.
> Also the number of files that I have in the data directory is about
> ~300, so it should not be exceeding the open files limit.
> I don't know if this is relevant. I just have two column families,
> counter_object and counter_time. I am using very wide columns, so row
> sizes can be huge. You can see from the log link, that the *.db files
> are sometimes pretty big.
> Please help! Thank you!
> --
> Regards,
> Ajeet

Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of DataStax, the source for professional Cassandra support

View raw message