incubator-cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Edward Capriolo <edlinuxg...@gmail.com>
Subject Re: get_count - cassandra 0.7.x predicate limit bug?
Date Tue, 30 Nov 2010 18:13:22 GMT
On Tue, Nov 30, 2010 at 1:00 AM, Tyler Hobbs <tyler@riptano.com> wrote:
> What error are you getting?
>
> Remember, get_count() is still just about as much work for cassandra as
> getting the whole row; the only advantage is it doesn't have to send the
> whole row back to the client.
>
> If you're counting 3+ million columns frequently, it's time to take a look
> at counters.
>
> - Tyler
>
> On Fri, Nov 26, 2010 at 10:33 AM, Marcin <marcin@33concept.com> wrote:
>>
>> Hi guys,
>>
>> I have a key with 3million+ columns but when I am trying to run get_count
>> on it its getting me error if setting limit more than 46000+ any ideas?
>>
>> In previous API there was no predicate at all so it was simply counting
>> number of columns now its not so simple any more.
>>
>> Please let me know if that is a bug or I do something wrong.
>>
>>
>> cheers,
>> /Marcin
>
>

+1 Tyler. The problem is you can increase the clients socket timeout
as high as you like if socketTimeout < rpcTimeout you should see
SocketTimeoutExceptions if socketTimeout >= rcpTimeout you start
seeing Cassandra TimedOutExceptions. Raising the RPC Timeout is done
on the server. In any case you may have to range_slice to get through
a row this big and count. Also in my experience rows this large do not
work well. They are particularly dangerous when combined with RowCache
as bringing them into to memory and evicting them is both disk and
memory intensive.

Mime
View raw message