incubator-cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From cem <cayiro...@gmail.com>
Subject Re: maximum storage per node
Date Fri, 26 Jul 2013 08:20:18 GMT
I dont think it is a good idea to put multiple instance in same machine.
You may lose multiple instances at the same time if the machine goes
down. You can also specify multiple directories as storage in 1.2.

I am also not sure boot-strapping will be a big problem since the number
keys you will store is relatively small.

Why didnt you partition your data according to time instead of using your
own compactor?

Cem



On Fri, Jul 26, 2013 at 3:50 AM, sankalp kohli <kohlisankalp@gmail.com>wrote:

> Try putting multiple instances per machine with each instance mapped to
> its own disk. This might not work with v-nodes
>
>
> On Thu, Jul 25, 2013 at 9:04 AM, Pruner, Anne (Anne) <pruner@avaya.com>wrote:
>
>>  I actually wrote my own compactor that deals with this problem.****
>>
>> ** **
>>
>> Anne****
>>
>> ** **
>>
>> *From:* cem [mailto:cayiroglu@gmail.com]
>> *Sent:* Thursday, July 25, 2013 11:59 AM
>>
>> *To:* user@cassandra.apache.org
>> *Subject:* Re: maximum storage per node****
>>
>> ** **
>>
>> You will suffer from long compactions if you are planning to get rid of
>> from old records by TTL.****
>>
>> ** **
>>
>> Best Regards,****
>>
>> Cem.****
>>
>> ** **
>>
>> On Thu, Jul 25, 2013 at 5:51 PM, Kanwar Sangha <kanwar@mavenir.com>
>> wrote:****
>>
>> Issues with large data nodes would be –****
>>
>>  ****
>>
>> ·         Nodetool repair will be impossible to run****
>>
>> ·         Your read i/o will suffer since you will almost always go to
>> disk (each read will take 3 IOPS worst case)****
>>
>> ·         Boot-straping the node in case of failures will take days/weeks
>> ****
>>
>>  ****
>>
>>  ****
>>
>> *From:* Pruner, Anne (Anne) [mailto:pruner@avaya.com]
>> *Sent:* 25 July 2013 10:45
>> *To:* user@cassandra.apache.org
>> *Subject:* RE: maximum storage per node****
>>
>>  ****
>>
>> We’re storing fairly large files (about 1MB apiece) for a few months and
>> then deleting the oldest to get more space to add new ones.  We have large
>> requirements (maybe up to 100 TB), so having a 1TB limit would be
>> unworkable.****
>>
>>  ****
>>
>> What is the reason for the limit?  Does something fail after that?****
>>
>>  ****
>>
>> If there are hardware issues, what’s recommended?****
>>
>>  ****
>>
>> BTW, we’re using Cassandra 1.2****
>>
>>  ****
>>
>> Anne****
>>
>>  ****
>>
>> *From:* cem [mailto:cayiroglu@gmail.com <cayiroglu@gmail.com>]
>> *Sent:* Thursday, July 25, 2013 11:41 AM
>> *To:* user@cassandra.apache.org
>> *Subject:* Re: maximum storage per node****
>>
>>  ****
>>
>> Between 500GB - 1TB is recommended. ****
>>
>>  ****
>>
>> But it depends also your hardware, traffic characteristics and
>> requirements. Can you give some details on that?****
>>
>>  ****
>>
>> Best Regards,****
>>
>> Cem****
>>
>>  ****
>>
>> On Thu, Jul 25, 2013 at 5:35 PM, Pruner, Anne (Anne) <pruner@avaya.com>
>> wrote:****
>>
>> Does anyone have opinions on the maximum amount of data reasonable to
>> store on one Cassandra node?  If there are limitations, what are the
>> reasons for it?****
>>
>>  ****
>>
>> Thanks,****
>>
>> Anne****
>>
>>  ****
>>
>> ** **
>>
>
>

Mime
View raw message