cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Sebastian Estevez <sebastian.este...@datastax.com>
Subject Re: COPY command to export a table to CSV file
Date Mon, 20 Apr 2015 21:43:59 GMT
Try Brian's cassandra-unloader
<https://github.com/brianmhess/cassandra-loader#cassandra-unloader>

All the best,


[image: datastax_logo.png] <http://www.datastax.com/>

Sebastián Estévez

Solutions Architect | 954 905 8615 | sebastian.estevez@datastax.com

[image: linkedin.png] <https://www.linkedin.com/company/datastax> [image:
facebook.png] <https://www.facebook.com/datastax> [image: twitter.png]
<https://twitter.com/datastax> [image: g+.png]
<https://plus.google.com/+Datastax/about>
<http://feeds.feedburner.com/datastax>

<http://cassandrasummit-datastax.com/>

DataStax is the fastest, most scalable distributed database technology,
delivering Apache Cassandra to the world’s most innovative enterprises.
Datastax is built to be agile, always-on, and predictably scalable to any
size. With more than 500 customers in 45 countries, DataStax is the
database technology and transactional backbone of choice for the worlds
most innovative companies such as Netflix, Adobe, Intuit, and eBay.

On Mon, Apr 20, 2015 at 12:31 PM, Neha Trivedi <nehajtrivedi@gmail.com>
wrote:

> Does the nproc,nofile,memlock settings in
> /etc/security/limits.d/cassandra.conf are set to optimum value ?
> it's all default.
>
> What is the consistency level ?
> CL = Qurom
>
> Is there any other way to export a table to CSV?
>
> regards
> Neha
>
> On Mon, Apr 20, 2015 at 12:21 PM, Kiran mk <coolkiran2006@gmail.com>
> wrote:
>
>> Hi,
>>
>> Thanks for the info,
>>
>> Does the nproc,nofile,memlock settings in
>> /etc/security/limits.d/cassandra.conf are set to optimum value ?
>>
>> What is the consistency level ?
>>
>> Best Regardds,
>> Kiran.M.K.
>>
>>
>> On Mon, Apr 20, 2015 at 11:55 AM, Neha Trivedi <nehajtrivedi@gmail.com>
>> wrote:
>>
>>> hi,
>>>
>>> What is the count of records in the column-family ?
>>>       We have about 38,000 Rows in the column-family for which we are
>>> trying to export
>>> What  is the Cassandra Version ?
>>>      We are using Cassandra 2.0.11
>>>
>>> MAX_HEAP_SIZE and HEAP_NEWSIZE is the default .
>>> The Server is 8 GB.
>>>
>>> regards
>>> Neha
>>>
>>> On Mon, Apr 20, 2015 at 11:39 AM, Kiran mk <coolkiran2006@gmail.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> check  the MAX_HEAP_SIZE configuration in cassandra-env.sh environment
>>>> file
>>>>
>>>> Also HEAP_NEWSIZE ?
>>>>
>>>> What is the Consistency Level you are using ?
>>>>
>>>> Best REgards,
>>>> Kiran.M.K.
>>>>
>>>> On Mon, Apr 20, 2015 at 11:13 AM, Kiran mk <coolkiran2006@gmail.com>
>>>> wrote:
>>>>
>>>>> Seems like the is related to JAVA HEAP Memory.
>>>>>
>>>>> What is the count of records in the column-family ?
>>>>>
>>>>> What  is the Cassandra Version ?
>>>>>
>>>>> Best Regards,
>>>>> Kiran.M.K.
>>>>>
>>>>> On Mon, Apr 20, 2015 at 11:08 AM, Neha Trivedi <nehajtrivedi@gmail.com
>>>>> > wrote:
>>>>>
>>>>>> Hello all,
>>>>>>
>>>>>> We are getting the OutOfMemoryError on one of the Node and the Node
>>>>>> is down, when we run the export command to get all the data from
a table.
>>>>>>
>>>>>>
>>>>>> Regards
>>>>>> Neha
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> ERROR [ReadStage:532074] 2015-04-09 01:04:00,603 CassandraDaemon.java
>>>>>> (line 199) Exception in thread Thread[ReadStage:532074,5,main]
>>>>>> java.lang.OutOfMemoryError: Java heap space
>>>>>>         at
>>>>>> org.apache.cassandra.io.util.RandomAccessReader.readBytes(RandomAccessReader.java:347)
>>>>>>         at
>>>>>> org.apache.cassandra.utils.ByteBufferUtil.read(ByteBufferUtil.java:392)
>>>>>>         at
>>>>>> org.apache.cassandra.utils.ByteBufferUtil.readWithLength(ByteBufferUtil.java:355)
>>>>>>         at
>>>>>> org.apache.cassandra.db.ColumnSerializer.deserializeColumnBody(ColumnSerializer.java:124)
>>>>>>         at
>>>>>> org.apache.cassandra.db.OnDiskAtom$Serializer.deserializeFromSSTable(OnDiskAtom.java:85)
>>>>>>         at
>>>>>> org.apache.cassandra.db.Column$1.computeNext(Column.java:75)
>>>>>>         at
>>>>>> org.apache.cassandra.db.Column$1.computeNext(Column.java:64)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
>>>>>>         at
>>>>>> org.apache.cassandra.db.columniterator.SimpleSliceReader.computeNext(SimpleSliceReader.java:88)
>>>>>>         at
>>>>>> org.apache.cassandra.db.columniterator.SimpleSliceReader.computeNext(SimpleSliceReader.java:37)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
>>>>>>         at
>>>>>> org.apache.cassandra.db.columniterator.SSTableSliceIterator.hasNext(SSTableSliceIterator.java:82)
>>>>>>         at
>>>>>> org.apache.cassandra.db.columniterator.LazyColumnIterator.computeNext(LazyColumnIterator.java:82)
>>>>>>         at
>>>>>> org.apache.cassandra.db.columniterator.LazyColumnIterator.computeNext(LazyColumnIterator.java:59)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
>>>>>>         at
>>>>>> org.apache.cassandra.db.filter.QueryFilter$2.getNext(QueryFilter.java:157)
>>>>>>         at
>>>>>> org.apache.cassandra.db.filter.QueryFilter$2.hasNext(QueryFilter.java:140)
>>>>>>         at
>>>>>> org.apache.cassandra.utils.MergeIterator$OneToOne.computeNext(MergeIterator.java:200)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
>>>>>>         at
>>>>>> org.apache.cassandra.db.filter.SliceQueryFilter.collectReducedColumns(SliceQueryFilter.java:185)
>>>>>>         at
>>>>>> org.apache.cassandra.db.filter.QueryFilter.collateColumns(QueryFilter.java:122)
>>>>>>         at
>>>>>> org.apache.cassandra.db.filter.QueryFilter.collateOnDiskAtom(QueryFilter.java:80)
>>>>>>         at
>>>>>> org.apache.cassandra.db.RowIteratorFactory$2.getReduced(RowIteratorFactory.java:101)
>>>>>>         at
>>>>>> org.apache.cassandra.db.RowIteratorFactory$2.getReduced(RowIteratorFactory.java:75)
>>>>>>         at
>>>>>> org.apache.cassandra.utils.MergeIterator$ManyToOne.consume(MergeIterator.java:115)
>>>>>>         at
>>>>>> org.apache.cassandra.utils.MergeIterator$ManyToOne.computeNext(MergeIterator.java:98)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143)
>>>>>>         at
>>>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138)
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best Regards,
>>>>> Kiran.M.K.
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Best Regards,
>>>> Kiran.M.K.
>>>>
>>>
>>>
>>
>>
>> --
>> Best Regards,
>> Kiran.M.K.
>>
>
>

Mime
View raw message