cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Шамим <>
Subject Re: Filter data on row key in Cassandra Hadoop's Random Partitioner
Date Wed, 12 Dec 2012 08:45:50 GMT
You can use Apache PIG to load data and filter it by row key, filter in pig is very fast.

11.12.2012, 20:46, "Ayush V." <>:
> I'm working on Cassandra Hadoop intergration (MapReduce). We have used Random
> Partioner to insert data to gain faster write. Now we have to read that data
> from cassandra in MapReduce and perform some calculation on it.
> From the lots of data we have in cassandra we wan't to fetch data only for
> particular ROW-KEYs but we are unable to do it due to RandomPartioner -
> assertion is there in code.
> Can anyone please guide me how should I filter data based on RowKey on
> Cassandra level itself (I know data is distributed across regions using Hash
> of the RowKey)?
> Does using secondary indexes (still trying to understand how it works) will
> solve my problem or is there some other way around?
> I will be really appreciated if someone could answer my queries.
> Thanks
> AV
> --
> View this message in context:
> Sent from the mailing list archive at

View raw message