cassandra-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Joost Ouwerkerk <jo...@openplaces.org>
Subject Re: Help with MapReduce
Date Tue, 20 Apr 2010 01:34:52 GMT
And when retrieving only one supercolumn?  Can I further specify which
subcolumns to retrieve?

On Mon, Apr 19, 2010 at 9:29 PM, Jonathan Ellis <jbellis@gmail.com> wrote:

> the latter, if you are retrieving multiple supercolumns.
>
> On Mon, Apr 19, 2010 at 8:10 PM, Joost Ouwerkerk <joost@openplaces.org>
> wrote:
> > hmm, might be too much data.  In the case of a supercolumn, how do I
> specify
> > which sub-columns to retrieve?  Or can I only retrieve entire
> supercolumns?
> > On Mon, Apr 19, 2010 at 8:47 PM, Jonathan Ellis <jbellis@gmail.com>
> wrote:
> >>
> >> Possibly you are asking it to retrieve too many columns per row.
> >>
> >> Possibly there is something else causing poor performance, like
> swapping.
> >>
> >> On Mon, Apr 19, 2010 at 7:12 PM, Joost Ouwerkerk <joost@openplaces.org>
> >> wrote:
> >> > I'm slowly getting somewhere with Cassandra... I have successfully
> >> > imported
> >> > 1.5 million rows using MapReduce.  This took about 8 minutes on an
> >> > 8-node
> >> > cluster, which is comparable to the time it takes with HBase.
> >> > Now I'm having trouble scanning this data.  I've created a simple
> >> > MapReduce
> >> > job that counts rows in my ColumnFamily.  The Job fails with most
> tasks
> >> > throwing the following Exception.  Anyone have any ideas what's going
> >> > wrong?
> >> > java.lang.RuntimeException: TimedOutException()
> >> >
> >> >       at
> >> >
> >> >
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$RowIterator.maybeInit(ColumnFamilyRecordReader.java:165)
> >> >       at
> >> >
> >> >
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$RowIterator.computeNext(ColumnFamilyRecordReader.java:215)
> >> >       at
> >> >
> >> >
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$RowIterator.computeNext(ColumnFamilyRecordReader.java:97)
> >> >       at
> >> >
> >> >
> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:135)
> >> >       at
> >> >
> >> >
> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:130)
> >> >       at
> >> >
> >> >
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader.nextKeyValue(ColumnFamilyRecordReader.java:91)
> >> >       at
> >> >
> >> >
> org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.nextKeyValue(MapTask.java:423)
> >> >       at
> >> >
> org.apache.hadoop.mapreduce.MapContext.nextKeyValue(MapContext.java:67)
> >> >       at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:143)
> >> >       at
> org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:583)
> >> >       at org.apache.hadoop.mapred.MapTask.run(MapTask.java:305)
> >> >       at org.apache.hadoop.mapred.Child.main(Child.java:170)
> >> > Caused by: TimedOutException()
> >> >       at
> >> >
> >> >
> org.apache.cassandra.thrift.Cassandra$get_range_slices_result.read(Cassandra.java:11015)
> >> >       at
> >> >
> >> >
> org.apache.cassandra.thrift.Cassandra$Client.recv_get_range_slices(Cassandra.java:623)
> >> >       at
> >> >
> >> >
> org.apache.cassandra.thrift.Cassandra$Client.get_range_slices(Cassandra.java:597)
> >> >       at
> >> >
> >> >
> org.apache.cassandra.hadoop.ColumnFamilyRecordReader$RowIterator.maybeInit(ColumnFamilyRecordReader.java:142)
> >> >       ... 11 more
> >> >
> >> > On Sun, Apr 18, 2010 at 6:01 PM, Stu Hood <stu.hood@rackspace.com>
> >> > wrote:
> >> >>
> >> >> In 0.6.0 and trunk, it is located at
> >> >> src/java/org/apache/cassandra/hadoop/ColumnFamilyInputFormat.java
> >> >>
> >> >> You might be using a pre-release version of 0.6 if you are seeing a
> fat
> >> >> client based InputFormat.
> >> >>
> >> >>
> >> >> -----Original Message-----
> >> >> From: "Joost Ouwerkerk" <joost@openplaces.org>
> >> >> Sent: Sunday, April 18, 2010 4:53pm
> >> >> To: user@cassandra.apache.org
> >> >> Subject: Re: Help with MapReduce
> >> >>
> >> >> Where is the ColumnFamilyInputFormat that uses Thrift?  I don't
> >> >> actually
> >> >> have a preference about client, I just want to be consistent with
> >> >> ColumnInputFormat.
> >> >>
> >> >> On Sun, Apr 18, 2010 at 5:37 PM, Stu Hood <stu.hood@rackspace.com>
> >> >> wrote:
> >> >>
> >> >> > ColumnFamilyInputFormat no longer uses the fat client API, and
> >> >> > instead
> >> >> > uses
> >> >> > Thrift. There are still some significant problems with the fat
> >> >> > client,
> >> >> > so it
> >> >> > shouldn't be used without a good understanding of those problems.
> >> >> >
> >> >> > If you still want to use it, check out contrib/bmt_example, but
I'd
> >> >> > recommend that you use thrift for now.
> >> >> >
> >> >> > -----Original Message-----
> >> >> > From: "Joost Ouwerkerk" <joost@openplaces.org>
> >> >> > Sent: Sunday, April 18, 2010 2:59pm
> >> >> > To: user@cassandra.apache.org
> >> >> > Subject: Help with MapReduce
> >> >> >
> >> >> > I'm a Cassandra noob trying to validate Cassandra as a viable
> >> >> > alternative
> >> >> > to
> >> >> > HBase (which we've been using for over a year) for our application.
> >> >> >  So
> >> >> > far,
> >> >> > I've had no success getting Cassandra working with MapReduce.
> >> >> >
> >> >> > My first step is inserting data into Cassandra.  I've created
a
> >> >> > MapRed
> >> >> > job
> >> >> > based using the fat client API.  I'm using the fat client
> >> >> > (StorageProxy)
> >> >> > because that's what ColumnFamilyInputFormat uses and I want to
use
> >> >> > the
> >> >> > same
> >> >> > API for both read and write jobs.
> >> >> >
> >> >> > When I call StorageProxy.mutate(), nothing happens.  The job
> >> >> > completes
> >> >> > as
> >> >> > if
> >> >> > it had done something, but in fact nothing has changed in the
> >> >> > cluster.
> >> >> >  When
> >> >> > I call StorageProxy.mutateBlocking(), I get an IOException
> >> >> > complaining
> >> >> > that
> >> >> > there is no connection to the cluster.  I've concluded with the
> >> >> > debugger
> >> >> > that StorageService is not connecting to the cluster, even though
> >> >> > I've
> >> >> > specified the correct seed and ListenAddress (I've using the exact
> >> >> > same
> >> >> > storage-conf.xml as the nodes in the cluster).
> >> >> >
> >> >> > I'm sure I'm missing something obvious in the configuration or
my
> >> >> > setup,
> >> >> > but
> >> >> > since I'm new to Cassandra, I can't see what it is.
> >> >> >
> >> >> > Any help appreciated,
> >> >> > Joost
> >> >> >
> >> >> >
> >> >> >
> >> >>
> >> >>
> >> >
> >> >
> >
> >
>

Mime
View raw message