What is the consistency level you are using ?

And as Ed said, if you can provide the stacktrace that would help too.

On Sat, Apr 9, 2011 at 7:02 PM, aaron morton <aaron@thelastpickle.com> wrote:
btw, the nodes are a tad out of balance was that deliberate ? 

http://wiki.apache.org/cassandra/Operations#Token_selection
http://wiki.apache.org/cassandra/Operations#Load_balancing


Aaron

On 10 Apr 2011, at 08:44, Ed Anuff wrote:

Sounds like the problem might be on the hector side.  Lots of hector
users on this list, but usually not a bad idea to ask on
hector-users@googlegroups.com (cc'd).

The jetty servers stopping responding is a bit vague, somewhere in
your logs is an error message that should shed some light on where
things are going awry.  If you can find the exception that's being
thrown in hector and post that, it'd make it much easier to help you
out.

Ed

On Sat, Apr 9, 2011 at 12:11 PM, Vram Kouramajian
<vram.kouramajian@gmail.com> wrote:
The hector clients are used as part of our jetty servers. And, the
jetty servers stop responding when one of the Cassandra nodes go down.

Vram

On Sat, Apr 9, 2011 at 11:54 AM, Joe Stump <joe@joestump.net> wrote:
Did the Cassandra cluster go down or did you start getting failures from the client when it routed queries to the downed node? The key in the client is to keep working around the ring if the initial node is down.

--Joe

On Apr 9, 2011, at 12:52 PM, Vram Kouramajian wrote:

We have a 5 Cassandra nodes with the following configuration:

Casandra Version: 0.6.11
Number of Nodes: 5
Replication Factor: 3
Client: Hector 0.6.0-14
Write Consistency Level: Quorum
Read Consistency Level: Quorum
Ring Topology:
  Owns    Range                                      Ring

132756707369141912386052673276321963528
192.168.89.153Up         4.15 GB       33.87%
20237398133070283622632741498697119875     |<--|
192.168.89.155Up         5.17 GB       18.29%
51358066040236348437506517944084891398     |   ^
192.168.89.154Up         7.41 GB       33.97%
109158969152851862753910401160326064203    v   |
192.168.89.152Up         5.07 GB       6.34%
119944993359936402983569623214763193674    |   ^
192.168.89.151Up         4.22 GB       7.53%
132756707369141912386052673276321963528    |-->|

We believe that our setup should survive the crash of one of the
Cassandra nodes. But, we had few crashes and the system stopped
functioning until we brought back the Cassandra nodes.

Any clues?

Vram