giraph-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jerome richard <>
Subject Scaling Problem
Date Fri, 26 Jul 2013 12:02:25 GMT

I encountered a critical scaling problem using Giraph. I made a very simple algorithm to test
Giraph on large graphs : a connexity test. It works on relatively large graphs (3 072 441
nodes and 117 185 083 edges) but not on very large graph (52 000 000 nodes and 2 000 000 000
edges). In fact, during the processing of the biggest graph, Giraph core seems to fail after
the superstep 14 (15 on some jobs). The input graph size is 30 GB stored as text and the output
is also stored as text. 9 working jobs are used to compute the graph.
Here is the tracktrace of jobs (this is the same for the 9 jobs):    java.lang.IllegalStateException:
run: Caught an unrecoverable exception exists: Failed to check /_hadoopBsp/job_201307260439_0006/_applicationAttemptsDir/0/_superstepDir/97/_addressesAndPartitions
after 3 tries!        at   
    at org.apache.hadoop.mapred.MapTask.runNewMapper(        at
       at org.apache.hadoop.mapred.Child$        at
Method)        at Source)        at
       at org.apache.hadoop.mapred.Child.main(    Caused by: java.lang.IllegalStateException:
exists: Failed to check /_hadoopBsp/job_201307260439_0006/_applicationAttemptsDir/0/_superstepDir/97/_addressesAndPartitions
after 3 tries!        at org.apache.giraph.zk.ZooKeeperExt.exists( 
      at org.apache.giraph.worker.BspServiceWorker.startSuperstep(
       at org.apache.giraph.graph.GraphTaskManager.execute(    
   at        ... 7 more
Could you help me to solve this problem?If you need the code of the program, I can put that
here (the code is relatively tiny).
Thanks, Jérôme.
View raw message