hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Billy Pearson" <sa...@pearsonwholesale.com>
Subject Re: completebulkload problems
Date Thu, 03 May 2012 05:21:45 GMT
Looks like it was a time out issue I seen the bonding log messages more then 
once so.

Does anyone know what the timeout setting name is on completebulkload?

Billy



"Billy Pearson" <sales@pearsonwholesale.com> 
wrote in message news:jnslmk$9ic$1@dough.gmane.org...
> ok I got a MR job I am trying to import into hbase it works with small 
> input loads with in secs and the command line returns
> when  Iadd more input to the map reduce job the completebulkload  hangs on 
> the command line never returning.
>
> when I run a large completebulkload it keeps trying to copy the input 
> files many times
>
> cut out of regionserver log
>
> 2012-05-02 18:56:11,629 INFO org.apache.hadoop.hbase.regionserver.Store: 
> File hdfs://node1/bulkoutput/idjuice/436678602442940864 on different 
> filesystem than destination store - moving to this filesystem.
> 2012-05-02 18:56:12,287 INFO org.apache.hadoop.hbase.regionserver.Store: 
> Copied to temporary path on dst filesystem: 
> hdfs://node1.hadoop.compspy.com/hbase/Repo/751e4b3c4a27e680e8d481be3e11507e/.tmp/6957718427566497105
> 2012-05-02 18:56:12,288 INFO org.apache.hadoop.hbase.regionserver.Store: 
> Renaming bulk load file 
> hdfs://node1.hadoop.compspy.com/hbase/Repo/751e4b3c4a27e680e8d481be3e11507e/.tmp/6957718427566497105

> to 
> hdfs://node1.hadoop.compspy.com/hbase/Repo/751e4b3c4a27e680e8d481be3e11507e/idjuice/499207355457162362
> 2012-05-02 18:56:12,310 INFO org.apache.hadoop.hbase.regionserver.Store: 
> Moved hfile 
> hdfs://node1.hadoop.compspy.com/hbase/Repo/751e4b3c4a27e680e8d481be3e11507e/.tmp/6957718427566497105

> into store directory 
> hdfs://node1.hadoop.compspy.com/hbase/Repo/751e4b3c4a27e680e8d481be3e11507e/idjuice 
>  - updating store file list.
>
> I can see the same input file load over and over in the logs
>
> The final size with TableMapReduceUtil.initTableReducerJob of the table 
> was around 680m on all the stores total reported by the regionserver
> I ran the import on a larger dataset same data format same mr job just 
> more input and I killed the regionserver and  truncate the table and 
> restart the whole cluster
> the after the table gets to 5GB
>
>
> the mapreduce job maps output is set by 
> HFileOutputFormat.configureIncrementalLoad(job, table);
>
>
> the only difference I can see in the two sets of files is one has 1 file 
> to import per column family and the larget one has 8 file for each column 
> family.
>
> Any suggestions on where the problem could be?
> kind of odd that the small data input would work so easy and the large one 
> would just run out of control.
>
> running
> hbase-0.90.4-cdh3u3.jar
> export HBASE_CLASSPATH=/etc/hbase/conf:/etc/hadoop/conf:/etc/zookeeper
>
>
>
> 


Mime
View raw message