hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Gang Luo <lgpub...@yahoo.com.cn>
Subject Re: Several questions about Hadoop
Date Sun, 06 Jun 2010 13:10:17 GMT
1. try to use JobClient.submitJob(JobConf). It will submit the job to hadoop without waiting
for its completion. 

2. no

3. the task running on the nodes which fail in the fly will be re-scheduled on other nodes.
The incomplete result will not be used.


----- 原始邮件 ----
发件人: Kevin Tse <kevintse.onjee@gmail.com>
收件人: common-user@hadoop.apache.org
发送日期: 2010/6/5 (周六) 10:00:32 下午
主   题: Several questions about Hadoop

I have successfully run Hadoop on a cluster of 3 nodes on RedHat linux, I
have several questions to ask.

1. When I submit an MR job using "hadoop jar mr-job.jar", it starts printing
out log messages to the stdout, how do I make it run in the
background?("hadoop jar mr-job.jar > log &" does not work). If it can be put
in the background, where do I find those log messages that it used to print
to the stdout?
2. While the MR job is being executed, will the MR job process be
affected/killed if I press "Ctrl-c"? it seems not since I can see the
tasktracker is still running, but I am not sure.
3. While the MR job is being executed, if I stop one of the
tasktrackers/nodes in the cluster using hadoop-daemon.sh, will the result of
the maps and reduces semi-completed by that tasktracker/node be submitted to
the namenode to be merged with the results completed by other tasktrackers?
Is it possible to restart a tasktracker at a point where it was stopped?

Thank you in advance.
- Kevin Tse


View raw message