hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Rasit OZDAS <rasitoz...@gmail.com>
Subject Re: empty log file...
Date Fri, 20 Feb 2009 08:22:18 GMT
I've looked at my datanode logs on the slaves, but they are all in quite
small sizes, although we've run many jobs on them.
And running 2 new jobs also didn't add anything to them.
(As I understand from the contents of the logs, hadoop logs especially
operations about DFS performance tests.)


2009/2/20 zander1013 <zander1013@gmail.com>

> hi,
> i am setting up hadoop for the first time on multi-node cluster. right now
> i
> have two nodes. the two node cluster consists of two laptops connected via
> ad-hoc wifi network. they they do not have access to the internet. i
> formated the datanodes on both machines prior to startup...
> output form the commands /usr/local/hadoop/bin/start-all.sh, jps (on both
> machines), and /usr/local/hadoop/bin/stop-all.sh all appear normal. however
> the file /usr/local/hadoop/logs/hadoop-hadoop-datanode-node1.log (the slave
> node) is empty.
> the same file for the master node shows the startup and shutdown events as
> normal and without error.
> is it okay that the log file on the slave is empty?
> zander
> --
> View this message in context:
> http://www.nabble.com/empty-log-file...-tp22113398p22113398.html
> Sent from the Hadoop core-user mailing list archive at Nabble.com.

M. Raşit ÖZDAŞ

  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message