hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From slitz <slitzferr...@gmail.com>
Subject Using NFS without HDFS
Date Fri, 11 Apr 2008 11:40:22 GMT
I'm trying to assemble a simple setup of 3 nodes using NFS as Distributed

Box A:, this box is either the NFS server and working as a slave
Box B:, this box is only JobTracker
Box C:, this box is only slave

Obviously all three nodes can access the NFS shared, and the path to the
share is /home/slitz/warehouse in all three.

My hadoop-site.xml file were copied over all nodes and looks like this:






 The name of the default file system. Either the literal string

"local" or a host:port for NDFS.







 The host and port that the MapReduce job

tracker runs at. If "local", then jobs are

 run in-process as a single map and reduce task.









As one can see, i'm not using HDFS at all.
(Because all the free space i have is located in only one node, so using
HDFS would be unnecessary overhead)

I've copied the input folder from hadoop to /home/slitz/warehouse/input.
When i try to run the example line

bin/hadoop jar hadoop-*-examples.jar grep /home/slitz/warehouse/input/
/home/slitz/warehouse/output 'dfs[a-z.]+'

the job starts and finish okay but at the end i get this error:

org.apache.hadoop.mapred.InvalidInputException: Input path doesn't exist :
at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:508)
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:753)
(...the error stack continues...)

i don't know why the input path being looked is in the local path
/home/slitz/hadoop(...) instead of /home/slitz/warehouse/(...)

Maybe something is missing in my hadoop-site.xml?


  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message