hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Vijay Kumar Adhikari <vijay...@gmail.com>
Subject Re: hadoop dfs -ls shows local files
Date Fri, 17 Jul 2009 20:55:01 GMT
Thanks, Todd. That gave a hint of what exactly was wrong. The  
permission on conf directory was wrong.

On Jul 17, 2009, at 1:38 PM, Todd Lipcon wrote:

> Hi Vijay,
>
> It sounds like your user is seeing a different configuration than  
> the hadoop
> user, such that fs.default.name isn't getting set. Did you set a
> HADOOP_CONF_DIR environment variable in the bashrc of the user that  
> runs the
> hadoop process? If so, you should instead set that variable in  
> hadoop-env.sh
>
> -Todd
>
>
> On Fri, Jul 17, 2009 at 11:09 AM, Vijay Kumar Adhikari
> <vijaykcm@gmail.com>wrote:
>
>> I have an installation of hadoop where when I log into the account
>> that runs the hadoop process, everything looks fine. I can copy local
>> files to the dfs, I can view the files inside dfs etc etc...
>>
>> When I log into a different user, I can still run all the dfs  
>> commands
>> but dfs shows files from my current local directory. If I am inside  
>> my
>> home directory in the local machine, and issue "hadoop dfs -ls" it
>> lists all the files inside my local home directory. If I change to
>> some other local directory and issue the same command all files from
>> that directory are shown. When I issue "hadoop dfs -copyFromLocal
>> <filename>", it says the uri already exists.
>>
>> What is wrong here? What do I need to fix?
>>
>> --
>> Vijay
>>


Mime
View raw message