hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Olga Natkovich" <ol...@yahoo-inc.com>
Subject RE: How can I tell that I have "permission denied" error returned to me
Date Tue, 25 Mar 2008 00:44:49 GMT
Doug, thanks. Would what I proposed be reasonable short term?

Olga

> -----Original Message-----
> From: Doug Cutting [mailto:cutting@apache.org] 
> Sent: Monday, March 24, 2008 5:40 PM
> To: core-dev@hadoop.apache.org
> Subject: Re: How can I tell that I have "permission denied" 
> error returned to me
> 
> This is a shortcoming of Hadoop RPC.  Ideally exceptions 
> thrown on the server would be re-thrown on the client, but 
> the concern is that their class might not exist there, so we 
> instead transmit the just class name and the error string and 
> do not attempt to re-throw the original exception and instead 
> throw a RemoteException.  HDFS could patch around this, but 
> it would really be best to fix it in the RPC layer.
> 
> Doug
> 
> Olga Natkovich wrote:
> > Hi,
> >  
> > In my code, I want to be able to differentiate access 
> control problems
> > and give a meaningful message to the users. I noticed that 
> in this case
> > org.apache.hadoop.fs.permission.AccessControlException is thrown but
> > then it gets wrapped into other exceptions such as 
> java.io.IOException
> > or org.apache.hadoop.ipc.RemoteException. 
> >  
> > One way to figure it out is to recursively check for the 
> cause of the
> > exception to be of type
> > org.apache.hadoop.fs.permission.AccessControlException. Is this the
> > right/best way to go about it?
> >  
> > Thanks,
> >  
> > Olga
> > 
> 
> 

Mime
View raw message