hbase-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Harsh J <ha...@cloudera.com>
Subject Re: HBase Security Configuration
Date Tue, 17 Apr 2012 12:28:21 GMT
Hey Konrad,

Make sure your HBase's classpath also has the Hadoop conf dir on it
(specifically hdfs-site.xml and core-site.xml). It it already does
have that, make sure they are populated with the right HDFS cluster
values (core-site needs two properties that toggle security ON, and
hdfs-site needs the HDFS server principals configured inside it -
basically just copy these core-site and hdfs-site files from your
secured HDFS cluster config over to the HBase machines/classpath).

On Tue, Apr 17, 2012 at 5:38 PM, Konrad Tendera <emails@tendera.eu> wrote:
> Hello,
> I'm trying to configure secure HBase using following instruction: https://ccp.cloudera.com/display/CDHDOC/HBase+Security+Configuration.
Our cluster uses Kerberos and everything in Hadoop work fine. But when I start HBase following
exception is thrown
>
> FATAL org.apache.hadoop.hbase.master.HMaster: Unhandled exception. Starting shutdown.
> org.apache.hadoop.security.AccessControlException: Authentication is required
>        at org.apache.hadoop.ipc.Client.call(Client.java:1028)
>        at org.apache.hadoop.ipc.WritableRpcEngine$Invoker.invoke(WritableRpcEngine.java:198)
>        at $Proxy9.getProtocolVersion(Unknown Source)
>        at org.apache.hadoop.ipc.WritableRpcEngine.getProxy(WritableRpcEngine.java:235)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:275)
>        at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:249)
>        at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:161)
>        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:278)
>        at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:245)
>        at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:109)
>        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1792)
>        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:76)
>        at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:1826)
>        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1808)
>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:265)
>        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:189)
>        at org.apache.hadoop.hbase.util.FSUtils.getRootDir(FSUtils.java:471)
>        at org.apache.hadoop.hbase.master.MasterFileSystem.<init>(MasterFileSystem.java:94)
>        at org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:448)
>        at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:326)
>        at java.lang.Thread.run(Thread.java:662)
>
> I can't find any info about it. I'm using Hbase 0.92 with Hadoop 0.22
>
> --
> Konrad Tendera



-- 
Harsh J

Mime
View raw message