hadoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Michael Segel <michael_se...@hotmail.com>
Subject Re: using hadoop on zLinux (Linux on S390)
Date Tue, 11 Dec 2012 14:31:57 GMT
Well, on the surface.... 

It looks like its either a missing class, or you don't have your class path set up right.


I'm assuming you got this version of Hadoop from IBM, so I would suggest contacting their
support and opening up a ticket. 


On Dec 11, 2012, at 8:23 AM, Emile Kao <emilekao@gmx.net> wrote:

> Hello community,
> I am trying to use hadoop 1.1.0 on a SLES 11 (zLinux) running on IBM S390.
> The java provided is "java-s390x-60" 64Bit.
> While trying to format the namenode I got the following error:
> 
> $:/opt/flume_hadoop/hadoop-1.1.0> bin/hadoop namenode -format
> 12/12/11 14:16:31 INFO namenode.NameNode: STARTUP_MSG:
> /************************************************************
> STARTUP_MSG: Starting NameNode
> STARTUP_MSG:   host = xxxxxxxxx
> STARTUP_MSG:   args = [-format]
> STARTUP_MSG:   version = 1.1.0
> STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.1
-r 1394289; compiled by 'hortonfo' on Thu Oct  4 22:06:49 UTC 2012
> ************************************************************/
> Re-format filesystem in /opt/hadoop_data/name ? (Y or N) Y
> 12/12/11 14:16:34 INFO util.GSet: VM type       = 64-bit
> 12/12/11 14:16:34 INFO util.GSet: 2% max memory = 20.0 MB
> 12/12/11 14:16:34 INFO util.GSet: capacity      = 2^21 = 2097152 entries
> 12/12/11 14:16:34 INFO util.GSet: recommended=2097152, actual=2097152
> 12/12/11 14:16:34 ERROR security.UserGroupInformation: Unable to find JAAS classes:com.ibm.security.auth.LinuxPrincipal
> 12/12/11 14:16:35 ERROR namenode.NameNode: java.io.IOException: failure to login
>        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:501)
>        at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:463)
>        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setConfigurationParameters(FSNamesystem.java:491)
>        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:480)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1198)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1391)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1412)
> Caused by: javax.security.auth.login.LoginException: java.lang.NullPointerException:
invalid null Class provided
>        at javax.security.auth.Subject.getPrincipals(Subject.java:809)
>        at org.apache.hadoop.security.UserGroupInformation$HadoopLoginModule.getCanonicalUser(UserGroupInformation.java:86)
>        at org.apache.hadoop.security.UserGroupInformation$HadoopLoginModule.commit(UserGroupInformation.java:123)
>        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:48)
>        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>        at java.lang.reflect.Method.invoke(Method.java:600)
>        at javax.security.auth.login.LoginContext.invoke(LoginContext.java:795)
>        at javax.security.auth.login.LoginContext.access$000(LoginContext.java:209)
>        at javax.security.auth.login.LoginContext$5.run(LoginContext.java:732)
>        at java.security.AccessController.doPrivileged(AccessController.java:284)
>        at javax.security.auth.login.LoginContext.invokeCreatorPriv(LoginContext.java:729)
>        at javax.security.auth.login.LoginContext.login(LoginContext.java:600)
>        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:482)
>        at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:463)
>        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setConfigurationParameters(FSNamesystem.java:491)
>        at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:480)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1198)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1391)
>        at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1412)
> 
>        at javax.security.auth.login.LoginContext.invoke(LoginContext.java:898)
>        at javax.security.auth.login.LoginContext.access$000(LoginContext.java:209)
>        at javax.security.auth.login.LoginContext$5.run(LoginContext.java:732)
>        at java.security.AccessController.doPrivileged(AccessController.java:284)
>        at javax.security.auth.login.LoginContext.invokeCreatorPriv(LoginContext.java:729)
>        at javax.security.auth.login.LoginContext.login(LoginContext.java:600)
>        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:482)
>        ... 6 more
> 
> 12/12/11 14:16:35 INFO namenode.NameNode: SHUTDOWN_MSG:
> /************************************************************
> SHUTDOWN_MSG: Shutting down NameNode at xxxxxxxxxxxxxxxx
> ************************************************************/
> $:/opt/flume_hadoop/hadoop-1.1.0>
> 
> Question:
> 
> 1)@developer
> Are you aware of this behavior?
> 2)It there a way to overcome this problem with a workaround?
> 3)IS it a security issue? --> I was able to issue ssh on localhost without error.
> 


Mime
View raw message