hadoop-hdfs-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Björn-Elmar Macek <ma...@cs.uni-kassel.de>
Subject Re: Hadoop Configuration Issues
Date Fri, 27 Apr 2012 10:39:05 GMT
Hi Alex,

as i have written, i already did so! The problem is as already stated in 
my mail before, that all the Variables ${bla}  seem to be UNSET - not 
only SECURITY_TYPE. As i dont really understand those parameters, i 
would like to use the default ones, which afaik should be configured in 
the hadoop-env.sh. But obviously they are not.

Best,
Björn

Am 27.04.2012 12:12, schrieb alo alt:
> Hi,
>
> Invalid attribute value for hadoop.security.authentication of ${SECURITY_TYPE}
> Set it to simple and it should work (default is kerberos).
>
> - Alex
>
> --
> Alexander Lorenz
> http://mapredit.blogspot.com
>
> On Apr 27, 2012, at 12:01 PM, Björn-Elmar Macek wrote:
>
>> Hello,
>>
>> i have recently installed Hadoop on my and a second machine in order to test the
setup and develop little programs locally before deploying them to the cluster. I stumbled
over several difficulties, which i could fix with some internet research. But once again im
stuck and i think i can nail the problem down:
>>
>> When Hadoop evaluates the config files in /etc/hadoop it does not have any default
values for all the variables used within:
>>
>> \________ First Error:
>> hadoop namenode -format
>> Warning: $HADOOP_HOME is deprecated.
>>
>> 12/04/27 11:31:41 INFO namenode.NameNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting NameNode
>> STARTUP_MSG:   host = ubuntu/127.0.1.1
>> STARTUP_MSG:   args = [-format]
>> STARTUP_MSG:   version = 1.0.1
>> STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0
-r 1243785; compiled by 'hortonfo' on Tue Feb 14 08:13:52 UTC 2012
>> ************************************************************/
>> 12/04/27 11:31:41 INFO util.GSet: VM type       = 32-bit
>> 12/04/27 11:31:41 INFO util.GSet: 2% max memory = 2.475 MB
>> 12/04/27 11:31:41 INFO util.GSet: capacity      = 2^19 = 524288 entries
>> 12/04/27 11:31:41 INFO util.GSet: recommended=524288, actual=524288
>> 12/04/27 11:31:41 ERROR namenode.NameNode: java.lang.IllegalArgumentException: Invalid
attribute value for hadoop.security.authentication of ${SECURITY_TYPE}
>>     at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:202)
>>     at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:187)
>>     at org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(UserGroupInformation.java:239)
>>     at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:438)
>>     at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:424)
>>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setConfigurationParameters(FSNamesystem.java:473)
>>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:462)
>>     at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1162)
>>     at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1271)
>>     at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>>
>> 12/04/27 11:31:41 INFO namenode.NameNode: SHUTDOWN_MSG:
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>> ************************************************************/
>>
>>
>> \_________ Solution
>> I removed the variable and exchanged it with the value "simple".
>> Then the next error occurred:
>>
>>
>> \_________ Error 2
>> hadoop namenode -format
>> Warning: $HADOOP_HOME is deprecated.
>>
>> 12/04/27 11:46:33 INFO namenode.NameNode: STARTUP_MSG:
>> /************************************************************
>> STARTUP_MSG: Starting NameNode
>> STARTUP_MSG:   host = ubuntu/127.0.1.1
>> STARTUP_MSG:   args = [-format]
>> STARTUP_MSG:   version = 1.0.1
>> STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0
-r 1243785; compiled by 'hortonfo' on Tue Feb 14 08:13:52 UTC 2012
>> ************************************************************/
>> 12/04/27 11:46:33 INFO util.GSet: VM type       = 32-bit
>> 12/04/27 11:46:33 INFO util.GSet: 2% max memory = 2.475 MB
>> 12/04/27 11:46:33 INFO util.GSet: capacity      = 2^19 = 524288 entries
>> 12/04/27 11:46:33 INFO util.GSet: recommended=524288, actual=524288
>> 12/04/27 11:46:33 ERROR namenode.NameNode: java.lang.ExceptionInInitializerError
>>     at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:212)
>>     at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:187)
>>     at org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(UserGroupInformation.java:239)
>>     at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:438)
>>     at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:424)
>>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.setConfigurationParameters(FSNamesystem.java:473)
>>     at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.<init>(FSNamesystem.java:462)
>>     at org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:1162)
>>     at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1271)
>>     at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1288)
>> Caused by: java.util.regex.PatternSyntaxException: Illegal repetition near index
8
>> [jt]t@.*${KERBEROS_REALM}
>>         ^
>>     at java.util.regex.Pattern.error(Pattern.java:1730)
>>     at java.util.regex.Pattern.closure(Pattern.java:2792)
>>     at java.util.regex.Pattern.sequence(Pattern.java:1906)
>>     at java.util.regex.Pattern.expr(Pattern.java:1769)
>>     at java.util.regex.Pattern.compile(Pattern.java:1477)
>>     at java.util.regex.Pattern.<init>(Pattern.java:1150)
>>     at java.util.regex.Pattern.compile(Pattern.java:840)
>>     at org.apache.hadoop.security.KerberosName$Rule.<init>(KerberosName.java:188)
>>     at org.apache.hadoop.security.KerberosName.parseRules(KerberosName.java:324)
>>     at org.apache.hadoop.security.KerberosName.setConfiguration(KerberosName.java:343)
>>     at org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:212)
>>     at org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:187)
>>     at org.apache.hadoop.security.UserGroupInformation.isSecurityEnabled(UserGroupInformation.java:239)
>>     at org.apache.hadoop.security.KerberosName.<clinit>(KerberosName.java:83)
>>     ... 10 more
>>
>> 12/04/27 11:46:33 INFO namenode.NameNode: SHUTDOWN_MSG:
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down NameNode at ubuntu/127.0.1.1
>> ************************************************************/
>>
>> \______ And once again...
>> ... a variable seems to be undefined. And i guess, if i found a suitable value for
this property the next one will be undefined. I hope there are any default values present,
because i have no idea, what to fill to the value slots. All books and instructions i read
on hadoop never discussed these issues.
>> BTW: HADOOP_HOME is defined, although the log tells different.
>>
>> I hope you can assist me.
>>
>> Best regards,
>> Björn-Elmar Macek
>>
>


Mime
View raw message