hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Sitaraman Vilayannur <vrsitaramanietfli...@gmail.com>
Subject Re: Getting error unrecognized option -jvm on starting nodemanager
Date Tue, 24 Dec 2013 15:18:17 GMT
I did press Y tried it several times and once more now.
Sitaraman


On Tue, Dec 24, 2013 at 8:38 PM, Nitin Pawar <nitinpawar432@gmail.com>wrote:

> see the error .. it says not formatted
> did you press Y or y ?
> try again :)
>
>
> On Tue, Dec 24, 2013 at 8:35 PM, Sitaraman Vilayannur <
> vrsitaramanietflists@gmail.com> wrote:
>
>> Hi Nitin,
>>  Even after formatting using hdfs namenode -format, i keep seeing
>> namenode not formatted in the logs when i try to start namenode........
>> 12/24 20:33:26 INFO namenode.FSNamesystem: supergroup=supergroup
>> 13/12/24 20:33:26 INFO namenode.FSNamesystem: isPermissionEnabled=true
>> 13/12/24 20:33:26 INFO namenode.NameNode: Caching file names occuring
>> more than 10 times
>> 13/12/24 20:33:26 INFO namenode.NNStorage: Storage directory
>> /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode has been successfully
>> formatted.
>> 13/12/24 20:33:26 INFO namenode.FSImage: Saving image file
>> /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/current/fsimage.ckpt_0000000000000000000
>> using no compression
>> 13/12/24 20:33:26 INFO namenode.FSImage: Image file of size 124 saved in
>> 0 seconds.
>> 13/12/24 20:33:26 INFO namenode.NNStorageRetentionManager: Going to
>> retain 1 images with txid >= 0
>> 13/12/24 20:33:26 INFO util.ExitUtil: Exiting with status 0
>> 13/12/24 20:33:26 INFO namenode.NameNode: SHUTDOWN_MSG:
>>
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
>> ************************************************************/
>>
>>
>> 2013-12-24 20:33:46,337 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Lock on
>> /usr/local/Software/hadoop-2.2.0/data/hdfs/namenode/in_use.lock acquired by
>> nodename 7518@localhost.localdomain
>> 2013-12-24 20:33:46,339 INFO org.mortbay.log: Stopped
>> SelectChannelConnector@0.0.0.0:50070
>> 2013-12-24 20:33:46,340 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
>> metrics system...
>> 2013-12-24 20:33:46,340 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>> stopped.
>> 2013-12-24 20:33:46,340 INFO
>> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system
>> shutdown complete.
>> 2013-12-24 20:33:46,340 FATAL
>> org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode join
>> java.io.IOException: NameNode is not formatted.
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:210)
>>
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
>>         at
>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
>> 2013-12-24 20:33:46,342 INFO org.apache.hadoop.util.ExitUtil: Exiting
>> with status 1
>> 2013-12-24 20:33:46,343 INFO
>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>
>> /************************************************************
>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
>> ************************************************************/
>>
>>
>>
>>
>> On Tue, Dec 24, 2013 at 3:13 PM, Nitin Pawar <nitinpawar432@gmail.com>wrote:
>>
>>> the issue here is you tried one version of hadoop and then changed to a
>>> different version.
>>>
>>> You can not do that directly with hadoop. You need to follow a process
>>> while upgrading hadoop versions.
>>>
>>> For now as you are just starting with hadoop, I would recommend just run
>>> a dfs format and start the hdfs again
>>>
>>>
>>> On Tue, Dec 24, 2013 at 2:57 PM, Sitaraman Vilayannur <
>>> vrsitaramanietflists@gmail.com> wrote:
>>>
>>>> When i run namenode with upgrade option i get the following error and
>>>> and namenode dosent start...
>>>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
>>>> STATE* Network topology has 0 racks and 0 datanodes
>>>> 2013-12-24 14:48:38,595 INFO org.apache.hadoop.hdfs.StateChange:
>>>> STATE* UnderReplicatedBlocks has 0 blocks
>>>> 2013-12-24 14:48:38,631 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> Responder: starting
>>>> 2013-12-24 14:48:38,632 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> listener on 9000: starting
>>>> 2013-12-24 14:48:38,633 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: NameNode RPC up at:
>>>> 192.168.1.2/192.168.1.2:9000
>>>> 2013-12-24 14:48:38,633 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Starting services
>>>> required for active state
>>>> 2013-12-24 14:50:50,060 ERROR
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: RECEIVED SIGNAL 15:
>>>> SIGTERM
>>>> 2013-12-24 14:50:50,062 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>>> /************************************************************
>>>> SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/127.0.0.1
>>>> ************************************************************/
>>>>
>>>>
>>>> On 12/24/13, Sitaraman Vilayannur <vrsitaramanietflists@gmail.com>
>>>> wrote:
>>>> > Found it,
>>>> >  I get the following error on starting namenode in 2.2
>>>> >
>>>> 10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar:/usr/local/Software/hadoop-0.23.10/contrib/capacity-scheduler/*.jar
>>>> > STARTUP_MSG:   build = https://svn.apache.org/repos/asf/hadoop/common
>>>> > -r 1529768; compiled by 'hortonmu' on 2013-10-07T06:28Z
>>>> > STARTUP_MSG:   java = 1.7.0_45
>>>> > ************************************************************/
>>>> > 2013-12-24 13:25:48,876 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX
>>>> > signal handlers for [TERM, HUP, INT]
>>>> > 2013-12-24 13:25:49,042 INFO
>>>> > org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from
>>>> > hadoop-metrics2.properties
>>>> > 2013-12-24 13:25:49,102 INFO
>>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot
>>>> > period at 10 second(s).
>>>> > 2013-12-24 13:25:49,102 INFO
>>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>>>> > system started
>>>> > 2013-12-24 13:25:49,232 WARN org.apache.hadoop.util.NativeCodeLoader:
>>>> > Unable to load native-hadoop library for your platform... using
>>>> > builtin-java classes where applicable
>>>> > 2013-12-24 13:25:49,375 INFO org.mortbay.log: Logging to
>>>> > org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>>>> > org.mortbay.log.Slf4jLog
>>>> > 2013-12-24 13:25:49,410 INFO org.apache.hadoop.http.HttpServer: Added
>>>> > global filter 'safety'
>>>> > (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter)
>>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>>>> > filter static_user_filter
>>>> >
>>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>>>> > to context hdfs
>>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>>>> > filter static_user_filter
>>>> >
>>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>>>> > to context static
>>>> > 2013-12-24 13:25:49,412 INFO org.apache.hadoop.http.HttpServer: Added
>>>> > filter static_user_filter
>>>> >
>>>> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter)
>>>> > to context logs
>>>> > 2013-12-24 13:25:49,422 INFO org.apache.hadoop.http.HttpServer:
>>>> > dfs.webhdfs.enabled = false
>>>> > 2013-12-24 13:25:49,432 INFO org.apache.hadoop.http.HttpServer: Jetty
>>>> > bound to port 50070
>>>> > 2013-12-24 13:25:49,432 INFO org.mortbay.log: jetty-6.1.26
>>>> > 2013-12-24 13:25:49,459 WARN org.mortbay.log: Can't reuse
>>>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08, using
>>>> > /tmp/Jetty_0_0_0_0_50070_hdfs____w2cu08_2787234685293301311
>>>> > 2013-12-24 13:25:49,610 INFO org.mortbay.log: Started
>>>> > SelectChannelConnector@0.0.0.0:50070
>>>> > 2013-12-24 13:25:49,611 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
>>>> > 0.0.0.0:50070
>>>> > 2013-12-24 13:25:49,628 WARN
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one image
>>>> > storage directory (dfs.namenode.name.dir) configured. Beware of
>>>> > dataloss due to lack of redundant storage directories!
>>>> > 2013-12-24 13:25:49,628 WARN
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Only one
>>>> > namespace edits storage directory (dfs.namenode.edits.dir) configured.
>>>> > Beware of dataloss due to lack of redundant storage directories!
>>>> > 2013-12-24 13:25:49,668 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read includes:
>>>> > HostSet(
>>>> > )
>>>> > 2013-12-24 13:25:49,669 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.HostFileManager: read excludes:
>>>> > HostSet(
>>>> > )
>>>> > 2013-12-24 13:25:49,670 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager:
>>>> > dfs.block.invalidate.limit=1000
>>>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: Computing
>>>> > capacity for map BlocksMap
>>>> > 2013-12-24 13:25:49,672 INFO org.apache.hadoop.util.GSet: VM type
>>>>   =
>>>> > 64-bit
>>>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: 2.0% max
>>>> > memory = 889 MB
>>>> > 2013-12-24 13:25:49,673 INFO org.apache.hadoop.util.GSet: capacity
>>>> >  = 2^21 = 2097152 entries
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > dfs.block.access.token.enable=false
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > defaultReplication         = 1
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > maxReplication             = 512
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > minReplication             = 1
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > maxReplicationStreams      = 2
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > shouldCheckForEnoughRacks  = false
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > replicationRecheckInterval = 3000
>>>> > 2013-12-24 13:25:49,677 INFO
>>>> > org.apache.hadoop.hdfs.server.blockmanagement.BlockManager:
>>>> > encryptDataTransfer        = false
>>>> > 2013-12-24 13:25:49,681 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: fsOwner
>>>> >   = sitaraman (auth:SIMPLE)
>>>> > 2013-12-24 13:25:49,681 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: supergroup
>>>> >   = supergroup
>>>> > 2013-12-24 13:25:49,681 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> > isPermissionEnabled = true
>>>> > 2013-12-24 13:25:49,681 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: HA Enabled: false
>>>> > 2013-12-24 13:25:49,682 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Append Enabled:
>>>> > true
>>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: Computing
>>>> > capacity for map INodeMap
>>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: VM type
>>>>   =
>>>> > 64-bit
>>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: 1.0% max
>>>> > memory = 889 MB
>>>> > 2013-12-24 13:25:49,801 INFO org.apache.hadoop.util.GSet: capacity
>>>> >  = 2^20 = 1048576 entries
>>>> > 2013-12-24 13:25:49,802 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Caching file names
>>>> > occuring more than 10 times
>>>> > 2013-12-24 13:25:49,804 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> > dfs.namenode.safemode.threshold-pct = 0.9990000128746033
>>>> > 2013-12-24 13:25:49,804 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> > dfs.namenode.safemode.min.datanodes = 0
>>>> > 2013-12-24 13:25:49,804 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> > dfs.namenode.safemode.extension     = 30000
>>>> > 2013-12-24 13:25:49,805 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache on
>>>> > namenode is enabled
>>>> > 2013-12-24 13:25:49,805 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Retry cache will
>>>> > use 0.03 of total heap and retry cache entry expiry time is 600000
>>>> > millis
>>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: Computing
>>>> > capacity for map Namenode Retry Cache
>>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: VM type
>>>>   =
>>>> > 64-bit
>>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet:
>>>> > 0.029999999329447746% max memory = 889 MB
>>>> > 2013-12-24 13:25:49,807 INFO org.apache.hadoop.util.GSet: capacity
>>>> >  = 2^15 = 32768 entries
>>>> > 2013-12-24 13:25:49,816 INFO
>>>> > org.apache.hadoop.hdfs.server.common.Storage: Lock on
>>>> > /usr/local/Software/hadoop-0.23.10/data/hdfs/namenode/in_use.lock
>>>> > acquired by nodename 19170@localhost.localdomain
>>>> > 2013-12-24 13:25:49,861 INFO org.mortbay.log: Stopped
>>>> > SelectChannelConnector@0.0.0.0:50070
>>>> > 2013-12-24 13:25:49,964 INFO
>>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode
>>>> > metrics system...
>>>> > 2013-12-24 13:25:49,965 INFO
>>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>>>> > system stopped.
>>>> > 2013-12-24 13:25:49,965 INFO
>>>> > org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics
>>>> > system shutdown complete.
>>>> > 2013-12-24 13:25:49,965 FATAL
>>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: Exception in namenode
>>>> > join
>>>> > java.io.IOException:
>>>> > File system image contains an old layout version -39.
>>>> > An upgrade to version -47 is required.
>>>> > Please restart NameNode with -upgrade option.
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:221)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:787)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:568)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:443)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:491)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:684)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:669)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1254)
>>>> >       at
>>>> >
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1320)
>>>> > 2013-12-24 13:25:49,967 INFO org.apache.hadoop.util.ExitUtil: Exiting
>>>> > with status 1
>>>> > 2013-12-24 13:25:49,968 INFO
>>>> > org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG:
>>>> > /************************************************************
>>>> > SHUTDOWN_MSG: Shutting down NameNode at localhost.localdomain/
>>>> 127.0.0.1
>>>> > ************************************************************/
>>>> >
>>>> > On 12/24/13, Sitaraman Vilayannur <vrsitaramanietflists@gmail.com>
>>>> wrote:
>>>> >> The line beginning with ulimit that i have appended below, i thought
>>>> >> was the log file?
>>>> >>
>>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>>>> >> Sitaraman
>>>> >> On 12/24/13, Nitin Pawar <nitinpawar432@gmail.com> wrote:
>>>> >>> Without log, very hard to guess what's happening.
>>>> >>>
>>>> >>> Can you clean up the log directory and then start over and check
>>>> for the
>>>> >>> logs again.
>>>> >>>
>>>> >>>
>>>> >>> On Tue, Dec 24, 2013 at 11:44 AM, Sitaraman Vilayannur <
>>>> >>> vrsitaramanietflists@gmail.com> wrote:
>>>> >>>
>>>> >>>> Hi Nitin,
>>>> >>>>  I moved to the release 2.2.0 on starting node manager it
remains
>>>> >>>> silent without errors but nodemanager dosent start....while
it
>>>> does in
>>>> >>>> the earlier 0.23 version
>>>> >>>>
>>>> >>>>
>>>> >>>> ./hadoop-daemon.sh start namenode
>>>> >>>> starting namenode, logging to
>>>> >>>>
>>>> >>>>
>>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>>>> >>>> Java HotSpot(TM) 64-Bit Server VM warning: You have loaded
library
>>>> >>>> /usr/local/Software/hadoop-2.2.0/lib/native/libhadoop.so.1.0.0
>>>> which
>>>> >>>> might have disabled stack guard. The VM will try to fix
the stack
>>>> >>>> guard now.
>>>> >>>> It's highly recommended that you fix the library with 'execstack
-c
>>>> >>>> <libfile>', or link it with '-z noexecstack'.
>>>> >>>> [sitaraman@localhost sbin]$ jps
>>>> >>>> 13444 Jps
>>>> >>>> [sitaraman@localhost sbin]$ vi
>>>> >>>>
>>>> >>>>
>>>> /usr/local/Software/hadoop-2.2.0/logs/hadoop-sitaraman-namenode-localhost.localdomain.out
>>>> >>>>
>>>> >>>>
>>>> >>>> ulimit -a for user sitaraman
>>>> >>>> core file size          (blocks, -c) 0
>>>> >>>> data seg size           (kbytes, -d) unlimited
>>>> >>>> scheduling priority             (-e) 0
>>>> >>>> file size               (blocks, -f) unlimited
>>>> >>>> pending signals                 (-i) 135104
>>>> >>>> max locked memory       (kbytes, -l) 32
>>>> >>>> max memory size         (kbytes, -m) unlimited
>>>> >>>> open files                      (-n) 1024
>>>> >>>> pipe size            (512 bytes, -p) 8
>>>> >>>> POSIX message queues     (bytes, -q) 819200
>>>> >>>> real-time priority              (-r) 0
>>>> >>>> stack size              (kbytes, -s) 10240
>>>> >>>> cpu time               (seconds, -t) unlimited
>>>> >>>> max user processes              (-u) 135104
>>>> >>>> virtual memory          (kbytes, -v) unlimited
>>>> >>>> file locks                      (-x) unlimited
>>>> >>>>
>>>> >>>>
>>>> >>>> On 12/24/13, Nitin Pawar <nitinpawar432@gmail.com>
wrote:
>>>> >>>> > For now you can ignore this warning,
>>>> >>>> > it was your first program so you can try building other
things
>>>> and
>>>> >>>> > slowly
>>>> >>>> > run the commands mentioned the log message to fix these
small
>>>> >>>> > warnings.
>>>> >>>> >
>>>> >>>> >
>>>> >>>> > On Tue, Dec 24, 2013 at 10:07 AM, Sitaraman Vilayannur
<
>>>> >>>> > vrsitaramanietflists@gmail.com> wrote:
>>>> >>>> >
>>>> >>>> >> Thanks Nitin, That worked,
>>>> >>>> >> When i run the Pi example, i get the following
warning at the
>>>> end,
>>>> >>>> >> what must i do about this warning....thanks much
for your help.
>>>> >>>> >> Sitaraman
>>>> >>>> >> inished in 20.82 seconds
>>>> >>>> >> Java HotSpot(TM) 64-Bit Server VM warning: You
have loaded
>>>> library
>>>> >>>> >> /usr/local/Software/hadoop-0.23.10/lib/native/libhadoop.so.1.0.0
>>>> >>>> >> which
>>>> >>>> >> might have disabled stack guard. The VM will try
to fix the
>>>> stack
>>>> >>>> >> guard now.
>>>> >>>> >> It's highly recommended that you fix the library
with
>>>> 'execstack -c
>>>> >>>> >> <libfile>', or link it with '-z noexecstack'.
>>>> >>>> >> 13/12/24 10:05:19 WARN util.NativeCodeLoader: Unable
to load
>>>> >>>> >> native-hadoop library for your platform... using
builtin-java
>>>> >>>> >> classes
>>>> >>>> >> where applicable
>>>> >>>> >> Estimated value of Pi is 3.14127500000000000000
>>>> >>>> >> [sitaraman@localhost mapreduce]$
>>>> >>>> >>
>>>> >>>> >> On 12/23/13, Nitin Pawar <nitinpawar432@gmail.com>
wrote:
>>>> >>>> >> > Can you try starting the process as non root
user.
>>>> >>>> >> > Give proper permissions to the user and start
it as a
>>>> different
>>>> >>>> >> > user.
>>>> >>>> >> >
>>>> >>>> >> > Thanks,
>>>> >>>> >> > Nitin
>>>> >>>> >> >
>>>> >>>> >> >
>>>> >>>> >> > On Mon, Dec 23, 2013 at 2:15 PM, Sitaraman
Vilayannur <
>>>> >>>> >> > vrsitaramanietflists@gmail.com> wrote:
>>>> >>>> >> >
>>>> >>>> >> >> Hi,
>>>> >>>> >> >>  When i attempt to start nodemanager i
get the following
>>>> error.
>>>> >>>> >> >> Any
>>>> >>>> >> >> help
>>>> >>>> >> >> appreciated.   I was able to start resource
manager datanode,
>>>> >>>> namenode
>>>> >>>> >> >> and
>>>> >>>> >> >> secondarynamenode,
>>>> >>>> >> >>
>>>> >>>> >> >>
>>>> >>>> >> >>    ./yarn-daemon.sh start nodemanager
>>>> >>>> >> >> starting nodemanager, logging to
>>>> >>>> >> >>
>>>> >>>> >>
>>>> >>>>
>>>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
>>>> >>>> >> >> Unrecognized option: -jvm
>>>> >>>> >> >> Error: Could not create the Java Virtual
Machine.
>>>> >>>> >> >> Error: A fatal exception has occurred.
Program will exit.
>>>> >>>> >> >> [root@localhost sbin]# emacs
>>>> >>>> >> >>
>>>> >>>> >>
>>>> >>>>
>>>> /usr/local/Software/hadoop-0.23.10/logs/yarn-root-nodemanager-localhost.localdomain.out
>>>> >>>> >> >> &
>>>> >>>> >> >> [4] 29004
>>>> >>>> >> >> [root@localhost sbin]# jps
>>>> >>>> >> >> 28402 SecondaryNameNode
>>>> >>>> >> >> 30280 Jps
>>>> >>>> >> >> 28299 DataNode
>>>> >>>> >> >> 6729 Main
>>>> >>>> >> >> 26044 ResourceManager
>>>> >>>> >> >> 28197 NameNode
>>>> >>>> >> >>
>>>> >>>> >> >
>>>> >>>> >> >
>>>> >>>> >> >
>>>> >>>> >> > --
>>>> >>>> >> > Nitin Pawar
>>>> >>>> >> >
>>>> >>>> >>
>>>> >>>> >
>>>> >>>> >
>>>> >>>> >
>>>> >>>> > --
>>>> >>>> > Nitin Pawar
>>>> >>>> >
>>>> >>>>
>>>> >>>
>>>> >>>
>>>> >>>
>>>> >>> --
>>>> >>> Nitin Pawar
>>>> >>>
>>>> >>
>>>> >
>>>>
>>>
>>>
>>>
>>> --
>>> Nitin Pawar
>>>
>>
>>
>
>
> --
> Nitin Pawar
>

Mime
View raw message