hama-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Edward J. Yoon" <edwardy...@apache.org>
Subject Re: failed to run hama pi example
Date Wed, 12 Sep 2012 13:10:18 GMT
I'm use 0.20.2 and 0.20.2-cdh versions. There's no problem.

Sent from my iPad

On Sep 12, 2012, at 4:59 PM, Thomas Jungblut <thomas.jungblut@gmail.com> wrote:

> Anyone tested the compatibility to Hadoop 20.2 with Hama 0.5?
> [1] says it is compatible.
> 
> [1] http://wiki.apache.org/hama/CompatibilityTable
> 
> 2012/9/12 顾荣 <gurongwalker@gmail.com>
> 
>> Okay, I'll try Hadoop 1.0.3 with Hama 0.50.
>> Thanks Thomas. I can't wait to explore the Hama world now.
>> 
>> walker.
>> 
>> 2012/9/12 Thomas Jungblut <thomas.jungblut@gmail.com>
>> 
>>> oh okay. I'm not sure if 0.5.0 is really compatible to 20.2, personally I
>>> have installed 1.0.3 and it works fine.
>>> Sorry to let you install all the different versions.
>>> 
>>> 2012/9/12 顾荣 <gurongwalker@gmail.com>
>>> 
>>>> Thanks Thomas. The HDFS works well. I even put a file from local to it
>>>> successfully. It absolutely left the safemode. The namenode starting
>> log
>>> is
>>>> as below:
>>>> 
>>>> 2012-09-12 15:10:39,002 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG:
>>>> /************************************************************
>>>> STARTUP_MSG: Starting NameNode
>>>> STARTUP_MSG:   host = slave021/192.168.1.21
>>>> STARTUP_MSG:   args = []
>>>> STARTUP_MSG:   version = 0.20.2
>>>> STARTUP_MSG:   build =
>>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r
>>>> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010
>>>> ************************************************************/
>>>> 2012-09-12 15:10:39,092 INFO org.apache.hadoop.ipc.metrics.RpcMetrics:
>>>> Initializing RPC Metrics with hostName=NameNode, port=54310
>>>> 2012-09-12 15:10:39,098 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at:
>>> slave021/
>>>> 192.168.1.21:54310
>>>> 2012-09-12 15:10:39,100 INFO org.apache.hadoop.metrics.jvm.JvmMetrics:
>>>> Initializing JVM Metrics with processName=NameNode, sessionId=null
>>>> 2012-09-12 15:10:39,101 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics:
>>>> Initializing NameNodeMeterics using context
>>>> object:org.apache.hadoop.metrics.spi.NullContext
>>>> 2012-09-12 15:10:39,143 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> fsOwner=hadoop,hadoop_user,wheel
>>>> 2012-09-12 15:10:39,144 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>> supergroup=supergroup
>>>> 2012-09-12 15:10:39,144 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem:
>>>> isPermissionEnabled=true
>>>> 2012-09-12 15:10:39,150 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics:
>>>> Initializing FSNamesystemMetrics using context
>>>> object:org.apache.hadoop.metrics.spi.NullContext
>>>> 2012-09-12 15:10:39,151 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered
>>>> FSNamesystemStatusMBean
>>>> 2012-09-12 15:10:39,177 INFO
>>> org.apache.hadoop.hdfs.server.common.Storage:
>>>> Number of files = 1
>>>> 2012-09-12 15:10:39,181 INFO
>>> org.apache.hadoop.hdfs.server.common.Storage:
>>>> Number of files under construction = 0
>>>> 2012-09-12 15:10:39,181 INFO
>>> org.apache.hadoop.hdfs.server.common.Storage:
>>>> Image file of size 96 loaded in 0 seconds.
>>>> 2012-09-12 15:10:39,181 INFO
>>> org.apache.hadoop.hdfs.server.common.Storage:
>>>> Edits file
>>>> /home/hadoop/gurong/hadoop-0.20.2/hadoop_dir/dfs/name/current/edits of
>>> size
>>>> 4 edits # 0 loaded in 0 seconds.
>>>> 2012-09-12 15:10:39,236 INFO
>>> org.apache.hadoop.hdfs.server.common.Storage:
>>>> Image file of size 96 saved in 0 seconds.
>>>> 2012-09-12 15:10:39,439 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading
>>>> FSImage in 312 msecs
>>>> 2012-09-12 15:10:39,441 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of
>>> blocks
>>>> = 0
>>>> 2012-09-12 15:10:39,441 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid
>>>> blocks = 0
>>>> 2012-09-12 15:10:39,441 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>>>> under-replicated blocks = 0
>>>> 2012-09-12 15:10:39,441 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>>>> over-replicated blocks = 0
>>>> 2012-09-12 15:10:39,441 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>>> Leaving safe mode after 0 secs.
>>>> 2012-09-12 15:10:39,441 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>>> Network topology has 0 racks and 0 datanodes
>>>> 2012-09-12 15:10:39,441 INFO org.apache.hadoop.hdfs.StateChange: STATE*
>>>> UnderReplicatedBlocks has 0 blocks
>>>> 2012-09-12 15:10:39,554 INFO org.mortbay.log: Logging to
>>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>>>> org.mortbay.log.Slf4jLog
>>>> 2012-09-12 15:10:39,603 INFO org.apache.hadoop.http.HttpServer: Port
>>>> returned by webServer.getConnectors()[0].getLocalPort() before open()
>> is
>>>> -1. Opening the listener on 50070
>>>> 2012-09-12 15:10:39,604 INFO org.apache.hadoop.http.HttpServer:
>>>> listener.getLocalPort() returned 50070
>>>> webServer.getConnectors()[0].getLocalPort() returned 50070
>>>> 2012-09-12 15:10:39,604 INFO org.apache.hadoop.http.HttpServer: Jetty
>>> bound
>>>> to port 50070
>>>> 2012-09-12 15:10:39,604 INFO org.mortbay.log: jetty-6.1.14
>>>> 2012-09-12 15:10:48,662 INFO org.mortbay.log: Started
>>>> SelectChannelConnector@0.0.0.0:50070
>>>> 2012-09-12 15:10:48,663 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at:
>>>> 0.0.0.0:50070
>>>> 2012-09-12 15:10:48,666 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> listener on 54310: starting
>>>> 2012-09-12 15:10:48,667 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> Responder: starting
>>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 0 on 54310: starting
>>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 1 on 54310: starting
>>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 2 on 54310: starting
>>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 3 on 54310: starting
>>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 4 on 54310: starting
>>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 5 on 54310: starting
>>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 6 on 54310: starting
>>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 7 on 54310: starting
>>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 8 on 54310: starting
>>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server
>>>> handler 9 on 54310: starting
>>>> 2012-09-12 15:10:48,700 INFO org.apache.hadoop.ipc.Server: Error
>> register
>>>> getProtocolVersion
>>>> java.lang.IllegalArgumentException: Duplicate
>>>> metricsName:getProtocolVersion
>>>>    at
>>>> 
>>> 
>> org.apache.hadoop.metrics.util.MetricsRegistry.add(MetricsRegistry.java:53)
>>>>    at
>>>> 
>>>> 
>>> 
>> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:89)
>>>>    at
>>>> 
>>>> 
>>> 
>> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:99)
>>>>    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:523)
>>>>    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)
>>>>    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)
>>>>    at java.security.AccessController.doPrivileged(Native Method)
>>>>    at javax.security.auth.Subject.doAs(Subject.java:416)
>>>>    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)
>>>> 2012-09-12 15:11:05,298 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
>>>> NameSystem.registerDatanode: node registration from
>>>> 192.168.1.21:50010storage
>>>> DS-1416037815-192.168.1.21-50010-1347433865293
>>>> 2012-09-12 15:11:05,300 INFO org.apache.hadoop.net.NetworkTopology:
>>> Adding
>>>> a new node: /default-rack/192.168.1.21:50010
>>>> 2012-09-12 15:11:15,069 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>>> ugi=webuser,webgroup    ip=/192.168.1.21    cmd=listStatus    src=/
>>>> dst=null    perm=null
>>>> 2012-09-12 15:12:05,034 WARN org.apache.hadoop.ipc.Server: Incorrect
>>> header
>>>> or version mismatch from 192.168.1.21:56281 got version 4 expected
>>>> version 3
>>>> 2012-09-12 15:14:51,535 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit:
>>>> ugi=hadoop,hadoop_user,wheel    ip=/192.168.1.21    cmd=listStatus
>>>> src=/    dst=null    perm=null
>>>> 2012-09-12 15:15:10,158 INFO
>>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
>>>> transactions: 2 Total time for transactions(ms): 0Number of
>> transactions
>>>> batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0
>>>> 
>>>> I used Hama 0.5 and hadoop 0.20.2. Has somebody test this match can
>> work
>>>> well?
>>>> 
>>>> thanks very much.
>>>> 
>>>> walker
>>>> 
>>>> 2012/9/12 Thomas Jungblut <thomas.jungblut@gmail.com>
>>>> 
>>>>> Still it says:
>>>>>> 
>>>>>> " 2012-09-12 14:41:16,218 ERROR org.apache.hama.bsp.BSPMaster:
>> Can't
>>>> get
>>>>>> connection to Hadoop Namenode! "
>>>>> 
>>>>> 
>>>>> Can you verify that the namenode is not in safemode and has correctly
>>>>> started up?
>>>>> Have a look into the namenode logs please!
>>>>> 
>>>>> 2012/9/12 顾荣 <gurongwalker@gmail.com>
>>>>> 
>>>>>> By the way, the fs.default.name is 192.168.1.21:54310. I checked
>> the
>>>>> HDFS,
>>>>>> it works well. I installed and ran both HDFS and Hama using the
>> same
>>>>> linux
>>>>>> account.
>>>>>> 
>>>>>> 2012/9/12 顾荣 <gurongwalker@gmail.com>
>>>>>> 
>>>>>>> Thanks so much, Edward.
>>>>>>> 
>>>>>>> I fellowed your suggestion and instanlled a hadoop 0.20.2 instead
>>> for
>>>>>>> Hama. However, this time when I start Hama, a fatal happened
and
>>> the
>>>>>>> bspmaster daemon can not start up. The corresponding error
>> message
>>> in
>>>>> the
>>>>>>> baspmaster log file shows as below.
>>>>>>> 
>>>>>>> ************************************************************/
>>>>>>> 2012-09-12 14:40:38,238 INFO org.apache.hama.BSPMasterRunner:
>>>>>> STARTUP_MSG:
>>>>>>> /************************************************************
>>>>>>> STARTUP_MSG: Starting BSPMaster
>>>>>>> STARTUP_MSG:   host = slave021/192.168.1.21
>>>>>>> STARTUP_MSG:   args = []
>>>>>>> STARTUP_MSG:   version = 1.0.0
>>>>>>> STARTUP_MSG:   build =
>>>>>>> 
>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0-r
>>>>>>> 1214675; compiled by 'hortonfo' on Fri Dec 16 20:01:27 UTC 2011
>>>>>>> ************************************************************/
>>>>>>> 2012-09-12 14:40:38,414 INFO org.apache.hama.bsp.BSPMaster: RPC
>>>>>> BSPMaster:
>>>>>>> host slave021 port 40000
>>>>>>> 2012-09-12 14:40:38,502 INFO org.apache.hadoop.ipc.Server:
>> Starting
>>>>>>> SocketReader
>>>>>>> 2012-09-12 14:40:38,542 INFO org.mortbay.log: Logging to
>>>>>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via
>>>>>>> org.mortbay.log.Slf4jLog
>>>>>>> 2012-09-12 14:40:38,583 INFO org.apache.hama.http.HttpServer:
>> Port
>>>>>>> returned by webServer.getConnectors()[0].getLocalPort() before
>>> open()
>>>>> is
>>>>>>> -1. Opening the listener on 40013
>>>>>>> 2012-09-12 14:40:38,584 INFO org.apache.hama.http.HttpServer:
>>>>>>> listener.getLocalPort() returned 40013
>>>>>>> webServer.getConnectors()[0].getLocalPort() returned 40013
>>>>>>> 2012-09-12 14:40:38,584 INFO org.apache.hama.http.HttpServer:
>> Jetty
>>>>> bound
>>>>>>> to port 40013
>>>>>>> 2012-09-12 14:40:38,584 INFO org.mortbay.log: jetty-6.1.14
>>>>>>> 2012-09-12 14:40:38,610 INFO org.mortbay.log: Extract
>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> jar:file:/home/hadoop/hama_installs/hama-0.5.0/hama-core-0.5.0.jar!/webapp/bspmaster/
>>>>>>> to /tmp/Jetty_slave021_40013_bspmaster____.1tzgsz/webapp
>>>>>>> 2012-09-12 14:41:16,073 INFO org.mortbay.log: Started
>>>>>>> SelectChannelConnector@slave021:40013
>>>>>>> 2012-09-12 14:41:16,218 ERROR org.apache.hama.bsp.BSPMaster:
>> Can't
>>>> get
>>>>>>> connection to Hadoop Namenode!
>>>>>>> java.io.IOException: Call to /192.168.1.21:54310 failed on local
>>>>>>> exception: java.io.EOFException
>>>>>>>    at
>> org.apache.hadoop.ipc.Client.wrapException(Client.java:1103)
>>>>>>>    at org.apache.hadoop.ipc.Client.call(Client.java:1071)
>>>>>>>    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>>>>>>    at $Proxy5.getProtocolVersion(Unknown Source)
>>>>>>>    at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396)
>>>>>>>    at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379)
>>>>>>>    at
>>>>>>> 
>>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119)
>>>>>>>    at
>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238)
>>>>>>>    at
>> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203)
>>>>>>>    at
>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89)
>>>>>>>    at
>>>>>>> 
>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386)
>>>>>>>    at
>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
>>>>>>>    at
>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404)
>>>>>>>    at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
>>>>>>>    at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:123)
>>>>>>>    at org.apache.hama.bsp.BSPMaster.<init>(BSPMaster.java:299)
>>>>>>>    at
>>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:454)
>>>>>>>    at
>>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:449)
>>>>>>>    at
>> org.apache.hama.BSPMasterRunner.run(BSPMasterRunner.java:46)
>>>>>>>    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>>>>>>    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
>>>>>>>    at
>>> org.apache.hama.BSPMasterRunner.main(BSPMasterRunner.java:56)
>>>>>>> Caused by: java.io.EOFException
>>>>>>>    at java.io.DataInputStream.readInt(DataInputStream.java:392)
>>>>>>>    at
>>>>>>> 
>>>>> 
>>> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:800)
>>>>>>>    at
>> org.apache.hadoop.ipc.Client$Connection.run(Client.java:745)
>>>>>>> 2012-09-12 14:41:16,222 FATAL org.apache.hama.BSPMasterRunner:
>>>>>>> java.lang.NullPointerException
>>>>>>>    at
>>> org.apache.hama.bsp.BSPMaster.getSystemDir(BSPMaster.java:862)
>>>>>>>    at org.apache.hama.bsp.BSPMaster.<init>(BSPMaster.java:308)
>>>>>>>    at
>>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:454)
>>>>>>>    at
>>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:449)
>>>>>>>    at
>> org.apache.hama.BSPMasterRunner.run(BSPMasterRunner.java:46)
>>>>>>>    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>>>>>>    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
>>>>>>>    at
>>> org.apache.hama.BSPMasterRunner.main(BSPMasterRunner.java:56)
>>>>>>> 
>>>>>>> 2012-09-12 14:41:16,223 INFO org.apache.hama.BSPMasterRunner:
>>>>>>> SHUTDOWN_MSG:
>>>>>>> /************************************************************
>>>>>>> SHUTDOWN_MSG: Shutting down BSPMaster at slave021/192.168.1.21
>>>>>>> ************************************************************/
>>>>>>> 
>>>>>>> Would please give me some tips again?
>>>>>>> 
>>>>>>> Thanks,again.
>>>>>>> 
>>>>>>> walker
>>>>>>> 
>>>>>>> 2012/9/12 Edward J. Yoon <edwardyoon@apache.org>
>>>>>>> 
>>>>>>> Unfortunately we don't support Hadoop secure version yet.
>>>>>>>> 
>>>>>>>> Instead of 0.20.205, Please use hadoop non-secure 0.20.2
or
>> 1.0.3
>>>>>>>> versions.
>>>>>>>> 
>>>>>>>> Thanks.
>>>>>>>> 
>>>>>>>> On Wed, Sep 12, 2012 at 11:25 AM, 顾荣 <gurongwalker@gmail.com>
>>>> wrote:
>>>>>>>>> Hi,all.
>>>>>>>>> 
>>>>>>>>> I set up a hama cluster of 3 nodes and start hama
>> successfully.
>>>>>> However,
>>>>>>>>> when I run the pi example, the job failed with a very
strange
>>>>> message
>>>>>> as
>>>>>>>>> below.
>>>>>>>>> 
>>>>>>>>> hama jar
>>>>>> /home/hadoop/hama_installs/hama-0.5.0/hama-examples-0.5.0.jar
>>>>>>>> pi
>>>>>>>>> org.apache.hadoop.ipc.RemoteException: java.io.IOException:
>>>>>>>>> java.lang.NoSuchMethodException:
>>>>>>>>> 
>>>>>> 
>>> org.apache.hadoop.hdfs.protocol.ClientProtocol.create(java.lang.String,
>>>>>>>>> org.apache.hadoop.fs.permission.FsPermission,
>> java.lang.String,
>>>>>> boolean,
>>>>>>>>> boolean, short, long)
>>>>>>>>>    at java.lang.Class.getMethod(Class.java:1605)
>>>>>>>>>    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:557)
>>>>>>>>>    at
>>>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388)
>>>>>>>>>    at
>>>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384)
>>>>>>>>>    at java.security.AccessController.doPrivileged(Native
>>> Method)
>>>>>>>>>    at javax.security.auth.Subject.doAs(Subject.java:396)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059)
>>>>>>>>>    at
>>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382)
>>>>>>>>> 
>>>>>>>>>    at org.apache.hadoop.ipc.Client.call(Client.java:1066)
>>>>>>>>>    at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225)
>>>>>>>>>    at $Proxy2.create(Unknown Source)
>>>>>>>>>    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>> Method)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>>>>    at java.lang.reflect.Method.invoke(Method.java:616)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
>>>>>>>>>    at $Proxy2.create(Unknown Source)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.<init>(DFSClient.java:3245)
>>>>>>>>>    at
>>> org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:713)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:182)
>>>>>>>>>    at
>>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:555)
>>>>>>>>>    at
>>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:536)
>>>>>>>>>    at
>>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:443)
>>>>>>>>>    at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:229)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>> 
>>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1195)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>> 
>>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1171)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>> 
>>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1143)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>> org.apache.hama.bsp.BSPJobClient.submitJobInternal(BSPJobClient.java:349)
>>>>>>>>>    at
>>>>>> org.apache.hama.bsp.BSPJobClient.submitJob(BSPJobClient.java:294)
>>>>>>>>>    at org.apache.hama.bsp.BSPJob.submit(BSPJob.java:218)
>>>>>>>>>    at
>>>> org.apache.hama.bsp.BSPJob.waitForCompletion(BSPJob.java:225)
>>>>>>>>>    at
>>>>> org.apache.hama.examples.PiEstimator.main(PiEstimator.java:139)
>>>>>>>>>    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>> Method)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>>>>    at java.lang.reflect.Method.invoke(Method.java:616)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68)
>>>>>>>>>    at
>>>>>>>> 
>>> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139)
>>>>>>>>>    at
>>>>>>>> 
>> org.apache.hama.examples.ExampleDriver.main(ExampleDriver.java:39)
>>>>>>>>>    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>> Method)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>>>>>>>    at
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>>>>    at java.lang.reflect.Method.invoke(Method.java:616)
>>>>>>>>>    at org.apache.hama.util.RunJar.main(RunJar.java:147)
>>>>>>>>> 
>>>>>>>>> My hama verison is 0.5 and hadoop version is 0.20.205.
This
>>> error
>>>>>> seems
>>>>>>>> to
>>>>>>>>> comes from the
>>>>> "org.apache.hadoop.hdfs.protocol.ClientProtocol.create"
>>>>>>>>> method, this is a normal method. I am kind of confused...
>>>>>>>>> 
>>>>>>>>> Thanks in advance.
>>>>>>>>> 
>>>>>>>>> walker
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Best Regards, Edward J. Yoon
>>>>>>>> @eddieyoon
>>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>> 
>>>>> 
>>>> 
>>> 
>> 

Mime
View raw message