hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From yeyu1899 <yeyu1...@163.com>
Subject Error for Pseudo-distributed Mode
Date Tue, 12 Feb 2013 12:57:58 GMT
Hi all,
I installed a redhat_enterprise-linux-x86 in VMware Workstation, and set the virtual machine
1G memory. 


Then I followed steps guided by "Installing CDH4 on a Single Linux Node in Pseudo-distributed
Mode" —— https://ccp.cloudera.com/display/CDH4DOC/Installing+CDH4+on+a+Single+Linux+Node+in+Pseudo-distributed+Mode.


When at last, I ran an example Hadoop job with the command "$ hadoop jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar
grep input output23 'dfs[a-z.]+'"


then the screen showed as follows, 
depending "AttemptID:attempt_1360528029309_0001_r_000000_0 Timed out after 600 secs" and I
wonder is that because my virtual machine's memory too little~~??


[hadoop@localhost hadoop-mapreduce]$ hadoop jar /usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar
grep input output23 'dfs[a-z]+'                                                          
                                
13/02/11 04:30:44 WARN mapreduce.JobSubmitter: No job jar file set.  User classes may not
be found. See Job or Job#setJar(String).                                                 
                                              
13/02/11 04:30:44 INFO input.FileInputFormat: Total input paths to process : 4           
                       
13/02/11 04:30:45 INFO mapreduce.JobSubmitter: number of splits:4                        
                       
13/02/11 04:30:45 WARN conf.Configuration: mapred.output.value.class is deprecated. Instead,
use mapreduce.job.output.value.class                                                     
                                           
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.combine.class is deprecated. Instead,
use mapreduce.job.combine.class                                                          
                                             
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.map.class is deprecated. Instead, use
mapreduce.job.map.class                                                                  
                                             
13/02/11 04:30:45 WARN conf.Configuration: mapred.job.name is deprecated. Instead, use mapreduce.job.name
       
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.reduce.class is deprecated. Instead,
use mapreduce.job.reduce.class                                                           
                                              
13/02/11 04:30:45 WARN conf.Configuration: mapred.input.dir is deprecated. Instead, use mapreduce.input.fileinputformat.inputdir
                                                                                         
       
13/02/11 04:30:45 WARN conf.Configuration: mapred.output.dir is deprecated. Instead, use mapreduce.output.fileoutputformat.outputdir
                                                                                         
   
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.outputformat.class is deprecated. Instead,
use mapreduce.job.outputformat.class                                                     
                                        
13/02/11 04:30:45 WARN conf.Configuration: mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps
      
13/02/11 04:30:45 WARN conf.Configuration: mapred.output.key.class is deprecated. Instead,
use mapreduce.job.output.key.class                                                       
                                             
13/02/11 04:30:45 WARN conf.Configuration: mapred.working.dir is deprecated. Instead, use
mapreduce.job.working.dir                                                                
                                              
13/02/11 04:30:46 INFO mapred.YARNRunner: Job jar is not present. Not adding any jar to the
list of resources.   
13/02/11 04:30:46 INFO mapred.ResourceMgrDelegate: Submitted application application_1360528029309_0001
to ResourceManager at /0.0.0.0:8032                                                      
                                
13/02/11 04:30:46 INFO mapreduce.Job: The url to track the job: http://localhost.localdomain:8088/proxy/application_1360528029309_0001/
                                                                                         

13/02/11 04:30:46 INFO mapreduce.Job: Running job: job_1360528029309_0001                
                       
13/02/11 04:31:01 INFO mapreduce.Job: Job job_1360528029309_0001 running in uber mode : false
                   
13/02/11 04:31:01 INFO mapreduce.Job:  map 0% reduce 0%                                  
                       
13/02/11 04:47:22 INFO mapreduce.Job: Task Id : attempt_1360528029309_0001_r_000000_0, Status
: FAILED           
AttemptID:attempt_1360528029309_0001_r_000000_0 Timed out after 600 secs                 
                       
cleanup failed for container container_1360528029309_0001_01_000006 : java.lang.reflect.UndeclaredThrowableException
                                                                                         
                   
        at org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl.unwrapAndThrowException(YarnRemoteExceptionPBImpl.java:135)
                                                                                       
        at org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagerPBClientImpl.stopContainer(ContainerManagerPBClientImpl.java:114)
                                                                                         
 
        at org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.kill(ContainerLauncherImpl.java:209)
                                                                                         
            
        at org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:394)
                                                                                         
        
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
                      
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
                      
        at java.lang.Thread.run(Thread.java:722)                                         
                       
Caused by: com.google.protobuf.ServiceException: java.net.SocketTimeoutException: Call From
localhost.localdomain/127.0.0.1 to localhost.localdomain:54113 failed on socket timeout exception:
java.net.SocketTimeoutException: 60000 millis timeout while waiting for channel to be ready
for read. ch : java.nio.channels.SocketChannel[connected local=/127.0.0.1:60976 remote=localhost.localdomain/127.0.0.1:54113];
For more details see:  http://wiki.apache.org/hadoop/SocketTimeout                       
                                                                 
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:212)
                   
        at $Proxy29.stopContainer(Unknown Source)                                        
                       
        at org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagerPBClientImpl.stopContainer(ContainerManagerPBClientImpl.java:111)
                                                                                         
 
        ... 5 more                                                                       
                       
Caused by: java.net.SocketTimeoutException: Call From localhost.localdomain/127.0.0.1 to localhost.localdomain:54113
failed on socket timeout exception: java.net.SocketTimeoutException: 60000 millis timeout
while waiting for channel to be ready for read. ch : java.nio.channels.SocketChannel[connected
local=/127.0.0.1:60976 remote=localhost.localdomain/127.0.0.1:54113]; For more details see:
 http://wiki.apache.org/hadoop/SocketTimeout              
        at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:742)               
                       
        at org.apache.hadoop.ipc.Client.call(Client.java:1228)                           
                       
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
                   
        ... 7 more                                                                       
                       
Caused by: java.net.SocketTimeoutException: 60000 millis timeout while waiting for channel
to be ready for read. ch : java.nio.channels.SocketChannel[connected local=/127.0.0.1:60976
remote=localhost.localdomain/127.0.0.1:54113]                                            
                                                                  
        at org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:165)  
                       
        at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:154)      
                       
        at org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:127)      
                       
        at java.io.FilterInputStream.read(FilterInputStream.java:133)                    
                       
        at java.io.FilterInputStream.read(FilterInputStream.java:133)                    
                       
        at org.apache.hadoop.ipc.Client$Connection$PingInputStream.read(Client.java:408) 
                       
        at java.io.BufferedInputStream.fill(BufferedInputStream.java:235)                
                       
        at java.io.BufferedInputStream.read(BufferedInputStream.java:254)                
                       
        at java.io.FilterInputStream.read(FilterInputStream.java:83)                     
                       
        at com.google.protobuf.AbstractMessageLite$Builder.mergeDelimitedFrom(AbstractMessageLite.java:276)
     
        at com.google.protobuf.AbstractMessage$Builder.mergeDelimitedFrom(AbstractMessage.java:760)
             
        at com.google.protobuf.AbstractMessageLite$Builder.mergeDelimitedFrom(AbstractMessageLite.java:288)
     
        at com.google.protobuf.AbstractMessage$Builder.mergeDelimitedFrom(AbstractMessage.java:752)
             
        at org.apache.hadoop.ipc.protobuf.RpcPayloadHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcPayloadHeaderProtos.java:985)
                                                                                      
        at org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:937)      
                       
        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:835)
..........



Mime
View raw message