Hi all,
I installed a redhat_enterprise-linux-x86 in VMware Workstation, and set the 
virtual machine 1G memory. 


Then I followed steps guided by "Installing CDH4 on a Single Linux Node in 
Pseudo-distributed Mode" —— 
https://ccp.cloudera.com/display/CDH4DOC/Installing+CDH4+on+a+Single+Linux+Node+in+Pseudo-distributed+Mode.


When at last, I ran an example Hadoop job with the command "$ hadoop jar 
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar grep input output23 
'dfs[a-z.]+'"


then the screen showed as follows, 
depending "AttemptID:attempt_1360528029309_0001_r_000000_0 Timed out after 600 
secs" and I wonder is that because my virtual machine's memory too little~~??


[hadoop@localhost hadoop-mapreduce]$ hadoop jar 
/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar grep input output23 
'dfs[a-z]+'                                                                     
                      
13/02/11 04:30:44 WARN mapreduce.JobSubmitter: No job jar file set.  User 
classes may not be found. See Job or Job#setJar(String).                        
                                                                        
13/02/11 04:30:44 INFO input.FileInputFormat: Total input paths to process : 4  
                                 
13/02/11 04:30:45 INFO mapreduce.JobSubmitter: number of splits:4               
                                 
13/02/11 04:30:45 WARN conf.Configuration: mapred.output.value.class is 
deprecated. Instead, use mapreduce.job.output.value.class                       
                                                                          
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.combine.class is 
deprecated. Instead, use mapreduce.job.combine.class                            
                                                                            
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.map.class is deprecated. 
Instead, use mapreduce.job.map.class                                            
                                                                    
13/02/11 04:30:45 WARN conf.Configuration: mapred.job.name is deprecated. 
Instead, use mapreduce.job.name        
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.reduce.class is 
deprecated. Instead, use mapreduce.job.reduce.class                             
                                                                             
13/02/11 04:30:45 WARN conf.Configuration: mapred.input.dir is deprecated. 
Instead, use mapreduce.input.fileinputformat.inputdir                           
                                                                       
13/02/11 04:30:45 WARN conf.Configuration: mapred.output.dir is deprecated. 
Instead, use mapreduce.output.fileoutputformat.outputdir                        
                                                                      
13/02/11 04:30:45 WARN conf.Configuration: mapreduce.outputformat.class is 
deprecated. Instead, use mapreduce.job.outputformat.class                       
                                                                       
13/02/11 04:30:45 WARN conf.Configuration: mapred.map.tasks is deprecated. 
Instead, use mapreduce.job.maps       
13/02/11 04:30:45 WARN conf.Configuration: mapred.output.key.class is 
deprecated. Instead, use mapreduce.job.output.key.class                         
                                                                            
13/02/11 04:30:45 WARN conf.Configuration: mapred.working.dir is deprecated. 
Instead, use mapreduce.job.working.dir                                          
                                                                     
13/02/11 04:30:46 INFO mapred.YARNRunner: Job jar is not present. Not adding 
any jar to the list of resources.   
13/02/11 04:30:46 INFO mapred.ResourceMgrDelegate: Submitted application 
application_1360528029309_0001 to ResourceManager at /0.0.0.0:8032              
                                                                         
13/02/11 04:30:46 INFO mapreduce.Job: The url to track the job: 
http://localhost.localdomain:8088/proxy/application_1360528029309_0001/         
                                                                                
  
13/02/11 04:30:46 INFO mapreduce.Job: Running job: job_1360528029309_0001       
                                 
13/02/11 04:31:01 INFO mapreduce.Job: Job job_1360528029309_0001 running in 
uber mode : false                    
13/02/11 04:31:01 INFO mapreduce.Job:  map 0% reduce 0%                         
                                 
13/02/11 04:47:22 INFO mapreduce.Job: Task Id : 
attempt_1360528029309_0001_r_000000_0, Status : FAILED           
AttemptID:attempt_1360528029309_0001_r_000000_0 Timed out after 600 secs        
                                 
cleanup failed for container container_1360528029309_0001_01_000006 : 
java.lang.reflect.UndeclaredThrowableException                                  
                                                                            
        at 
org.apache.hadoop.yarn.exceptions.impl.pb.YarnRemoteExceptionPBImpl.unwrapAndThrowException(YarnRemoteExceptionPBImpl.java:135)
                                                                                
        
        at 
org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagerPBClientImpl.stopContainer(ContainerManagerPBClientImpl.java:114)
                                                                                
            
        at 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$Container.kill(ContainerLauncherImpl.java:209)
                                                                                
                       
        at 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncherImpl$EventProcessor.run(ContainerLauncherImpl.java:394)
                                                                                
                   
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110) 
                      
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603) 
                      
        at java.lang.Thread.run(Thread.java:722)                                
                                 
Caused by: com.google.protobuf.ServiceException: 
java.net.SocketTimeoutException: Call From localhost.localdomain/127.0.0.1 to 
localhost.localdomain:54113 failed on socket timeout exception: 
java.net.SocketTimeoutException: 60000 millis timeout while waiting for channel 
to be ready for read. ch : java.nio.channels.SocketChannel[connected 
local=/127.0.0.1:60976 remote=localhost.localdomain/127.0.0.1:54113]; For more 
details see:  http://wiki.apache.org/hadoop/SocketTimeout                       
                                                                  
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:212)
                    
        at $Proxy29.stopContainer(Unknown Source)                               
                                 
        at 
org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagerPBClientImpl.stopContainer(ContainerManagerPBClientImpl.java:111)
                                                                                
            
        ... 5 more                                                              
                                 
Caused by: java.net.SocketTimeoutException: Call From 
localhost.localdomain/127.0.0.1 to localhost.localdomain:54113 failed on socket 
timeout exception: java.net.SocketTimeoutException: 60000 millis timeout while 
waiting for channel to be ready for read. ch : 
java.nio.channels.SocketChannel[connected local=/127.0.0.1:60976 
remote=localhost.localdomain/127.0.0.1:54113]; For more details see:  
http://wiki.apache.org/hadoop/SocketTimeout              
        at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:742)      
                                 
        at org.apache.hadoop.ipc.Client.call(Client.java:1228)                  
                                 
        at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:202)
                    
        ... 7 more                                                              
                                 
Caused by: java.net.SocketTimeoutException: 60000 millis timeout while waiting 
for channel to be ready for read. ch : 
java.nio.channels.SocketChannel[connected local=/127.0.0.1:60976 
remote=localhost.localdomain/127.0.0.1:54113]                                   
                                                                            
        at 
org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:165)    
                      
        at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:154)        
                      
        at 
org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:127)        
                      
        at java.io.FilterInputStream.read(FilterInputStream.java:133)           
                                 
        at java.io.FilterInputStream.read(FilterInputStream.java:133)           
                                 
        at 
org.apache.hadoop.ipc.Client$Connection$PingInputStream.read(Client.java:408)   
                      
        at java.io.BufferedInputStream.fill(BufferedInputStream.java:235)       
                                 
        at java.io.BufferedInputStream.read(BufferedInputStream.java:254)       
                                 
        at java.io.FilterInputStream.read(FilterInputStream.java:83)            
                                 
        at 
com.google.protobuf.AbstractMessageLite$Builder.mergeDelimitedFrom(AbstractMessageLite.java:276)
      
        at 
com.google.protobuf.AbstractMessage$Builder.mergeDelimitedFrom(AbstractMessage.java:760)
              
        at 
com.google.protobuf.AbstractMessageLite$Builder.mergeDelimitedFrom(AbstractMessageLite.java:288)
      
        at 
com.google.protobuf.AbstractMessage$Builder.mergeDelimitedFrom(AbstractMessage.java:752)
              
        at 
org.apache.hadoop.ipc.protobuf.RpcPayloadHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcPayloadHeaderProtos.java:985)
                                                                                
       
        at 
org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:937)        
                      
        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:835)
..........



Reply via email to