I'm use 0.20.2 and 0.20.2-cdh versions. There's no problem. Sent from my iPad
On Sep 12, 2012, at 4:59 PM, Thomas Jungblut <[email protected]> wrote: > Anyone tested the compatibility to Hadoop 20.2 with Hama 0.5? > [1] says it is compatible. > > [1] http://wiki.apache.org/hama/CompatibilityTable > > 2012/9/12 顾荣 <[email protected]> > >> Okay, I'll try Hadoop 1.0.3 with Hama 0.50. >> Thanks Thomas. I can't wait to explore the Hama world now. >> >> walker. >> >> 2012/9/12 Thomas Jungblut <[email protected]> >> >>> oh okay. I'm not sure if 0.5.0 is really compatible to 20.2, personally I >>> have installed 1.0.3 and it works fine. >>> Sorry to let you install all the different versions. >>> >>> 2012/9/12 顾荣 <[email protected]> >>> >>>> Thanks Thomas. The HDFS works well. I even put a file from local to it >>>> successfully. It absolutely left the safemode. The namenode starting >> log >>> is >>>> as below: >>>> >>>> 2012-09-12 15:10:39,002 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: STARTUP_MSG: >>>> /************************************************************ >>>> STARTUP_MSG: Starting NameNode >>>> STARTUP_MSG: host = slave021/192.168.1.21 >>>> STARTUP_MSG: args = [] >>>> STARTUP_MSG: version = 0.20.2 >>>> STARTUP_MSG: build = >>>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-0.20 -r >>>> 911707; compiled by 'chrisdo' on Fri Feb 19 08:07:34 UTC 2010 >>>> ************************************************************/ >>>> 2012-09-12 15:10:39,092 INFO org.apache.hadoop.ipc.metrics.RpcMetrics: >>>> Initializing RPC Metrics with hostName=NameNode, port=54310 >>>> 2012-09-12 15:10:39,098 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Namenode up at: >>> slave021/ >>>> 192.168.1.21:54310 >>>> 2012-09-12 15:10:39,100 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: >>>> Initializing JVM Metrics with processName=NameNode, sessionId=null >>>> 2012-09-12 15:10:39,101 INFO >>>> org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics: >>>> Initializing NameNodeMeterics using context >>>> object:org.apache.hadoop.metrics.spi.NullContext >>>> 2012-09-12 15:10:39,143 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> fsOwner=hadoop,hadoop_user,wheel >>>> 2012-09-12 15:10:39,144 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>> supergroup=supergroup >>>> 2012-09-12 15:10:39,144 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: >>>> isPermissionEnabled=true >>>> 2012-09-12 15:10:39,150 INFO >>>> org.apache.hadoop.hdfs.server.namenode.metrics.FSNamesystemMetrics: >>>> Initializing FSNamesystemMetrics using context >>>> object:org.apache.hadoop.metrics.spi.NullContext >>>> 2012-09-12 15:10:39,151 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Registered >>>> FSNamesystemStatusMBean >>>> 2012-09-12 15:10:39,177 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: >>>> Number of files = 1 >>>> 2012-09-12 15:10:39,181 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: >>>> Number of files under construction = 0 >>>> 2012-09-12 15:10:39,181 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: >>>> Image file of size 96 loaded in 0 seconds. >>>> 2012-09-12 15:10:39,181 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: >>>> Edits file >>>> /home/hadoop/gurong/hadoop-0.20.2/hadoop_dir/dfs/name/current/edits of >>> size >>>> 4 edits # 0 loaded in 0 seconds. >>>> 2012-09-12 15:10:39,236 INFO >>> org.apache.hadoop.hdfs.server.common.Storage: >>>> Image file of size 96 saved in 0 seconds. >>>> 2012-09-12 15:10:39,439 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Finished loading >>>> FSImage in 312 msecs >>>> 2012-09-12 15:10:39,441 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Total number of >>> blocks >>>> = 0 >>>> 2012-09-12 15:10:39,441 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of invalid >>>> blocks = 0 >>>> 2012-09-12 15:10:39,441 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of >>>> under-replicated blocks = 0 >>>> 2012-09-12 15:10:39,441 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of >>>> over-replicated blocks = 0 >>>> 2012-09-12 15:10:39,441 INFO org.apache.hadoop.hdfs.StateChange: STATE* >>>> Leaving safe mode after 0 secs. >>>> 2012-09-12 15:10:39,441 INFO org.apache.hadoop.hdfs.StateChange: STATE* >>>> Network topology has 0 racks and 0 datanodes >>>> 2012-09-12 15:10:39,441 INFO org.apache.hadoop.hdfs.StateChange: STATE* >>>> UnderReplicatedBlocks has 0 blocks >>>> 2012-09-12 15:10:39,554 INFO org.mortbay.log: Logging to >>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >>>> org.mortbay.log.Slf4jLog >>>> 2012-09-12 15:10:39,603 INFO org.apache.hadoop.http.HttpServer: Port >>>> returned by webServer.getConnectors()[0].getLocalPort() before open() >> is >>>> -1. Opening the listener on 50070 >>>> 2012-09-12 15:10:39,604 INFO org.apache.hadoop.http.HttpServer: >>>> listener.getLocalPort() returned 50070 >>>> webServer.getConnectors()[0].getLocalPort() returned 50070 >>>> 2012-09-12 15:10:39,604 INFO org.apache.hadoop.http.HttpServer: Jetty >>> bound >>>> to port 50070 >>>> 2012-09-12 15:10:39,604 INFO org.mortbay.log: jetty-6.1.14 >>>> 2012-09-12 15:10:48,662 INFO org.mortbay.log: Started >>>> [email protected]:50070 >>>> 2012-09-12 15:10:48,663 INFO >>>> org.apache.hadoop.hdfs.server.namenode.NameNode: Web-server up at: >>>> 0.0.0.0:50070 >>>> 2012-09-12 15:10:48,666 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> listener on 54310: starting >>>> 2012-09-12 15:10:48,667 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> Responder: starting >>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 0 on 54310: starting >>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 1 on 54310: starting >>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 2 on 54310: starting >>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 3 on 54310: starting >>>> 2012-09-12 15:10:48,668 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 4 on 54310: starting >>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 5 on 54310: starting >>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 6 on 54310: starting >>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 7 on 54310: starting >>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 8 on 54310: starting >>>> 2012-09-12 15:10:48,669 INFO org.apache.hadoop.ipc.Server: IPC Server >>>> handler 9 on 54310: starting >>>> 2012-09-12 15:10:48,700 INFO org.apache.hadoop.ipc.Server: Error >> register >>>> getProtocolVersion >>>> java.lang.IllegalArgumentException: Duplicate >>>> metricsName:getProtocolVersion >>>> at >>>> >>> >> org.apache.hadoop.metrics.util.MetricsRegistry.add(MetricsRegistry.java:53) >>>> at >>>> >>>> >>> >> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:89) >>>> at >>>> >>>> >>> >> org.apache.hadoop.metrics.util.MetricsTimeVaryingRate.<init>(MetricsTimeVaryingRate.java:99) >>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:523) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) >>>> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at javax.security.auth.Subject.doAs(Subject.java:416) >>>> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) >>>> 2012-09-12 15:11:05,298 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* >>>> NameSystem.registerDatanode: node registration from >>>> 192.168.1.21:50010storage >>>> DS-1416037815-192.168.1.21-50010-1347433865293 >>>> 2012-09-12 15:11:05,300 INFO org.apache.hadoop.net.NetworkTopology: >>> Adding >>>> a new node: /default-rack/192.168.1.21:50010 >>>> 2012-09-12 15:11:15,069 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: >>>> ugi=webuser,webgroup ip=/192.168.1.21 cmd=listStatus src=/ >>>> dst=null perm=null >>>> 2012-09-12 15:12:05,034 WARN org.apache.hadoop.ipc.Server: Incorrect >>> header >>>> or version mismatch from 192.168.1.21:56281 got version 4 expected >>>> version 3 >>>> 2012-09-12 15:14:51,535 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: >>>> ugi=hadoop,hadoop_user,wheel ip=/192.168.1.21 cmd=listStatus >>>> src=/ dst=null perm=null >>>> 2012-09-12 15:15:10,158 INFO >>>> org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of >>>> transactions: 2 Total time for transactions(ms): 0Number of >> transactions >>>> batched in Syncs: 0 Number of syncs: 0 SyncTimes(ms): 0 >>>> >>>> I used Hama 0.5 and hadoop 0.20.2. Has somebody test this match can >> work >>>> well? >>>> >>>> thanks very much. >>>> >>>> walker >>>> >>>> 2012/9/12 Thomas Jungblut <[email protected]> >>>> >>>>> Still it says: >>>>>> >>>>>> " 2012-09-12 14:41:16,218 ERROR org.apache.hama.bsp.BSPMaster: >> Can't >>>> get >>>>>> connection to Hadoop Namenode! " >>>>> >>>>> >>>>> Can you verify that the namenode is not in safemode and has correctly >>>>> started up? >>>>> Have a look into the namenode logs please! >>>>> >>>>> 2012/9/12 顾荣 <[email protected]> >>>>> >>>>>> By the way, the fs.default.name is 192.168.1.21:54310. I checked >> the >>>>> HDFS, >>>>>> it works well. I installed and ran both HDFS and Hama using the >> same >>>>> linux >>>>>> account. >>>>>> >>>>>> 2012/9/12 顾荣 <[email protected]> >>>>>> >>>>>>> Thanks so much, Edward. >>>>>>> >>>>>>> I fellowed your suggestion and instanlled a hadoop 0.20.2 instead >>> for >>>>>>> Hama. However, this time when I start Hama, a fatal happened and >>> the >>>>>>> bspmaster daemon can not start up. The corresponding error >> message >>> in >>>>> the >>>>>>> baspmaster log file shows as below. >>>>>>> >>>>>>> ************************************************************/ >>>>>>> 2012-09-12 14:40:38,238 INFO org.apache.hama.BSPMasterRunner: >>>>>> STARTUP_MSG: >>>>>>> /************************************************************ >>>>>>> STARTUP_MSG: Starting BSPMaster >>>>>>> STARTUP_MSG: host = slave021/192.168.1.21 >>>>>>> STARTUP_MSG: args = [] >>>>>>> STARTUP_MSG: version = 1.0.0 >>>>>>> STARTUP_MSG: build = >>>>>>> >>> https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.0-r >>>>>>> 1214675; compiled by 'hortonfo' on Fri Dec 16 20:01:27 UTC 2011 >>>>>>> ************************************************************/ >>>>>>> 2012-09-12 14:40:38,414 INFO org.apache.hama.bsp.BSPMaster: RPC >>>>>> BSPMaster: >>>>>>> host slave021 port 40000 >>>>>>> 2012-09-12 14:40:38,502 INFO org.apache.hadoop.ipc.Server: >> Starting >>>>>>> SocketReader >>>>>>> 2012-09-12 14:40:38,542 INFO org.mortbay.log: Logging to >>>>>>> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >>>>>>> org.mortbay.log.Slf4jLog >>>>>>> 2012-09-12 14:40:38,583 INFO org.apache.hama.http.HttpServer: >> Port >>>>>>> returned by webServer.getConnectors()[0].getLocalPort() before >>> open() >>>>> is >>>>>>> -1. Opening the listener on 40013 >>>>>>> 2012-09-12 14:40:38,584 INFO org.apache.hama.http.HttpServer: >>>>>>> listener.getLocalPort() returned 40013 >>>>>>> webServer.getConnectors()[0].getLocalPort() returned 40013 >>>>>>> 2012-09-12 14:40:38,584 INFO org.apache.hama.http.HttpServer: >> Jetty >>>>> bound >>>>>>> to port 40013 >>>>>>> 2012-09-12 14:40:38,584 INFO org.mortbay.log: jetty-6.1.14 >>>>>>> 2012-09-12 14:40:38,610 INFO org.mortbay.log: Extract >>>>>>> >>>>>> >>>>> >>>> >>> >> jar:file:/home/hadoop/hama_installs/hama-0.5.0/hama-core-0.5.0.jar!/webapp/bspmaster/ >>>>>>> to /tmp/Jetty_slave021_40013_bspmaster____.1tzgsz/webapp >>>>>>> 2012-09-12 14:41:16,073 INFO org.mortbay.log: Started >>>>>>> SelectChannelConnector@slave021:40013 >>>>>>> 2012-09-12 14:41:16,218 ERROR org.apache.hama.bsp.BSPMaster: >> Can't >>>> get >>>>>>> connection to Hadoop Namenode! >>>>>>> java.io.IOException: Call to /192.168.1.21:54310 failed on local >>>>>>> exception: java.io.EOFException >>>>>>> at >> org.apache.hadoop.ipc.Client.wrapException(Client.java:1103) >>>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1071) >>>>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) >>>>>>> at $Proxy5.getProtocolVersion(Unknown Source) >>>>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396) >>>>>>> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379) >>>>>>> at >>>>>>> >>>> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119) >>>>>>> at >> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238) >>>>>>> at >> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203) >>>>>>> at >>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89) >>>>>>> at >>>>>>> >>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386) >>>>>>> at >>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >>>>>>> at >>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404) >>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) >>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:123) >>>>>>> at org.apache.hama.bsp.BSPMaster.<init>(BSPMaster.java:299) >>>>>>> at >>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:454) >>>>>>> at >>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:449) >>>>>>> at >> org.apache.hama.BSPMasterRunner.run(BSPMasterRunner.java:46) >>>>>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >>>>>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) >>>>>>> at >>> org.apache.hama.BSPMasterRunner.main(BSPMasterRunner.java:56) >>>>>>> Caused by: java.io.EOFException >>>>>>> at java.io.DataInputStream.readInt(DataInputStream.java:392) >>>>>>> at >>>>>>> >>>>> >>> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.java:800) >>>>>>> at >> org.apache.hadoop.ipc.Client$Connection.run(Client.java:745) >>>>>>> 2012-09-12 14:41:16,222 FATAL org.apache.hama.BSPMasterRunner: >>>>>>> java.lang.NullPointerException >>>>>>> at >>> org.apache.hama.bsp.BSPMaster.getSystemDir(BSPMaster.java:862) >>>>>>> at org.apache.hama.bsp.BSPMaster.<init>(BSPMaster.java:308) >>>>>>> at >>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:454) >>>>>>> at >>> org.apache.hama.bsp.BSPMaster.startMaster(BSPMaster.java:449) >>>>>>> at >> org.apache.hama.BSPMasterRunner.run(BSPMasterRunner.java:46) >>>>>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) >>>>>>> at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) >>>>>>> at >>> org.apache.hama.BSPMasterRunner.main(BSPMasterRunner.java:56) >>>>>>> >>>>>>> 2012-09-12 14:41:16,223 INFO org.apache.hama.BSPMasterRunner: >>>>>>> SHUTDOWN_MSG: >>>>>>> /************************************************************ >>>>>>> SHUTDOWN_MSG: Shutting down BSPMaster at slave021/192.168.1.21 >>>>>>> ************************************************************/ >>>>>>> >>>>>>> Would please give me some tips again? >>>>>>> >>>>>>> Thanks,again. >>>>>>> >>>>>>> walker >>>>>>> >>>>>>> 2012/9/12 Edward J. Yoon <[email protected]> >>>>>>> >>>>>>> Unfortunately we don't support Hadoop secure version yet. >>>>>>>> >>>>>>>> Instead of 0.20.205, Please use hadoop non-secure 0.20.2 or >> 1.0.3 >>>>>>>> versions. >>>>>>>> >>>>>>>> Thanks. >>>>>>>> >>>>>>>> On Wed, Sep 12, 2012 at 11:25 AM, 顾荣 <[email protected]> >>>> wrote: >>>>>>>>> Hi,all. >>>>>>>>> >>>>>>>>> I set up a hama cluster of 3 nodes and start hama >> successfully. >>>>>> However, >>>>>>>>> when I run the pi example, the job failed with a very strange >>>>> message >>>>>> as >>>>>>>>> below. >>>>>>>>> >>>>>>>>> hama jar >>>>>> /home/hadoop/hama_installs/hama-0.5.0/hama-examples-0.5.0.jar >>>>>>>> pi >>>>>>>>> org.apache.hadoop.ipc.RemoteException: java.io.IOException: >>>>>>>>> java.lang.NoSuchMethodException: >>>>>>>>> >>>>>> >>> org.apache.hadoop.hdfs.protocol.ClientProtocol.create(java.lang.String, >>>>>>>>> org.apache.hadoop.fs.permission.FsPermission, >> java.lang.String, >>>>>> boolean, >>>>>>>>> boolean, short, long) >>>>>>>>> at java.lang.Class.getMethod(Class.java:1605) >>>>>>>>> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:557) >>>>>>>>> at >>>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1388) >>>>>>>>> at >>>> org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1384) >>>>>>>>> at java.security.AccessController.doPrivileged(Native >>> Method) >>>>>>>>> at javax.security.auth.Subject.doAs(Subject.java:396) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1059) >>>>>>>>> at >>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:1382) >>>>>>>>> >>>>>>>>> at org.apache.hadoop.ipc.Client.call(Client.java:1066) >>>>>>>>> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) >>>>>>>>> at $Proxy2.create(Unknown Source) >>>>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>> Method) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:616) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) >>>>>>>>> at $Proxy2.create(Unknown Source) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.<init>(DFSClient.java:3245) >>>>>>>>> at >>> org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:713) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:182) >>>>>>>>> at >>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:555) >>>>>>>>> at >>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:536) >>>>>>>>> at >>> org.apache.hadoop.fs.FileSystem.create(FileSystem.java:443) >>>>>>>>> at org.apache.hadoop.fs.FileUtil.copy(FileUtil.java:229) >>>>>>>>> at >>>>>>>>> >>>>>> >>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1195) >>>>>>>>> at >>>>>>>>> >>>>>> >>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1171) >>>>>>>>> at >>>>>>>>> >>>>>> >>> org.apache.hadoop.fs.FileSystem.copyFromLocalFile(FileSystem.java:1143) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>> >> org.apache.hama.bsp.BSPJobClient.submitJobInternal(BSPJobClient.java:349) >>>>>>>>> at >>>>>> org.apache.hama.bsp.BSPJobClient.submitJob(BSPJobClient.java:294) >>>>>>>>> at org.apache.hama.bsp.BSPJob.submit(BSPJob.java:218) >>>>>>>>> at >>>> org.apache.hama.bsp.BSPJob.waitForCompletion(BSPJob.java:225) >>>>>>>>> at >>>>> org.apache.hama.examples.PiEstimator.main(PiEstimator.java:139) >>>>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>> Method) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:616) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68) >>>>>>>>> at >>>>>>>> >>> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139) >>>>>>>>> at >>>>>>>> >> org.apache.hama.examples.ExampleDriver.main(ExampleDriver.java:39) >>>>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>> Method) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>>>>> at >>>>>>>>> >>>>>>>> >>>>>> >>>>> >>>> >>> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:616) >>>>>>>>> at org.apache.hama.util.RunJar.main(RunJar.java:147) >>>>>>>>> >>>>>>>>> My hama verison is 0.5 and hadoop version is 0.20.205. This >>> error >>>>>> seems >>>>>>>> to >>>>>>>>> comes from the >>>>> "org.apache.hadoop.hdfs.protocol.ClientProtocol.create" >>>>>>>>> method, this is a normal method. I am kind of confused... >>>>>>>>> >>>>>>>>> Thanks in advance. >>>>>>>>> >>>>>>>>> walker >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards, Edward J. Yoon >>>>>>>> @eddieyoon >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >>
