I have been able to run the PageRank example on my Apache pseudo cluster on my 
laptop (1.0.x)
But unable to run it on my dev cluster running Hortonworks 2.0 (I built 0.6.4 
src  with  mvn clean install -Phadoop2 -Dhadoop.version=2.2.0
)
I get the following error - (even though I have no trouble with hdfs in 
putting/getting files)

attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 WARN conf.Configuration: 
/tmp/hadoop-kiru/bsp/local/groomServer/attempt_201405201413_0006_000000_0/job.xml:an
 attempt to override final parameter: 
mapreduce.job.end-notification.max.attempts;  Ignoring.
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: Starting 
Socket Reader #1 for port 61001
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server Responder: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server listener on 61001: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server handler 0 on 61001: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server handler 1 on 61001: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server handler 2 on 61001: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server handler 3 on 61001: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO 
message.HamaMessageManagerImpl: BSPPeer address:server02.infnet port:61001
attempt_201405201413_0006_000000_0: 14/05/20 14:41:31 INFO ipc.Server: IPC 
Server handler 4 on 61001: starting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:32 INFO sync.ZKSyncClient: 
Initializing ZK Sync Client
attempt_201405201413_0006_000000_0: 14/05/20 14:41:32 INFO 
sync.ZooKeeperSyncClientImpl: Start connecting to Zookeeper! At 
server02.infnet/192.168.1.85:61001
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 WARN hdfs.DFSClient: 
DataStreamer Exception
attempt_201405201413_0006_000000_0: 
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
/tmp/hama-parts/job_201405201413_0005/part-3/file-0 could only be replicated to 
0 nodes instead of minReplication (=1).  There are 1 datanode(s) running and no 
node(s) are excluded in this operation.
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2503)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2053)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
attempt_201405201413_0006_000000_0: at 
java.security.AccessController.doPrivileged(Native Method)
attempt_201405201413_0006_000000_0: at 
javax.security.auth.Subject.doAs(Subject.java:415)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2047)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1347)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1300)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
attempt_201405201413_0006_000000_0: at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
attempt_201405201413_0006_000000_0: at 
java.lang.reflect.Method.invoke(Method.java:606)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1226)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514)
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 ERROR bsp.BSPTask: Error 
running bsp setup and bsp function.
attempt_201405201413_0006_000000_0: 
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
/tmp/hama-parts/job_201405201413_0005/part-3/file-0 could only be replicated to 
0 nodes instead of minReplication (=1).  There are 1 datanode(s) running and no 
node(s) are excluded in this operation.
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2503)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2053)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
attempt_201405201413_0006_000000_0: at 
java.security.AccessController.doPrivileged(Native Method)
attempt_201405201413_0006_000000_0: at 
javax.security.auth.Subject.doAs(Subject.java:415)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2047)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1347)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1300)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
attempt_201405201413_0006_000000_0: at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
attempt_201405201413_0006_000000_0: at 
java.lang.reflect.Method.invoke(Method.java:606)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1226)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514)
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: Stopping 
server on 61001
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: IPC 
Server handler 1 on 61001: exiting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: IPC 
Server handler 0 on 61001: exiting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: IPC 
Server handler 2 on 61001: exiting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: IPC 
Server handler 4 on 61001: exiting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: IPC 
Server handler 3 on 61001: exiting
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: Stopping 
IPC Server listener on 61001
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO ipc.Server: Stopping 
IPC Server Responder
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 INFO 
Configuration.deprecation: mapred.cache.localFiles is deprecated. Instead, use 
mapreduce.job.cache.local.files
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 ERROR bsp.BSPTask: 
Shutting down ping service.
attempt_201405201413_0006_000000_0: 14/05/20 14:41:33 FATAL bsp.GroomServer: 
Error running child
attempt_201405201413_0006_000000_0: 
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
/tmp/hama-parts/job_201405201413_0005/part-3/file-0 could only be replicated to 
0 nodes instead of minReplication (=1).  There are 1 datanode(s) running and no 
node(s) are excluded in this operation.
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2503)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2053)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
attempt_201405201413_0006_000000_0: at 
java.security.AccessController.doPrivileged(Native Method)
attempt_201405201413_0006_000000_0: at 
javax.security.auth.Subject.doAs(Subject.java:415)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2047)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1347)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1300)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
attempt_201405201413_0006_000000_0: at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
attempt_201405201413_0006_000000_0: at 
java.lang.reflect.Method.invoke(Method.java:606)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1226)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514)
attempt_201405201413_0006_000000_0: 
org.apache.hadoop.ipc.RemoteException(java.io.IOException): File 
/tmp/hama-parts/job_201405201413_0005/part-3/file-0 could only be replicated to 
0 nodes instead of minReplication (=1).  There are 1 datanode(s) running and no 
node(s) are excluded in this operation.
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2503)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:59582)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2053)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2049)
attempt_201405201413_0006_000000_0: at 
java.security.AccessController.doPrivileged(Native Method)
attempt_201405201413_0006_000000_0: at 
javax.security.auth.Subject.doAs(Subject.java:415)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Server$Handler.run(Server.java:2047)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1347)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.Client.call(Client.java:1300)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
attempt_201405201413_0006_000000_0: at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
attempt_201405201413_0006_000000_0: at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
attempt_201405201413_0006_000000_0: at 
java.lang.reflect.Method.invoke(Method.java:606)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102)
attempt_201405201413_0006_000000_0: at com.sun.proxy.$Proxy10.addBlock(Unknown 
Source)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.addBlock(ClientNamenodeProtocolTranslatorPB.java:330)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1226)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078)
attempt_201405201413_0006_000000_0: at 
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514)
attempt_201405201413_0006_000000_0: log4j:WARN No appenders could be found for 
logger (org.apache.hadoop.hdfs.DFSClient).
attempt_201405201413_0006_000000_0: log4j:WARN Please initialize the log4j 
system properly.
attempt_201405201413_0006_000000_0: log4j:WARN See 
http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info.
14/05/20 14:42:40 INFO bsp.BSPJobClient: Job failed.
14/05/20 14:42:40 ERROR bsp.BSPJobClient: Error partitioning the input path.
java.io.IOException: Runtime partition failed for the job.
at org.apache.hama.bsp.BSPJobClient.partition(BSPJobClient.java:478)
at org.apache.hama.bsp.BSPJobClient.submitJobInternal(BSPJobClient.java:341)
at org.apache.hama.bsp.BSPJobClient.submitJob(BSPJobClient.java:296)
at org.apache.hama.bsp.BSPJob.submit(BSPJob.java:219)
at org.apache.hama.graph.GraphJob.submit(GraphJob.java:208)
at org.apache.hama.bsp.BSPJob.waitForCompletion(BSPJob.java:226)
at org.apache.hama.examples.PageRank.main(PageRank.java:160)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at 
org.apache.hama.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68)
at org.apache.hama.util.ProgramDriver.driver(ProgramDriver.java:139)
at org.apache.hama.examples.ExampleDriver.main(ExampleDriver.java:45)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.hama.util.RunJar.main(RunJar.java:146)


 
Regards,
- kiru


Kiru Pakkirisamy | webcloudtech.wordpress.com

Reply via email to