What is your suggestion here, should I keep going on this quest to fix hdfs or should I try to run using HttpFileSystem? On 11 Aug 2014, at 23:01, Telles Nobrega <[email protected]> wrote:
> The port is right?? 50700. I have no idea what is happening now. > > On 11 Aug 2014, at 22:33, Telles Nobrega <[email protected]> wrote: > >> Right now the error is the following: >> Exception in thread "main" java.io.IOException: Failed on local exception: >> com.google.protobuf.InvalidProtocolBufferException: Protocol message >> end-group tag did not match expected tag.; Host Details : local host is: >> "telles-samza-master/10.1.0.79"; destination host is: >> "telles-samza-master":50070; >> at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:764) >> at org.apache.hadoop.ipc.Client.call(Client.java:1410) >> at org.apache.hadoop.ipc.Client.call(Client.java:1359) >> at >> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206) >> at com.sun.proxy.$Proxy14.getFileInfo(Unknown Source) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:606) >> at >> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:186) >> at >> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:102) >> at com.sun.proxy.$Proxy14.getFileInfo(Unknown Source) >> at >> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getFileInfo(ClientNamenodeProtocolTranslatorPB.java:671) >> at org.apache.hadoop.hdfs.DFSClient.getFileInfo(DFSClient.java:1746) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1112) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1108) >> at >> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1108) >> at >> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111) >> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) >> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) >> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) >> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) >> at org.apache.samza.job.JobRunner.main(JobRunner.scala) >> Caused by: com.google.protobuf.InvalidProtocolBufferException: Protocol >> message end-group tag did not match expected tag. >> at >> com.google.protobuf.InvalidProtocolBufferException.invalidEndTag(InvalidProtocolBufferException.java:94) >> at >> com.google.protobuf.CodedInputStream.checkLastTagWas(CodedInputStream.java:124) >> at >> com.google.protobuf.AbstractParser.parsePartialFrom(AbstractParser.java:202) >> at >> com.google.protobuf.AbstractParser.parsePartialDelimitedFrom(AbstractParser.java:241) >> at >> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:253) >> at >> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:259) >> at >> com.google.protobuf.AbstractParser.parseDelimitedFrom(AbstractParser.java:49) >> at >> org.apache.hadoop.ipc.protobuf.RpcHeaderProtos$RpcResponseHeaderProto.parseDelimitedFrom(RpcHeaderProtos.java:2364) >> at >> org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1051) >> at org.apache.hadoop.ipc.Client$Connection.run(Client.java:945) >> >> I feel that I’m close to making it run. Thanks for the help in advance. >> On 11 Aug 2014, at 22:06, Telles Nobrega <[email protected]> wrote: >> >>> Hi, I downloaded hadoop-common-2.3.0.jar and it worked better. Now I’m >>> having a configuration problem with my host, but it looks like the hdfs is >>> not a problem anymore. >>> >>> >>> >>> >>> On 11 Aug 2014, at 22:04, Telles Nobrega <[email protected]> wrote: >>> >>>> So, I added hadoop-hdfs-2.3.0.jar as a maven dependency. Recompiled the >>>> project, extracted to deploy/samza and there problem still happens. I >>>> downloaded hadoop-client-2.3.0.jar and the problems still happens, >>>> hadoop-common is 2.2.0 does this is a problem? I will try with 2.3.0 >>>> >>>> Actually a lot of hadoop jars are 2.2.0 >>>> >>>> On 11 Aug 2014, at 21:33, Yan Fang <[email protected]> wrote: >>>> >>>>> <include>org.apache.hadoop:hadoop-hdfs</include> >>>> >>> >> >
