hi Mohan, could you please tell me the hadoop version and the spark version on which you are working on.
On Mon, Feb 10, 2014 at 3:37 PM, Amit Behera <amit.bd...@gmail.com> wrote: > Please go to hadoop configuration directory and open core-site.xml and > check the IP and port for HDFS, mentioned on the value of "fs.default.name" > . After that specify the same IP and Port number on your code. > format *hdfs://<ip>:port/* > > I hope it will work..... > > > On Mon, Feb 10, 2014 at 2:14 PM, mohankreddy <mre...@beanatomics.com>wrote: > >> I am getting the following error when trying to access my data using >> hdfs:// >> ....... Not sure how to fix this one. >> >> " java.io.IOException: Call to server1/10.85.85.17:9000 failed on local >> exception: java.io.EOFException >> at org.apache.hadoop.ipc.Client.wrapException(Client.java:1107) >> at org.apache.hadoop.ipc.Client.call(Client.java:1075) >> at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:225) >> at $Proxy8.getProtocolVersion(Unknown Source) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:396) >> at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:379) >> at >> org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:119) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:238) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:203) >> at >> >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:89) >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1386) >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66) >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1404) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254) >> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:187) >> at >> >> org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:176) >> at >> >> org.apache.hadoop.mapred.FileInputFormat.getSplits(FileInputFormat.java:208) >> at >> org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:140) >> at >> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:207) >> at >> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205) >> at scala.Option.getOrElse(Option.scala:120) >> at org.apache.spark.rdd.RDD.partitions(RDD.scala:205) >> at >> org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28) >> at >> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:207) >> at >> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205) >> at scala.Option.getOrElse(Option.scala:120) >> at org.apache.spark.rdd.RDD.partitions(RDD.scala:205) >> at >> org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28) >> at >> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:207) >> at >> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:205) >> at scala.Option.getOrElse(Option.scala:120) >> at org.apache.spark.rdd.RDD.partitions(RDD.scala:205) >> at org.apache.spark.mllib.recommendation.ALS.run(ALS.scala:139) >> at org.apache.spark.mllib.recommendation.ALS$.main(ALS.scala:594) >> at org.apache.spark.mllib.recommendation.ALS.main(ALS.scala) >> Caused by: java.io.EOFException >> at java.io.DataInputStream.readInt(DataInputStream.java:375) >> at >> org.apache.hadoop.ipc.Client$Connection.receiveResponse(Client.ja >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/EOF-Exception-when-trying-to-access-hdfs-tp1347.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> > >