I'm using Hadoop 1.0.4 and Spark 1.2.0. I'm facing a strange issue. I have a requirement to read a small file from HDFS and all it's content has to be read at one shot. So I'm using spark context's wholeTextFiles API passing the HDFS URL for the file.
When I try this from a spark shell it's works as mentioned in the documentation, but when I try the same through program (by submitting job to cluster) I get FileNotFoundException. I have all compatible JARs in place. Please help.