Please ask any questions if I am not clear above about the problem I am facing.
Thanks, Akhil akhil1988 wrote: > > Hi All! > > I want a directory to be present in the local working directory of the > task for which I am using the following statements: > > DistributedCache.addCacheArchive(new URI("/home/akhil1988/Config.zip"), > conf); > DistributedCache.createSymlink(conf); > >>> Here Config is a directory which I have zipped and put at the given >>> location in HDFS > > I have zipped the directory because the API doc of DistributedCache > (http://hadoop.apache.org/core/docs/r0.20.0/api/index.html) says that the > archive files are unzipped in the local cache directory : > > DistributedCache can be used to distribute simple, read-only data/text > files and/or more complex types such as archives, jars etc. Archives (zip, > tar and tgz/tar.gz files) are un-archived at the slave nodes. > > So, from my understanding of the API docs I expect that the Config.zip > file will be unzipped to Config directory and since I have SymLinked them > I can access the directory in the following manner from my map function: > > FileInputStream fin = new FileInputStream("Config/file1.config"); > > But I get the FileNotFoundException on the execution of this statement. > Please let me know where I am going wrong. > > Thanks, > Akhil > -- View this message in context: http://www.nabble.com/Using-addCacheArchive-tp24207739p24210836.html Sent from the Hadoop core-user mailing list archive at Nabble.com.