How are you putting files in distributed cache ? Sent from my iPhone
> On Mar 27, 2014, at 9:20 AM, Jonathan Poon <[email protected]> wrote: > > > Hi Stanley, > > Sorry about the confusion, but I'm trying to read a txt file into my Mapper > function. I am trying to copy the file using the -files option when > submitting the Hadoop job. > > I try to obtain the filename using the following lines of code in my Mapper: > > URI[] localPaths = context.getCacheFiles(); > String configFilename = localPaths[0].toString(); > > However, when I run the JAR in hadoop, I get a NullPointerException. > > Error: java.lang.NullPointerException > > I'm running Hadoop 2.2 in Single Node mode. Not sure if that affects > things... > > > > >> On Wed, Mar 26, 2014 at 8:21 PM, Stanley Shi <[email protected]> wrote: >> where did you get the error? from the compiler or the runtime? >> >> Regards, >> Stanley Shi, >> >> >> >>> On Thu, Mar 27, 2014 at 7:34 AM, Jonathan Poon <[email protected]> wrote: >>> Hi Everyone, >>> >>> I'm submitting a MapReduce job using the -files option to copy a text file >>> that contains properties I use for the map and reduce functions. >>> >>> I'm trying to obtain the local cache files in my mapper function using: >>> >>> Path[] paths = context.getLocalCacheFiles(); >>> >>> However, i get an error saying getLocalCacheFiles() is undefined. I've >>> imported the hadoop-mapreduce-client-core-2.2.0.jar as part of my build >>> environment in Eclipse. >>> >>> Any ideas on what could be incorrect? >>> >>> If I'm incorrectly using the distributed cache, could someone point me to >>> an example using the distributed cache with Hadoop 2.2.0? >>> >>> Thanks for your help! >>> >>> Jonathan >> >
