Dear friends I am new to hadoop on aws. I am using AWS Elastic mapreduce. I am 
trying to convert an old java program to elastic mapreduce. I would appreciate 
your help. My question is about:
1. How do I pass a "Properties" file which i used to pass via command line 
paramters - to elastic mapreduce?
2. how do I pass parameters from this properties file from main() to the Mapper 
and Reducer? Details:
1. properties file: 
My program used to read a properties file as follows:
org.apache.commons.configuration.Configuration config = new 
PropertiesConfiguration("my_app.properties");
How do I read this in Amazon EMR?
I tried launching the program with :
jar file: s3n://my.bucket.name/myjar-job.jar
args: s3n://my.bucket.name/my_app.properties I get an stderr of:
org.apache.commons.configuration.ConfigurationException: Cannot locate 
configuration source s3n://my.bucket.name/my_app.properties In order to debug, 
I tried to do the following:
String c = FileUtils.readFileToString(new File(remainingArgs[0]));
I got an exception:
Exception in thread "main" java.io.FileNotFoundException: File 
's3n:/my.bucket.name/my_app.propertiess' does not exist
Notice that the file name - even though I entered s3n:// is shown as s3n:/ - 
could the // be getting dropped?  2. Passing the properties file from my main() 
to Mapper and Reducer. How do I do that? Or pass other generic parameters?
-Sky                                      

Reply via email to