I didn't hard-codec any hdfs prefix, I've just used mahout-examples-0.5-job.jar (downloaded from mahout website) to run DistributedLanczosSolver.
The output suggest that the jar invoked FileSystem.get(conf) instead of FileSystem.get(uri, conf) to get my input matrix is it possible? 2011/11/10 Ted Dunning <[email protected]> > This looks like a hard-coded hdfs prefix in a path name construction > somewhere. > > On Wed, Nov 9, 2011 at 8:27 AM, motta <[email protected]> wrote: > > > Hi everybody, > > I have tried to run my first Mahout SVD Job (DistributedLanczosSolver) in > > Elastic Map Reduce. > > Before going to Amazon I've tried to run it on my local hadoop > installation > > and it works, so the problem is just on Amazon. > > > > Using the online Amazon Management console i passed the following > > parameters > > JAR: s3n://<my-bucket-name>/mahout-examples-0.5-job.jar > > JAR ARGUMENTS: > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver --input > > s3n://<my-bucket-name>/input --output s3n://<my-bucket-name>/output > > --numCols 24 --numRows 913327 --cleansvd true --rank 24 > > > > The reason why it fails is reported here. Thank you for your help > > > > Exception in thread "main" java.lang.IllegalArgumentException: This file > > system object (hdfs://ip-10-227-133-187.eu-west-1.compute.internal:9000) > > does not support access to the request path > > 's3n://polimi-neodata-svd/input' > > You possibly called FileSystem.get(conf) when you should have called > > FileSystem.get(uri, conf) to obtain a file system supporting your path. > > at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:351) > > at > > > > > org.apache.hadoop.hdfs.DistributedFileSystem.checkPath(DistributedFileSystem.java:99) > > at > > org.apache.hadoop.fs.FileSystem.makeQualified(FileSystem.java:263) > > at > > > > > org.apache.hadoop.hdfs.DistributedFileSystem.makeQualified(DistributedFileSystem.java:116) > > at > > > > > org.apache.mahout.math.hadoop.DistributedRowMatrix.setConf(DistributedRowMatrix.java:106) > > at > > > > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:198) > > at > > > > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:161) > > at > > > > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:120) > > at > > > > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver$DistributedLanczosSolverJob.run(DistributedLanczosSolver.java:290) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > > at > > > > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.main(DistributedLanczosSolver.java:296) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > at > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > at java.lang.reflect.Method.invoke(Method.java:597) > > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > > > > -- > > View this message in context: > > > http://lucene.472066.n3.nabble.com/Running-Mahout-SVD-on-Amazon-Elastic-Map-Reduce-tp3493514p3493514.html > > Sent from the Mahout User List mailing list archive at Nabble.com. > > >
