This looks like a hard-coded hdfs prefix in a path name construction somewhere.
On Wed, Nov 9, 2011 at 8:27 AM, motta <[email protected]> wrote: > Hi everybody, > I have tried to run my first Mahout SVD Job (DistributedLanczosSolver) in > Elastic Map Reduce. > Before going to Amazon I've tried to run it on my local hadoop installation > and it works, so the problem is just on Amazon. > > Using the online Amazon Management console i passed the following > parameters > JAR: s3n://<my-bucket-name>/mahout-examples-0.5-job.jar > JAR ARGUMENTS: > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver --input > s3n://<my-bucket-name>/input --output s3n://<my-bucket-name>/output > --numCols 24 --numRows 913327 --cleansvd true --rank 24 > > The reason why it fails is reported here. Thank you for your help > > Exception in thread "main" java.lang.IllegalArgumentException: This file > system object (hdfs://ip-10-227-133-187.eu-west-1.compute.internal:9000) > does not support access to the request path > 's3n://polimi-neodata-svd/input' > You possibly called FileSystem.get(conf) when you should have called > FileSystem.get(uri, conf) to obtain a file system supporting your path. > at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:351) > at > > org.apache.hadoop.hdfs.DistributedFileSystem.checkPath(DistributedFileSystem.java:99) > at > org.apache.hadoop.fs.FileSystem.makeQualified(FileSystem.java:263) > at > > org.apache.hadoop.hdfs.DistributedFileSystem.makeQualified(DistributedFileSystem.java:116) > at > > org.apache.mahout.math.hadoop.DistributedRowMatrix.setConf(DistributedRowMatrix.java:106) > at > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:198) > at > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:161) > at > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:120) > at > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver$DistributedLanczosSolverJob.run(DistributedLanczosSolver.java:290) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > at > > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.main(DistributedLanczosSolver.java:296) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > at > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:597) > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/Running-Mahout-SVD-on-Amazon-Elastic-Map-Reduce-tp3493514p3493514.html > Sent from the Mahout User List mailing list archive at Nabble.com. >
