I didn't hard-codec any hdfs prefix, I've just used mahout-examples-0.5-job.jar
(downloaded from mahout website) to run DistributedLanczosSolver.

The output suggest that the jar invoked FileSystem.get(conf) instead
of FileSystem.get(uri, conf) to get my input matrix

is it possible?

2011/11/10 Ted Dunning <[email protected]>

> This looks like a hard-coded hdfs prefix in a path name construction
> somewhere.
>
> On Wed, Nov 9, 2011 at 8:27 AM, motta <[email protected]> wrote:
>
> > Hi everybody,
> > I have tried to run my first Mahout SVD Job (DistributedLanczosSolver) in
> > Elastic Map Reduce.
> > Before going to Amazon I've tried to run it on my local hadoop
> installation
> > and it works, so the problem is just on Amazon.
> >
> > Using the online Amazon Management console i passed the following
> > parameters
> > JAR: s3n://<my-bucket-name>/mahout-examples-0.5-job.jar
> > JAR ARGUMENTS:
> > org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver --input
> > s3n://<my-bucket-name>/input --output s3n://<my-bucket-name>/output
> > --numCols 24 --numRows 913327 --cleansvd true --rank 24
> >
> > The reason why it fails is reported here. Thank you for your help
> >
> > Exception in thread "main" java.lang.IllegalArgumentException: This file
> > system object (hdfs://ip-10-227-133-187.eu-west-1.compute.internal:9000)
> > does not support access to the request path
> > 's3n://polimi-neodata-svd/input'
> > You possibly called FileSystem.get(conf) when you should have called
> > FileSystem.get(uri, conf) to obtain a file system supporting your path.
> >        at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:351)
> >        at
> >
> >
> org.apache.hadoop.hdfs.DistributedFileSystem.checkPath(DistributedFileSystem.java:99)
> >        at
> > org.apache.hadoop.fs.FileSystem.makeQualified(FileSystem.java:263)
> >        at
> >
> >
> org.apache.hadoop.hdfs.DistributedFileSystem.makeQualified(DistributedFileSystem.java:116)
> >        at
> >
> >
> org.apache.mahout.math.hadoop.DistributedRowMatrix.setConf(DistributedRowMatrix.java:106)
> >        at
> >
> >
> org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:198)
> >        at
> >
> >
> org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:161)
> >        at
> >
> >
> org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.run(DistributedLanczosSolver.java:120)
> >        at
> >
> >
> org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver$DistributedLanczosSolverJob.run(DistributedLanczosSolver.java:290)
> >        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
> >        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79)
> >        at
> >
> >
> org.apache.mahout.math.hadoop.decomposer.DistributedLanczosSolver.main(DistributedLanczosSolver.java:296)
> >        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >        at
> >
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> >        at
> >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> >        at java.lang.reflect.Method.invoke(Method.java:597)
> >        at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
> >
> > --
> > View this message in context:
> >
> http://lucene.472066.n3.nabble.com/Running-Mahout-SVD-on-Amazon-Elastic-Map-Reduce-tp3493514p3493514.html
> > Sent from the Mahout User List mailing list archive at Nabble.com.
> >
>

Reply via email to