Hi Darius,

On Sep 18, 2013, at 1:10am, Gokhan Capan wrote:

> It seems you hit a "Hadoop on Windows" issue, it might have something to do
> with how Hadoop sets file permissions.

From my experience, only the (old) 0.20.2 version of Hadoop works well with 
Cygwin, otherwise you run into file permissions issues like the one you 
mentioned.

If you want to give that version a try, and can't find a download, see 
http://scaleunlimited.com/downloads/3nn2pq/hadoop-0.20.2.tgz

-- Ken


> On Tue, Sep 17, 2013 at 3:02 PM, Darius Miliauskas <
> [email protected]> wrote:
> 
>> That's like a charm, Gokhan, your suggestion was on point again. However...
>> Despite the fact that the build is successful, the file is still empty,
>> and I got the exception as always on Windows:
>> 
>> java.io.IOException: Failed to set permissions of path:
>> \tmp\hadoop-DARIUS\mapred\staging\DARIUS331150778\.staging to 0777
>> at org.apache.hadoop.fs.FileUtil.checkReturnValue(FileUtil.java:689)
>> at org.apache.hadoop.fs.FileUtil.setPermission(FileUtil.java:670)
>> at
>> 
>> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:514)
>> at
>> org.apache.hadoop.fs.RawLocalFileSystem.mkdirs(RawLocalFileSystem.java:349)
>> at org.apache.hadoop.fs.FilterFileSystem.mkdirs(FilterFileSystem.java:189)
>> at
>> 
>> org.apache.hadoop.mapreduce.JobSubmissionFiles.getStagingDir(JobSubmissionFiles.java:116)
>> at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:918)
>> at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:912)
>> at java.security.AccessController.doPrivileged(Native Method)
>> at javax.security.auth.Subject.doAs(Subject.java:415)
>> at
>> 
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
>> at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:912)
>> at org.apache.hadoop.mapreduce.Job.submit(Job.java:500)
>> at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:530)
>> at
>> 
>> org.apache.mahout.vectorizer.DocumentProcessor.tokenizeDocuments(DocumentProcessor.java:93)
>> at
>> 
>> org.apache.mahout.vectorizer.SparseVectorsFromSequenceFiles.run(SparseVectorsFromSequenceFiles.java:257)
>> at
>> 
>> org.apache.mahout.mahoutnewsrecommender2.Recommender.myRecommender(Recommender.java:99)
>> at org.apache.mahout.mahoutnewsrecommender2.App.main(App.java:26)
>> 
>> BUILD SUCCESSFUL (total time: 3 seconds)
>> 
>> 
>> Thanks,
>> 
>> Darius
>> 
>> 
>> 
>> 
>> 2013/9/12 Gokhan Capan <[email protected]>
>> 
>>> Although Windows is not officially supported, your
>>> svsf.run(new String[]{inputPath.toString(), outputPath.toString()})
>>> should be
>>> svsf.run(new String[]{"-i",inputPath.toString(), "-o",
>>> outputPath.toString()}) anyway.
>>> 
>>> Best
>>> 
>>> 
>>> Gokhan
>>> 
>>> 
>>> On Thu, Sep 12, 2013 at 4:14 PM, Darius Miliauskas <
>>> [email protected]> wrote:
>>> 
>>>> Dear All,
>>>> 
>>>> I am trying to use SparseVectorsFromSequenceFiles () through Java code
>>>> (NetBeans 7&Windows 7) . here is my code (API):
>>>> 
>>>> //inputPath is the path of my SequenceFile
>>>> Path inputPath = new Path(""C:\\Users\\DARIUS\\forTest1.txt");
>>>> 
>>>> //outputPath where I expect some results
>>>> Path outputPath = new Path("C:\\Users\\DARIUS\\forTest2.txt");
>>>> 
>>>> SparseVectorsFromSequenceFiles svfsf = new
>> SparseVectorsFromSequenceFiles
>>>> ();
>>>> svfsf.run(new String []{inputPath.toString(), outputPath.toString()
>>>> });
>>>> 
>>>> Build is successful. However, at the end I got just the empty file what
>>> was
>>>> expected to be my output. Do you have any idea why the output file is
>>>> empty, and what I should change in the code to get the results?
>>>> 
>>>> 
>>>> Ciao,
>>>> 
>>>> Darius
>>>> 
>>> 
>> 

--------------------------
Ken Krugler
+1 530-210-6378
http://www.scaleunlimited.com
custom big data solutions & training
Hadoop, Cascading, Cassandra & Solr





Reply via email to