[ https://issues.apache.org/jira/browse/MAHOUT-249?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12803603#action_12803603 ]
Sean Owen commented on MAHOUT-249: ---------------------------------- Ready to commit this too: is there an extra slash in the path you now construct? looks like there is a "//chunk" in there. Also does this still work with local files, does a "file://" URL work? > Make WikipediaXmlSplitter able to write the chunks directly to HDFS or S3 > ------------------------------------------------------------------------- > > Key: MAHOUT-249 > URL: https://issues.apache.org/jira/browse/MAHOUT-249 > Project: Mahout > Issue Type: Improvement > Components: Classification > Affects Versions: 0.2 > Reporter: Olivier Grisel > Priority: Minor > Fix For: 0.3 > > Attachments: MAHOUT-249-v2.patch, > MAHOUT-249-WikipediaXMLSplitterHDFS.patch > > > By using the Hadoop FS abstraction it should be possible to avoid writing the > chunks on the local hard drive before uploading them to HDFS or S3. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.