Yes open a JIRA here with a patch: https://issues.apache.org/jira/browse/MAHOUT
If you're writing SequenceFiles, sure you can just write them straight to S3 by writing to a Path on s3:// -- is that something you've tried and doesn't work? should be that easy. On Thu, Jan 19, 2012 at 7:40 AM, Paul Rudin <p...@rudin.co.uk> wrote: > > I have a large lucene index from which I'm trying to extract term > vectors. I get a stackoverflow error, which is believe is caused by the > recursion in LuceneIterator.computeNext(). I could increase the stack > size, but with big enough data there could always be a problem. > > I have a modified version that uses a loop instead of the recursion > which seems to work OK. Should I put a patch somewhere? > > On a related note I'd quite like to be able to write the vectors > straight to s3 without writing to a local file first - is there a > practical way to do this? > >