Yes open a JIRA here with a patch: https://issues.apache.org/jira/browse/MAHOUT

If you're writing SequenceFiles, sure you can just write them straight
to S3 by writing to a Path on s3:// -- is that something you've tried
and doesn't work? should be that easy.

On Thu, Jan 19, 2012 at 7:40 AM, Paul Rudin <p...@rudin.co.uk> wrote:
>
> I have a large lucene index from which I'm trying to extract term
> vectors. I get a stackoverflow error, which is believe is caused by the
> recursion in LuceneIterator.computeNext().  I could increase the stack
> size, but with big enough data there could always be a problem.
>
> I have a modified version that uses a loop instead of the recursion
> which seems to work OK. Should I put a patch somewhere?
>
> On a related note I'd quite like to be able to write the vectors
> straight to s3 without writing to a local file first - is there a
> practical way to do this?
>
>

Reply via email to