Thanks for your idea, Ankur.

Unfortunately, I have many other folders under the root folder. Anyhow,
it's a very strange behavior.

I can also load the folders separately and join then with UNION. I can use
one of these alternatives for now. But I'd like to know whether it's a new
issue, a well-known bug of if I'm doing someting wrong.

Thanks,
Rodrigo.

2014-12-24 15:04 GMT-02:00 Ankur <ankur.kasliwal...@gmail.com>:

> Hi,
>
> Try giving path till root folder that is folder containing the mentioned
> four folders. ( bucket in your case ).
>
> This is a temporary solution to your problem.
>
> Thanks,
> Ankur
>
> Sent from my iPhone
>
> > On Dec 24, 2014, at 10:22 PM, Rodrigo Ferreira <web...@gmail.com> wrote:
> >
> > Hi everyone, happy holidays!
> >
> > I have a Pig script that reads from 4 different folders in Amazon S3.
> This
> > is the code:
> >
> > load_1 = LOAD 's3n://mybucket/{folder_1,folder_2,folder_3,folder_4}'
> > USING...;
> >
> > It happens that instead of reading each folder just once and appending
> the
> > files Pig/Hadoop reads each folder 4 times.
> >
> > The input should have 62174 records, but in the end I get 248696.
> >
> > Why is that? Any ideas?
> >
> > Thanks,
> > Rodrigo.
>

Reply via email to