by editing the hadoop-site.xml, you set the default. but I don't
recommend changing the default on EC2.
but you can specify the filesystem to use through the URL that
references your data (jobConf.addInputPath etc) for a particular job.
in the case of the S3 block filesystem, just use a
That's a good point, in fact it didn't occured me that i could access it
like that.
But some questions came to my mind:
How do i put something into the fs?
something like bin/hadoop fs -put input input will not work well since s3
is not the default fs, so i tried to do bin/hadoop fs -put input
How do i put something into the fs?
something like bin/hadoop fs -put input input will not work well
since s3
is not the default fs, so i tried to do bin/hadoop fs -put input
s3://ID:[EMAIL PROTECTED]/input (and some variations of it) but didn't
worked, i
always got an error complaining
Hello,
I've been trying to setup hadoop to use s3 as filesystem, i read in the wiki
that it's possible to choose either S3 native FileSystem or S3 Block
Filesystem. I would like to use S3 Block FileSystem to avoid the task of
manually transferring data from S3 to HDFS every time i want to run a