Re: Using S3 Block FileSystem as HDFS replacement

2008-07-01 Thread Chris K Wensel
by editing the hadoop-site.xml, you set the default. but I don't recommend changing the default on EC2. but you can specify the filesystem to use through the URL that references your data (jobConf.addInputPath etc) for a particular job. in the case of the S3 block filesystem, just use a

Re: Using S3 Block FileSystem as HDFS replacement

2008-07-01 Thread slitz
That's a good point, in fact it didn't occured me that i could access it like that. But some questions came to my mind: How do i put something into the fs? something like bin/hadoop fs -put input input will not work well since s3 is not the default fs, so i tried to do bin/hadoop fs -put input

Re: Using S3 Block FileSystem as HDFS replacement

2008-07-01 Thread Chris K Wensel
How do i put something into the fs? something like bin/hadoop fs -put input input will not work well since s3 is not the default fs, so i tried to do bin/hadoop fs -put input s3://ID:[EMAIL PROTECTED]/input (and some variations of it) but didn't worked, i always got an error complaining

Using S3 Block FileSystem as HDFS replacement

2008-06-30 Thread slitz
Hello, I've been trying to setup hadoop to use s3 as filesystem, i read in the wiki that it's possible to choose either S3 native FileSystem or S3 Block Filesystem. I would like to use S3 Block FileSystem to avoid the task of manually transferring data from S3 to HDFS every time i want to run a