Hi Yanbo,

Thank you for your reply..

Now I've made changes exactly , present in this link
http://wiki.apache.org/hadoop/AmazonS3

But my namenode is not coming up with exception (I tried it both locally
and inside EC2)

ERROR org.apache.hadoop.hdfs.server.namenode.NameNode:
java.net.BindException: Problem binding to <bucket-name>.
s3.amazonaws.com/207.171.163.14:8020 : Cannot assign requested address

fs.default.name = s3://<mybucket>    // Error : UnknownHost Exception in
Namenode!
or
fs.default.name = s3://<mybucket>.s3.amazonaws.com // Error BindException
in Namenode log!

Also I can't see any dir created inside my bucket. ( I'm able to execute
command $ bin/hadoop dfs -ls s3://<bucket>/ )

"bin/hadoop namenode -format " is saying succesfully formatted namenode dir
S3://bucket/hadoop/namenode , when it is not even existing there!


any suggestion?

Thanks again.

On Tue, Jul 24, 2012 at 4:11 PM, Yanbo Liang <yanboha...@gmail.com> wrote:

> I think you have made confusion about the integration of hadoop and S3.
> 1) If you set "dfs.data.dir=s3://******", it means that you set S3 as the
> DataNode local storage.
> You are still using HDFS as the under storage layer.
> As far as I know, it is not supported at present.
> 2)  The right way of using S3 integrated with Hadoop is to replace the
> HDFS with S3 just like you have tried.
> But I think you have missed some configuration parameters.
> "  fs.default.name=s3://<<mybucket> " is the most important parameter
> when you are using S3 to replace HDFS, but it is not enough. The detail
> configuration can be obtained from here
> http://wiki.apache.org/hadoop/AmazonS3
>
> Yanbo
>
>
> 2012/7/23 Alok Kumar <alok...@gmail.com>
>
>> Hello Group,
>>
>> I've hadoop setup locally running.
>>
>> Now I want to use Amazon s3://<mybucket> as my data store,
>> so i changed like " dfs.data.dir=s3://<mybucket>/hadoop/ " in my
>> hdfs-site.xml, Is it the correct way?
>> I'm getting error :
>>
>> WARN org.apache.hadoop.hdfs.server.datanode.DataNode: Invalid directory
>> in dfs.data.dir: can not create directory: s3://<mybucket>/hadoop
>> 2012-07-23 13:15:06,260 ERROR
>> org.apache.hadoop.hdfs.server.datanode.DataNode: All directories in
>> dfs.data.dir are invalid.
>>
>> and
>> when i changed like " dfs.data.dir=s3://<mybucket>/ "
>> I got error :
>>  ERROR org.apache.hadoop.hdfs.server.datanode.DataNode:
>> java.lang.IllegalArgumentException: Wrong FS: s3://<mybucket>/, expected:
>> file:///
>>     at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:381)
>>     at
>> org.apache.hadoop.fs.RawLocalFileSystem.pathToFile(RawLocalFileSystem.java:55)
>>     at
>> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:393)
>>     at
>> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:251)
>>     at
>> org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:146)
>>     at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:162)
>>     at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1574)
>>     at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1521)
>>     at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1539)
>>     at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:1665)
>>     at
>> org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:1682)
>>
>> Also,
>> When I'm changing fs.default.name=s3://<<mybucket> , Namenode is not
>> coming up with error : ERROR
>> org.apache.hadoop.hdfs.server.namenode.NameNode: java.net.BindException:
>> (Any way I want to run namenode locally, so I reverted it back to
>> hdfs://localhost:9000 )
>>
>> Your help is highly appreciated!
>> Thanks
>> --
>> Alok Kumar
>>
>
>


-- 
Alok Kumar

Reply via email to