[ 
https://issues.apache.org/jira/browse/HADOOP-5010?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12663388#action_12663388
 ] 

Marco Nicosia commented on HADOOP-5010:
---------------------------------------

bq. So maybe all we need is better documentation of what's passed over HTTP?

If there's a guarantee that subsequent connections need not always connect to 
the same server (ie, any session over the protocol is managed either via a 
single continuous HTTP/1.1 connection, cookies, or some other session 
management), then yes, more documentation on how the HTTP protocol is used will 
allow "creative" admins to use existing HTTP infrastructure in their Hadoop 
deployments.

bq. I think we cannot simply use standard HTTP because it does not support file 
system access.

If the limitation is that HTTP doesn't specify how to get/put structured data 
(such as a directory listing), why not use some well accepted standard, such as 
REST?

The reason I'm pushing for this is that the closer Hadoop comes to presenting 
some standards-compliant interface, the easier it becomes for users to 
integrate Hadoop into existing infrastructure(s). Currently, one of the least 
obvious points of integration is how to get data both onto, and back off of, an 
HDFS.


> Replace HFTP/HSFTP with plain HTTP/HTTPS
> ----------------------------------------
>
>                 Key: HADOOP-5010
>                 URL: https://issues.apache.org/jira/browse/HADOOP-5010
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: contrib/hdfsproxy
>    Affects Versions: 0.18.0
>            Reporter: Marco Nicosia
>
> In HADOOP-1563, [~cutting] wrote:
> bq. The URI for this should be something like hftp://host:port/a/b/c, since, 
> while HTTP will be used as the transport, this will not be a FileSystem for 
> arbitrary HTTP urls.
> Recently, we've been talking about implementing an HDFS proxy (HADOOP-4575) 
> which would be a secure way to make HFTP/HSFTP available. In so doing, we may 
> even remove HFTP/HSFTP from being offered on the HDFS itself (that's another 
> discussion).
> In the case of the HDFS proxy, does it make sense to do away with the 
> artificial HFTP/HSFTP protocols, and instead simply offer standard HTTP and 
> HTTPS? That would allow non-HDFS-specific clients, as well as using various 
> standard HTTP infrastructure, such as load balancers, etc.
> NB, to the best of my knowledge, HFTP is only documented on the 
> [distcp|http://hadoop.apache.org/core/docs/current/distcp.html] page, and 
> HSFTP is not documented at all?

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to