[
https://issues.apache.org/jira/browse/HADOOP-5010?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12663388#action_12663388
]
Marco Nicosia commented on HADOOP-5010:
---------------------------------------
bq. So maybe all we need is better documentation of what's passed over HTTP?
If there's a guarantee that subsequent connections need not always connect to
the same server (ie, any session over the protocol is managed either via a
single continuous HTTP/1.1 connection, cookies, or some other session
management), then yes, more documentation on how the HTTP protocol is used will
allow "creative" admins to use existing HTTP infrastructure in their Hadoop
deployments.
bq. I think we cannot simply use standard HTTP because it does not support file
system access.
If the limitation is that HTTP doesn't specify how to get/put structured data
(such as a directory listing), why not use some well accepted standard, such as
REST?
The reason I'm pushing for this is that the closer Hadoop comes to presenting
some standards-compliant interface, the easier it becomes for users to
integrate Hadoop into existing infrastructure(s). Currently, one of the least
obvious points of integration is how to get data both onto, and back off of, an
HDFS.
> Replace HFTP/HSFTP with plain HTTP/HTTPS
> ----------------------------------------
>
> Key: HADOOP-5010
> URL: https://issues.apache.org/jira/browse/HADOOP-5010
> Project: Hadoop Core
> Issue Type: Improvement
> Components: contrib/hdfsproxy
> Affects Versions: 0.18.0
> Reporter: Marco Nicosia
>
> In HADOOP-1563, [~cutting] wrote:
> bq. The URI for this should be something like hftp://host:port/a/b/c, since,
> while HTTP will be used as the transport, this will not be a FileSystem for
> arbitrary HTTP urls.
> Recently, we've been talking about implementing an HDFS proxy (HADOOP-4575)
> which would be a secure way to make HFTP/HSFTP available. In so doing, we may
> even remove HFTP/HSFTP from being offered on the HDFS itself (that's another
> discussion).
> In the case of the HDFS proxy, does it make sense to do away with the
> artificial HFTP/HSFTP protocols, and instead simply offer standard HTTP and
> HTTPS? That would allow non-HDFS-specific clients, as well as using various
> standard HTTP infrastructure, such as load balancers, etc.
> NB, to the best of my knowledge, HFTP is only documented on the
> [distcp|http://hadoop.apache.org/core/docs/current/distcp.html] page, and
> HSFTP is not documented at all?
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.