Hi Ankur,

Can you post on reviewboard? We can discuss more about the code there.

Tim

Sent from my iPhone

> On Nov 1, 2014, at 6:29 PM, Ankur Chauhan <[email protected]> wrote:
> 
> Hi Tim,
> 
> I don't think there is an issue which is directly in line with what i wanted 
> but the closest one that I could find in JIRA is 
> https://issues.apache.org/jira/browse/MESOS-1711
> 
> I have a branch ( 
> https://github.com/ankurcha/mesos/compare/prefer_hadoop_fetcher ) that has a 
> change that would enable users to specify whatever hdfs compatible uris to 
> the mesos-fetcher but maybe you can weight in on it. Do you think this is the 
> right track? if so, i would like to pick this issue and submit a patch for 
> review.
> 
> -- Ankur
> 
> 
>> On 1 Nov 2014, at 04:32, Tom Arnfeld <[email protected]> wrote:
>> 
>> Completely +1 to this. There are now quite a lot of hadoop compatible 
>> filesystem wrappers out in the wild and this would certainly be very useful.
>> 
>> I'm happy to contribute a patch. Here's a few related issues that might be 
>> of interest;
>> 
>> - https://issues.apache.org/jira/browse/MESOS-1887
>> - https://issues.apache.org/jira/browse/MESOS-1316
>> - https://issues.apache.org/jira/browse/MESOS-336
>> - https://issues.apache.org/jira/browse/MESOS-1248
>> 
>>> On 31 October 2014 22:39, Tim Chen <[email protected]> wrote:
>>> I believe there is already a JIRA ticket for this, if you search for 
>>> fetcher in Mesos JIRA I think you can find it.
>>> 
>>> Tim
>>> 
>>>> On Fri, Oct 31, 2014 at 3:27 PM, Ankur Chauhan <[email protected]> wrote:
>>>> Hi,
>>>> 
>>>> I have been looking at some of the stuff around the fetcher and saw 
>>>> something interesting. The code for fetcher::fetch method is dependent on 
>>>> a hard coded list of url schemes. No doubt that this works but is very 
>>>> restrictive.
>>>> Hadoop/HDFS in general is pretty flexible when it comes to being able to 
>>>> fetch stuff from urls and has the ability to fetch a large number of types 
>>>> of urls and can be extended by adding configuration into the 
>>>> conf/hdfs-site.xml and core-site.xml
>>>> 
>>>> What I am proposing is that we refactor the fetcher.cpp to prefer to use 
>>>> the hdfs (using hdfs/hdfs.hpp) to do all the fetching if HADOOP_HOME is 
>>>> set and $HADOOP_HOME/bin/hadoop is available. This logic already exists 
>>>> and we can just use it. The fallback logic for using net::download or 
>>>> local file copy is may be left in place for installations that do not have 
>>>> hadoop configured. This means that if hadoop is present we can directly 
>>>> fetch urls such as tachyon://... snackfs:// ... cfs:// .... ftp://... 
>>>> s3://... http:// ... file:// with no extra effort. This makes up for a 
>>>> much better experience when it comes to debugging and extensibility.
>>>> 
>>>> What do others think about this?
>>>> 
>>>> - Ankur
> 

Reply via email to