Hi Ankur,

There is a fetcher_tests.cpp in src/tests.

Tim

On Sat, Nov 1, 2014 at 7:27 PM, Ankur Chauhan <[email protected]> wrote:

> Hi Tim,
>
> I am trying to find/write some test cases. I couldn't find a
> fetcher_tests.{cpp|hpp} so once I have something, I'll post on review
> board. I am new to gmock/gtest so bear with me while i get up to speed.
>
> -- Ankur
>
> On 1 Nov 2014, at 19:23, Timothy Chen <[email protected]> wrote:
>
> Hi Ankur,
>
> Can you post on reviewboard? We can discuss more about the code there.
>
> Tim
>
> Sent from my iPhone
>
> On Nov 1, 2014, at 6:29 PM, Ankur Chauhan <[email protected]> wrote:
>
> Hi Tim,
>
> I don't think there is an issue which is directly in line with what i
> wanted but the closest one that I could find in JIRA is
> https://issues.apache.org/jira/browse/MESOS-1711
>
> I have a branch (
> https://github.com/ankurcha/mesos/compare/prefer_hadoop_fetcher ) that
> has a change that would enable users to specify whatever hdfs compatible
> uris to the mesos-fetcher but maybe you can weight in on it. Do you think
> this is the right track? if so, i would like to pick this issue and submit
> a patch for review.
>
> -- Ankur
>
>
> On 1 Nov 2014, at 04:32, Tom Arnfeld <[email protected]> wrote:
>
> Completely +1 to this. There are now quite a lot of hadoop compatible
> filesystem wrappers out in the wild and this would certainly be very useful.
>
> I'm happy to contribute a patch. Here's a few related issues that might be
> of interest;
>
> - https://issues.apache.org/jira/browse/MESOS-1887
> - https://issues.apache.org/jira/browse/MESOS-1316
> - https://issues.apache.org/jira/browse/MESOS-336
> - https://issues.apache.org/jira/browse/MESOS-1248
>
> On 31 October 2014 22:39, Tim Chen <[email protected]> wrote:
>
>> I believe there is already a JIRA ticket for this, if you search for
>> fetcher in Mesos JIRA I think you can find it.
>>
>> Tim
>>
>> On Fri, Oct 31, 2014 at 3:27 PM, Ankur Chauhan <[email protected]>
>> wrote:
>>
>>> Hi,
>>>
>>> I have been looking at some of the stuff around the fetcher and saw
>>> something interesting. The code for fetcher::fetch method is dependent on a
>>> hard coded list of url schemes. No doubt that this works but is very
>>> restrictive.
>>> Hadoop/HDFS in general is pretty flexible when it comes to being able to
>>> fetch stuff from urls and has the ability to fetch a large number of types
>>> of urls and can be extended by adding configuration into the
>>> conf/hdfs-site.xml and core-site.xml
>>>
>>> What I am proposing is that we refactor the fetcher.cpp to prefer to use
>>> the hdfs (using hdfs/hdfs.hpp) to do all the fetching if HADOOP_HOME is set
>>> and $HADOOP_HOME/bin/hadoop is available. This logic already exists and we
>>> can just use it. The fallback logic for using net::download or local file
>>> copy is may be left in place for installations that do not have hadoop
>>> configured. This means that if hadoop is present we can directly fetch urls
>>> such as tachyon://... snackfs:// ... cfs:// .... ftp://... s3://...
>>> http:// ... file:// with no extra effort. This makes up for a much
>>> better experience when it comes to debugging and extensibility.
>>>
>>> What do others think about this?
>>>
>>> - Ankur
>>
>>
>>
>
>
>

Reply via email to