>>How would a MapR NM container connect back to Yarn  without running
>>MFS inside the container?
>
> Currently NM doesn't run inside a (docker) container (both with and
> without Myriad).
>
> But the plan is to get there in the future to support multi tenancy.
> If you can, please share your story on why you'd want to run NM
> inside a container. It helps to work together and work on the right
> problems.

I would say that in our case, the desired outcome would be having the
ability to for example, run MapR-FS on baremetal, getting all non-OS
spindles of a host and using it as the backing of a highly available
file-store, without having to worry with yet another scale out system.

We think of it in terms of being able to run Nifi / ElasticSearch /
Kafka / <your_favourtite_apache_project> within a container, where
persistent files are backed by MapR-FS and resource allocation
coordinated not by specific platform but by application agnostic
framework.

It reflects our experience (as probably many others noticed before)
that a good amount of current scale out systems have very similar
requirements

http://doc.mapr.com/display/MapR/Planning+Cluster+Hardware
https://www.elastic.co/guide/en/elasticsearch/guide/current/hardware.html
http://docs.confluent.io/1.0/kafka/deployment.html
(the Kafka and ElasticSearch pages go to the length of having some
good chunks of identical text... )


This mean some of this kit is so similar to the technology next door
that you could simply reallocate them around, but this is messy and I
rather use ephemeral managed computing resources.



Why running inside of a container? Basic hygiene.

We have users that use Python UDFs in their Hive queries and as a
mater of principle I would love to be able to prevent them from
running code (albeit indirectly) on baremetal.


>> If that is not the case, this will have some interesting consequences
>> on MapR users adopting Myriad.
>
> Can you please elaborate more on this? Currently Myriad doesn't launch
> NMs inside docker containers. Myriad launches NMs via Mesos, but as
> physical processes.

I meant that running MapR-FS both on baremetal and within a container
is not only a technical waste but a financial waste as well.

If that (running MFS twice) ends up being the choice and MapR
continues to license as it does now, I would end up having to license
1 + N  MapR nodes (where N is the number of containers reaching the
CLDB).

However based on your email it seems to me this is not an issue yet,
as running NM within a container still unsupported and working just
through hacks.

> I think a cleaner solution would be to not enforce running MFS inside the
> containers. Trying to do that has other consequences, esp. around the need
> to pass the host's disks into docker containers for MFS to format.
>
> The right way to do this is to modify createTTVolume.sh to be container
> aware.

100% agree.


Thank you for the replies, it seems like it will be: Mesos here I go. :-)

Cheers

Reply via email to