Congratulations for the work Javi Roman !!
It seems to me a very interesting tool to accelerate and test the work of
the contributors. Personally, I would leave HDFS installed as a service to
make management and maintenance easier. For demonstrations and tests, I
think it would be interesting to put a spark-shell and update the README
with some map-reduce example. What do you think?

On 9 April 2018 at 19:34, Javi Roman <> wrote:

> Hi!
> I've just created a PR [1] with the fist version of DC/OS (tested
> v1.11.0) for Myriad development/deployment at DC/OS.
> The cluster created by default is something like that:
> $ vagrant status
> Current machine states:
> bt                         running (libvirt)
> m1                       running (libvirt)
> a1                        running (libvirt)
> a2                        running (libvirt)
> a3                        running (libvirt)
> a4                        running (libvirt)
> p1                        running (libvirt)
> The bootstrap machine (with DC/OS CLI installed). One master node, 4
> private agents, and one public.
> I haven't tested Myraid yet, because we have different options for it:
> - Deploy HDFS as a system service.
> - Deploy HDFS using Mesosphere Universe (we have to tune this one, the
> requirements don't fit in this small Vagrant cluster).
> - Deploy Myriad by means of Marathon and Docker.
> We have to test different options for this deployment. One interesting
> idea would be create a Myriad Universe with packages for HDFS tuned
> for this Vagrant, with upstream Apache Hadoop, and packages for
> Myriad. This could be a good point for demos.
> [1]
> --
> Javi Roman
> Twitter: @javiromanrh
> GitHub:
> Linkedin:
> Big Data Blog:

Reply via email to