Disclousre: I work at Weaveworks
Ajay - IP assignments are managed dynamically by Weave Net. There is NO manual intervention needed. Here are the docs - http://docs.weave.works/weave/latest_release/features.html#addressing Please: If you need help with Weave, get in touch. All, I'd like to highlight an important feature of Weave Net that I haven't seen in competing products. Weave does not require you to set and up and manage a distributed key-value store, such as etcd, in order to maintain network state. This makes Weave a lot easier to set up. Also, it means that with Weave you don't lose *any* availability during a network partition. Finally, you don't suffer the performance hit of running a consensus algorithm when you update your network. A reasonably good intro to this is here - http://blog.weave.works/2015/11/03/docker-networking-1-9-technical-deep-dive/ Please do feel free to direct questions about Weave to our team and community! See here: http://weave.works/help/index.html#mailing-list alexis On Thu, Nov 26, 2015 at 2:31 PM, Ajay Bhatnagar <[email protected]> wrote: > With Calico you only create the virtual subnets and ip assignments are > managed dynamically by Calico w/o any manual intervention needed. > > Cheers > > Ajay > > > > From: Paul Bell [mailto:[email protected]] > Sent: Thursday, November 26, 2015 9:05 AM > To: [email protected] > Subject: Re: Anyone try Weave in Mesos env ? > > > > Hi Weitao, > > > > I came up with this architecture as a way of distributing our application > across multiple nodes. Pre-Mesos, our application, delivered as a single > VMware VM, was not easily scalable. By breaking out the several application > components as Docker containers, we are now able (within limits imposed > chiefly by the application itself) to distribute & run those containers > across the several nodes in the Mesos cluster. Application containers that > need to talk to each other are connected via Weave's "overlay" (veth) > network. > > > > Not surprisingly, this architecture has some of the benefits that you'd > expect from Mesos, chief among them being high-availability (more on this > below), scalability, and hybrid Cloud deployment. > > > > The core unit of deployment is an Ubuntu image (14.04 LTS) that I've > configured with the appropriate components: > > > > Zookeeper > > Mesos-master > > Mesos-slave > > Marathon > > Docker > > Weave > > SSH (including RSA keys) > > Our application > > > > This images is presently downloaded by a customer as a VMware .ova file. We > typically ask the customer to convert the resulting VM to a so-called VMware > template from which she can easily deploy multiple VMs as needed. Please > note that although we've started with VMware as our virtualization platform, > I've successfully run cluster nodes on both EC2 and Azure. > > > > I tend to describe the Ubuntu image as "polymorphic", i.e., it can be told > to assume one of two roles, either a "master" role or a "slave" role. A > master runs ZK, mesos-master, and Marathon. A slave runs mesos-slave, > Docker, Weave, and the application. > > > > We presently offer 3 canned deployment options: > > single-host, no HA > multi-host, no HA (1 master, 3 slaves) > multi-host, HA (3 masters, 3 slaves) > > The single-host, no HA option exists chiefly to mimic the original pre-Mesos > deployment. But it has the added virtue, thanks to Mesos, of allowing us to > dynamically "grow" from a single-host to multiple hosts. > > > > The multi-host, no HA option is presently geared toward a sharded MongoDB > backend where each slave runs a mongod container that is a single partition > (shard) of the larger database. This deployment option also lends itself > very nicely to adding a new slave node at the cluster level, and a new > mongod container at the application level - all without any downtime > whatsoever. > > > > The multi-host, HA option offers the probably familiar cluster-level high > availability. I stress "cluster-level" because I think we have to > distinguish between HA at that level & HA at the application level. The > former is realized by the 3 master hosts, i.e., you can lose a master and > new one will self-elect thereby keeping the cluster up & running. But, to my > mind, at least, application level HA requires some co-operation on the part > of the application itself (e.g., checkpoint/restart). That said, it is > almost magical to watch Mesos re-launch an application container that has > crashed. But whether or not that re-launch results in coherent application > behavior is another matter. > > > > An important home-grown component here is a Java program that automates > these functions: > > > > create cluster - configures a host for a given role and starts Mesos > services. This is done via SSH > > start application - distributes application containers across slave hosts. > This is done by talking to the Marathon REST API > > stop application - again, via the Marathon REST API > > stop cluster - stops Mesos services. Again, via SSH > > destroy cluster - deconfigures the host (after which it has no defined > role); again, SSH > > > > As I write, I see Ajay's e-mail arrive about Calico. I am aware of this > project and it seems quite solid. But I've never understood the need to > "worry about networking containers in multihost setup". Weave runs as a > Docker container and It Just Works. I've "woven" together slaves nodes in a > cluster that spanned 3 different datacenters, one of them in EC2, without > any difficulty. Yes, I do have to assign Weave IP addresses to the several > containers, but this is hardly onerous. In fact, I've found it "liberating" > to select such addresses from a CIDR/8 address space, assigning them to > containers based on the container's purpose (e.g., MongoDB shard containers > might live at 10.4.0.X, etc.). Ultimately, this assignment boils down to > setting an environment variable that Marathon (or the mesos-slave executor) > will use when creating the container via "docker run". > > > > There is a whole lot more that I could say about the internals of this > architecture. But, if you're still interested, I'll await further questions > from you. > > > > HTH. > > > > Cordially, > > > > Paul > > > > > > On Thu, Nov 26, 2015 at 7:16 AM, Paul <[email protected]> wrote: > > Gladly, Weitao. It'd be my pleasure. > > But give me a few hours to find some free time. > > I am today tasked with cooking a Thanksgiving turkey. > > But I will try to find the time before noon today (I'm on the right coast in > the USA). > > -Paul > > >> On Nov 25, 2015, at 11:26 PM, Weitao <[email protected]> wrote: >> >> Hi, Paul. Can your share the total experience about the arch with us. I am >> trying to do the similar thing >> >> >>> 在 2015年11月26日,09:47,Paul <[email protected]> 写道: >>> >>> experience > >

