On 20/06/17 11:45, Jay Pipes wrote:
Good discussion, Zane. Comments inline.

++

On 06/20/2017 11:01 AM, Zane Bitter wrote:
On 20/06/17 10:08, Jay Pipes wrote:
On 06/20/2017 09:42 AM, Doug Hellmann wrote:
Does "service VM" need to be a first-class thing?  Akanda creates
them, using a service user. The VMs are tied to a "router" which
is the billable resource that the user understands and interacts with
through the API.

Frankly, I believe all of these types of services should be built as applications that run on OpenStack (or other) infrastructure. In other words, they should not be part of the infrastructure itself.

There's really no need for a user of a DBaaS to have access to the host or hosts the DB is running on. If the user really wanted that, they would just spin up a VM/baremetal server and install the thing themselves.

Hey Jay,
I'd be interested in exploring this idea with you, because I think everyone agrees that this would be a good goal, but at least in my mind it's not obvious what the technical solution should be. (Actually, I've read your email a bunch of times now, and I go back and forth on which one you're actually advocating for.) The two options, as I see it, are as follows:

1) The database VMs are created in the user's tena^W project. They connect directly to the tenant's networks, are governed by the user's quota, and are billed to the project as Nova VMs (on top of whatever additional billing might come along with the management services). A [future] feature in Nova (https://review.openstack.org/#/c/438134/) allows the Trove service to lock down access so that the user cannot actually interact with the server using Nova, but must go through the Trove API. On a cloud that doesn't include Trove, a user could run Trove as an application themselves and all it would have to do differently is not pass the service token to lock down the VM.

alternatively:

2) The database VMs are created in a project belonging to the operator of the service. They're connected to the user's network through <magic>, and isolated from other users' databases running in the same project through <security groups? hierarchical projects? magic?>. Trove has its own quota management and billing. The user cannot interact with the server using Nova since it is owned by a different project. On a cloud that doesn't include Trove, a user could run Trove as an application themselves, by giving it credentials for their own project and disabling all of the cross-tenant networking stuff.

None of the above :)

Don't think about VMs at all. Or networking plumbing. Or volume storage or any of that.

OK, but somebody has to ;)

Think only in terms of what a user of a DBaaS really wants. At the end of the day, all they want is an address in the cloud where they can point their application to write and read data from.

Do they want that data connection to be fast and reliable? Of course, but how that happens is irrelevant to them

Do they want that data to be safe and backed up? Of course, but how that happens is irrelevant to them.

Fair enough. The world has changed a lot since RDS (which was the model for Trove) was designed, it's certainly worth reviewing the base assumptions before embarking on a new design.

The problem with many of these high-level *aaS projects is that they consider their user to be a typical tenant of general cloud infrastructure -- focused on launching VMs and creating volumes and networks etc. And the discussions around the implementation of these projects always comes back to minutia about how to set up secure communication channels between a control plane message bus and the service VMs.

Incidentally, the reason that discussions always come back to that is because OpenStack isn't very good at it, which is a huge problem not only for the *aaS projects but for user applications in general running on OpenStack.

If we had fine-grained authorisation and ubiquitous multi-tenant asynchronous messaging in OpenStack then I firmly believe that we, and application developers, would be in much better shape.

If you create these projects as applications that run on cloud infrastructure (OpenStack, k8s or otherwise),

I'm convinced there's an interesting idea here, but the terminology you're using doesn't really capture it. When you say 'as applications that run on cloud infrastructure', it sounds like you mean they should run in a Nova VM, or in a Kubernetes cluster somewhere, rather than on the OpenStack control plane. I don't think that's what you mean though, because you can (and IIUC Rackspace does) deploy OpenStack services that way already, and it has no real effect on the architecture of those services.

then the discussions focus instead on how the real end-users -- the ones that actually call the APIs and utilize the service -- would interact with the APIs and not the underlying infrastructure itself.

Here's an example to think about...

What if a provider of this DBaaS service wanted to jam 100 database instances on a single VM and provide connectivity to those database instances to 100 different tenants?

Would those tenants know if those databases were all serviced from a single database server process running on the VM?

You bet they would when one (or all) of the other 99 decided to run a really expensive query at an inopportune moment :)

Or 100 contains each running a separate database server process? Or 10 containers running 10 database server processes each?

No, of course not. And the tenant wouldn't care at all, because the

Well, if they had any kind of regulatory (or even performance) requirements then the tenant might care really quite a lot. But I take your point that many might not and it would be good to be able to offer them lower cost options.

point of the DBaaS service is to get a database. It isn't to get one or more VMs/containers/baremetal servers.

I'm not sure I entirely agree here. There are two kinds of DBaaS. One is a data API: a multitenant database a la DynamoDB. Those are very cool, and I'm excited about the potential to reduce the granularity of billing to a minimum, in much the same way Swift does for storage, and I'm sad that OpenStack's attempt in this space (MagnetoDB) didn't work out. But Trove is not that.

People use Trove because they want to use a *particular* database, but still have all the upgrades, backups, &c. handled for them. Given that the choice of database is explicitly *not* abstracted away from them, things like how many different VMs/containers/baremetal servers the database is running on are very much relevant IMHO, because what you want depends on both the database and how you're trying to use it. And because (afaik) none of them have native multitenancy, it's necessary that no tenant should have to share with any other.

Essentially Trove operates at a moderate level of abstraction - somewhere between managing the database + the infrastructure it runs on yourself and just an API endpoint you poke data into. It also operates at the coarse end of a granularity spectrum running from VMs->Containers->pay as you go.

It's reasonable to want to move closer to the middle of the granularity spectrum. But you can't go all the way to the high abstraction/fine grained ends of the spectra (which turn out to be equivalent) without becoming something qualitatively different.

At the end of the day, I think Trove is best implemented as a hosted application that exposes an API to its users that is entirely separate from the underlying infrastructure APIs like Cinder/Nova/Neutron.

This is similar to Kevin's k8s Operator idea, which I support but in a generic fashion that isn't specific to k8s.

In the same way that k8s abstracts the underlying infrastructure (via its "cloud provider" concept), I think that Trove and similar projects need to use a similar abstraction and focus on providing a different API to their users that doesn't leak the underlying infrastructure API concepts out.

OK, so trying to summarise (stop me if I'm getting it wrong): essentially you support option (2) because it is a closed abstraction. Trove has its own quota management, billing, &c. and the user can't see the VM, so the operator is free to substitute a different backend that allocates compute capacity in finer-grained increments than Nova does.

Interestingly, that's only an issue because there is no finer-grained compute resource than a VM available through the OpenStack API. If there were an OpenStack API (or even just a Keystone-authenticated API) to a shared, multitenant container orchestration cluster, this wouldn't be an issue. But apart from OpenShift, I can't think of any cloud service that's doing that - AWS, Google, OpenStack are all using the model where the COE cluster is deployed on VMs that are owned by a particular tenant. Of all the things you could run in containers on shared servers, databases have arguably the most to lose (performance, security) and the least to gain (since they're by definition stateful). So my question is: if this is such a good idea for databases, why isn't anybody doing it for everything container-based? i.e. instead of Magnum/Zun should we just be working on a Keystone auth gateway for OpenShift (a.k.a. the _one_ thing that _everyone_ had hitherto agreed was definitely out of scope :D )?

Until then it seems to me that the tradeoff is between decoupling it from the particular cloud it's running on so that users can optionally deploy it standalone (essentially Vish's proposed solution for the *aaS services from many moons ago) vs. decoupling it from OpenStack in general so that the operator has more flexibility in how to deploy.

I'd love to be able to cover both - from a user using it standalone to spin up and manage a DB in containers on a shared PaaS, through to a user accessing it as a service to provide a DB running on a dedicated VM or bare metal server, and everything in between. I don't know is such a thing is feasible. I suspect we're going to have to talk a lot about VMs and network plumbing and volume storage :)

cheers,
Zane.

Best,
-jay

Of course the current situation, as Amrith alluded to, where the default is option (1) except without the lock-down feature in Nova, though some operators are deploying option (2) but it's not tested upstream... clearly that's the worst of all possible worlds, and AIUI nobody disagrees with that.

To my mind, (1) sounds more like "applications that run on OpenStack (or other) infrastructure", since it doesn't require stuff like the admin-only cross-project networking that makes it effectively "part of the infrastructure itself" - as evidenced by the fact that unprivileged users can run it standalone with little more than a simple auth middleware change. But I suspect you are going to use similar logic to argue for (2)? I'd be interested to hear your thoughts.

cheers,
Zane.

__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: [email protected]?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: [email protected]?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


__________________________________________________________________________
OpenStack Development Mailing List (not for usage questions)
Unsubscribe: [email protected]?subject:unsubscribe
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to