On 05/18/2012 01:07 PM, Itamar Heim wrote:
On 05/15/2012 05:34 PM, Gary Kotton wrote:
...

2. host management --> interface
2.1 you are suggesting to remove the vlan field with a fabric field?
are we sure this is something which shouldn't be presented in the main
view and only via extended properties?

> Yes, this is what I am suggesting. VLAN tagging is one method of network
> isolation. There are more, for example GRE tunneling.

that doesn't mean this is not an important enough piece of information to have in the subtab, rather than in an additional dialog (the fabric type could be represented in an icon for example)
miki/andrew/einav - thoughts on the UI aspect of this?
This is a very important piece of information. The network specific are very important. I think that they can be displayed on a different screen and do not need to clutter this specific view. Please not that currently Quantum does not return the network specifics, for example the VLAN tag. This is something that is being addressed by a blueprint for connecting to existing networks.


2.2 is the fabric really a host interface level property, or a cluster
wide property (would same logical network be implemented on different
hosts in the cluster via both vdsm and quantum)?
would live migration work in same cluster if one host has qunatum
fabric and another vdsm fabric for same logical network?
These are all up for discussion. I just wanted to provide something that
we can start to work with. It would be nice if there was some input from
product management on this issue (not exactly sure how this works in an
open source community :))

This goes to assumptions on qunatum integration and capabilities.
I'm missing this part in the beginning of the feature page.
I can add a link to the Quantum wiki or documentation. I am not sure that this will address this issue. Quantum is evolving. It started with simple layer 2 connectivity. Now layer 3 address management is being added. Next stop will be security policies and quality of service.
if quantum can support multiple types, and live migration between, then it can a host level config. if live migration wouldn't work, it would have to be a cluster level config. if it is limited to a single technology, it is kind of a system level config.

(and actually, after reading your reply to 3.7.1, i understand the plugin is an interface level config, since you can have more than a single one on the same host).

Quantum is a network service. It can work in all of the above scenarios - for example, it works in OpenStack which supports the above. It is just a matter of configuring it correctly.
same goes to implications to the logical network entity (see my comment in 2.4 below).

I'd also like to understand in that regard how do we match up host reporting it can support (i.e., installed and configured) to a specific technology. i.e., there needs to be some correlation between which logical networks can be deployed on which plugin. so also need to understand how this correlation is done automatically between logical network, to which fabric/interface it belongs to.
This is a gap in Quantum at the moment, that is, Quantum does not return logical network statistics. (https://blueprints.launchpad.net/quantum/+spec/network-statistics)

(I admit your reply to 3.7.1 confused me as to how quantum will decide on which plugin to create a specific attachment)?



2.3 you are describing the subtab of an interface, and mention a
qunatum pull down menu. is this in the subtab? in a dialog?
UI mockups are needed here.
I think that prior to getting mockups, we should ensure that we have the
idea crystallized.

I can live with that, goes back to what i wrote in 2.2.


2.4 also, is this pulldown correct at host level or cluster level
(would live migration work between hosts implementing different
plugins for same logical network in same cluster?)

as i stated above - this has to be clarified in assumptions.
doesn't have to be the current state, rather a clear understanding where it is going to, or as to what makes sense. (I am not sure requiring live migration between multiple technologies makes sense, since implementation of the logical network could be based on different concepts (vlan vs. gre). however, this means we need to define if the logical network (which is not limited to a cluster) should be limited to clusters supporting these technologies. i.e., a vlan based logical network can cross (in different clusters) UCS, openvswitch and bridges. but a gre based logical network cannot cross all of them, etc.
For a start we can limit live migration only for homogeneous hosts, that is the same Quantum agent. If the VM has characteristics that are supported by more than one plugin then I do not see any reason why this should be limited. For example, every plugin supports VLAN tagging. If a VM runs on a tagged network then there is no reason why the VM cannot be moved from a host running OVS to a Host that has UCS


on to backend:
3.1 "The Quantum Service is a process that runs the Quantum API web
server (port 9696)"

how is authentication between engine and quantum service is done (we
have a client certificate for engine which qunatum can verify i guess)
In my opinion this should be running locally on the same host as the
engine.

even on same server, we have authentication between components running in different processes.
Can you please point me to an example so that I can understand in more detail. The interface with the Quantum server is via REST API. If necesasry this can be done over SSL.
on top of that, one of our main goals is to allow multiple engines running side by side for scale. maybe quantum will have to be in active/passive mode for a while, but multiple nodes should be able to access it.
Do the engines have a common database or does each engine have their own database? If the service si the problem then a load balancer can be used to select the "best" server. The issue is the information stored in the database.

there is no need for fancy authentication/authorization. simply trusting ovirt engine CA certificate and validating engine client certificate should suffice (and i assume can be done via config in the http layer of quantum).

also, this can be a phase 2, but let's have a list of what is a must for merging the code (this isn't), to deployment (phase 2), to more than that (phase 3). under deployment i'd expect not requiring a different db technology, integrating with installer, an upgrade path, cleaner and secure communication channels, etc.
Some Quantum plugins have a database, others do not. The open source plugins are implemented in python whcih uses the SQL Alchemy database interface. The user can select the database type. I have current used this with the default mysql. If we use the postgres database and it is on the same database with the same authentication as the oVirt engine does this cover the authentication issues?


3.2 is there a config of the quantum service uri? should it be a
single instance or a per DC property?
(it sounds like a per DC property, can be a future thing)
I am sorry but I do not understand the question.

even if quantum is running on same host, you will have a configuration item of the URI to the service, and will configure this as part of installation flow. if this is not configured, I'd expect engine/ui to not show quantum options. is this a single uri or a DC instance level property goes to modeling/assumptions part. can we live with a single quantum service for multiple technologies, or will need different quantum for different technologies. this can also be a later phase, starting with a limitation of a single, system-wide quantum.
The way that I see this is that the user should know of the existance of a Quantum server. They just know that Quantum implemnts the networking. It would be ideal if the oVirt engine hides all of the interfaces with the server. This is just an implementation detail.


3.3 network creation

semantics: we create a network at DC level. we attach it at cluster
level. the UI allows you to "create at DC and attach to cluster" at
cluster level.

3.3.1 what if i create a network with same VLAN in two DC's? what if
we attach same logical network to multiple clusters - each will create
the logical network in qunatum again?
If I understand you correctly then I think that one Quantum network can
be created. This should work. In my opinion this should be managed by
the oVirt engine so the actual creation is not an issue. Ensuring that
each cluster has the correct network management configurations is what
is important.
3.3.2 what if the qunatum service is down when engine performs this
action or quantum returns an error?
The engine should be able to deal with this - similar to the way in
which it deals with a network creation when VDSM is down.

network creation is not affected when VDSM is down, as the admin configures it on the hosts after it is created (via the UI).
in quantum case, there is no admin step.
so need to solve engine monitoring quantum for list of networks relevant to a cluster are defined to mark non defined networks as non-operational, and have timer based or manual retry mechanism.
I personally think that the issue of "non operational" status has to be addressed. In my opinion the goal is to have VM's up. I'll give you an example. Say we have VM1, VM2 and VM3 running on one host. VM1 and VM2 communicate on an internal network. VM3 is connectd to the internal network and to the outside work. If there is a problem with the networking then the user is unable to do anything with VM1 and VM2 due to the fact that thir host is non operational... How does one monitor the state of a network? At the moment VDSM just sends a periodic interface update to the engine. The engine has a logic that defines if the host is up or down according to this. I think that it should suffice to the user to know that the network link is up or down. VMware does not have such a strict and limiting system. If a user takes a hardware device and connects it to the wrong switch port they are still able to work on the device. Why should this be different for a virtual machine?

I don't think this is a merge blocker to code, but it should be part of the phase 1 of this feature.
I am not sure that I am in sync with the phases. It would be great maybe if we could describe the scope of the various phases. This will help address the issues for each phase.

3.3.3 each creation of a DC creates a rhevm network - I assume these
are not created, since "no host in a cluster in the DC yet"?
This functionally can remain the same.

ok, goes to assumptions part.
"quantum integration will be done only for vm networks"

3.3.4 what happens on moving of a host to another cluster, or moving
of a cluster to another DC (possible if it's DC was removed iirc).
The Quantum agents take care of this. On each host there will be a
Quantum agent that treats the network management.

I meant in the aspect of provisioning needed configuration to the host (doing it when a host tries to go to up status is fine, but need to remember this use case when desiging, since it means this isn't a bootstrap aspect) also, please document this so flow so someone trying to test the integration covers it.
OK.

3.3.5 shouldn't this be limited to vm networks, or all networks are
relevant?
I think all Quantum networks are relevant

I understood in 3.3.3 that quantum networks are only relevant to vm networks (networks that can run virtual machines, rather than the management, live migration or storage networks)?

(I'd suggest keeping it simple and limited to vm networks for first phase)
I agree

important part here is code should not try to create this network via quantum i guess.
This is why the user should still be able to still have an option of the using the "VDSM style" networking.

3.3.6 what if the host with the qunatum fabric is in maint mode?
I do not understand. When a host is in maintenace mode do VM's receive
traffic? The Quantum port for the VM's can be set as DOWN

I'm offline - we'll need to re-read the wiki to reply.

3.3.7 could a host with qunatum fabric have a VDSM fabric on another
interface (for vm neworks? for non-vm networks)?
Yes. This is something that I would like. I also would like the Quantum

then this should go to assumptions part.

API to be updated so that we can indicate the phycal network interface
that the Quantum network will be running on.

i don't think this is needed.
when admin configures the interface or bond on the host to be 'fabric' or 'quantum' it calls vdsm to perform the network configuration change. this will flag the relevant interface in the host as fabric/quantum (either via config, or alias, etc.)
At the moment Quantum has one network interface for all of the networks. We would like Quantum to be extended so that it can indicate which network is run on which physical interface. For example if NIC 1 (eth0) is on a DMZ and NIC 2 (eth2) is on a provate network, we want the Quantum DMZ network to only be connected to NIC1.

but for modeling sake, can you please explain how this is supposed to look for a UCS host?


3.5 network deletion (detach network from a cluster)
3.5.1 what happens if qunatum is down or returned an error?
The engine should be able to deal with this - similarly to what it does
today

see my reply on 3.3.2. this is managed by the admin today.
so this is coding to be done that should be documented (it is not a merge blocker in my view, but it is part of phase 1.


3.6 vm creation
3.6.1 what about moving a VM from/to a cluster with a quantum fabric?
I do not see a problem here. The agenets running on VDSM will detect and
treat accordingly.

i meant in the aspect of the need to create the port for them in quantum?
I think that we need to discuss the port creation. Quantum enable the user to create a network a port on the network and the ability to attach an interface to the port. I thin that the port creation is something that should be hidden from the oVirt engine user. They just nee to know about networks and that a VM has been connected to the network.

3.6.2 what about import of a VM into a cluster with a qunatum fabric?
Same as above

indeed - need to handle port creations?

3.6.3 you have vm creation/removal - missing vnic addition/removal

please document this operation so someone testing this will have full picture.
:) OK - seems like I have quite a lot of documentation to do. The same test cases should be used as those with the existing VDSM implemntation. Our first goal is to have network parity. Once we have this we can expand the scope.

3.6.4 worth mentioning qunatum doesn't care about the vnic model?
about the mac address?
The Quantum driver on VDSM does take into account the MAC address. This
is used in some plugins - for example the openvswicth plugin.

ok, can you please explain the flow of data of how it gets it (simply vdsm needs to pass it to the driver?)
In the POC implementation that I did the Quantum treatment in VDSM received the following information:-
- mac address of the interface
- VM UUID
- Quantum ID's


3.7 vm start
3.7.1 why is the plugin parameter required at vm start?
This is to indicate to VDSM the operations to take place - for example
updating the openvswitch integration bridge

i thought the integration bridge is pre-configured at host level?
i thought the host already knows which interface is used for quantum, and for sure, which plugin is supposed to be used for that interface.
so why is this a vm start parameter?
When I spoke with the guys from VDSM they did not want me to change the VDSM configuration files. The idea was to deal with each plugin type at run time. This is logical and leaves changes on the VDSM side to a minimal. Less chance for configuration problems.

can multiple plugins be supported in parallel at vdsm/host level or by
the quantum service?
Yes/ Multiple agents can run on VDSM. In the engine a little work is

ok, please document this in the assumptions part.
i assume multiple agents, but still only one per fabric interface?
Correct.
also, if plugins are services, I want to understand when vdsm needs to start them (I assume when admin configures an interface of type fabric for a specific plugin type).
The agent will be running on VDSM. This service is started when the VDSM host boots. The installation of the agent shoud ensure this.

required to run multiple plugins.

in the engine or in quantum?
The plugin runs on the engine. The agent runs on the host.

I updated my reply on assumptions above based on this info as something that needs clearing up.

3.7.2 aren't network_uuid and port uuid redundant to attachment uuid
(I assume qunatum service knows from attachment the port and network)
- i have no objection to passing them to vdsm, just trying to
understand reasoning for this.
I am missing what happens at vdsm level in this point (even after
reading the matching vdsm part)
The network ID and the port ID are returned by the Quantum service. The
attachment ID is passed to the Quantum server. If this is unique then it
can be a key to the above (I am currently working on the scale issue
with Quantum and there are 2 options, the first is that it is unique,
the second is that all three are passed to the agent). It is a few extra
bytes passed. I think that it is better to be safe and have all of the
information on VDSM for future use.

3.8 vm suspend/resume (it's fine to say it behaves like X, but still
need to be covered to not cause bugs/regressions)
The Quantum port status can be updated.

please document this operation so someone testing this will have full picture.

OK

3.9 vm stop
Same as above

please document this operation so someone testing this will have full picture.
OK

3.9.1 need to make sure vm stop when engine is down is handled correctly.

this requires adhering to. again, not a merge blocker to code. but in my view part of phase 1 (my view of phase 1 is deployment is still manual, but flows are handled correctly and do not leave leftovers around) please document this operation so someone testing this will have full picture.

3.9.2 what happens if qunatum service is down? unlike start vm or
network creation the operation in this case cannot be
stopped/rolledback, only rolled forward.
We have to ensure that it is up.

we have several use cases requiring to be able to start virtual machines even when engine is down (regardless of HA aspects for the engine). this means admin wants to go to a host, and start the VM from command line. the VM definition is recovered from the OVF for such a manual emergency procedure.
sanlock is supposed to protect against vm corruption in this case.
but the VMs will need networking to be useful.
so what's the emergency flow to starting a VM?
If the user is usingt he VDSM command line for VM creation then the same support should be offered for Quantum. The underlying assumption here is that the VM is using a network that exists. If this is the case then we should be OK here.


3.10 hot plug nic?
Each new NIC has an attachment ID - the agent know that a new NIC is
added and treats accordingly.

true, but i assume you would need to change the api here to get the quantum details like the attachemnt_id (and the rest). also, all affected use cases should be documented for someone testing this to have full picture of scope.
The oVirt engine for allocating the attachment ID. The fact that the engine does the hot nic assignment ensure that it will also assigned the attachment. I have yet to understand the hot vnic flow - the code that I was working on did not have this support. I should also test this with Quantum.


3.11 vm migration
3.11.1 ok, so this is the first time i understand hosts can be mixed
in same cluster. worth specifying this in the beginning (still not
clear if mixed plugins can exist)
If the networks are configured with the same characteristics then this
should be OK. As mentioned above we are working on a blueprint to deal
with connecting to existing networks - i.e. enable the user/admin to
configure the VLAN tags

please define under assumptions part what are the network characteristics, how they are mapped to which plugins support which characteristics to be relevant. also, monitoring wise, you are implying each host must be able to accommodate all types of characteristics in the networks in the cluster.
Monitoring whould be discussed. Statistics are currently not reported.

3.11.2 afair, we don't deal with engine talking to both hosts during
live migration. only to host A, who is then communicating with host B.
so why not always have the VM configuration at vm start (and hot plug
nic) have the quantum details so live migration can occur at will
without additional information?
I do not understand can you please explain. VDSM creates the tap device
and builds the libvirt files. The agents detect the tap device and
attach to the network. I do not understand why it is a problem for the
live migration. This is also driven by the libvirt XML's being created.
Is this correct?

that is fine. iirc, the wiki states engine needs to pass information to target host. we don't have such a flow in live migration (which is complex enough). live migration should be based on the information already available to vdsm in source host. I don't know if quantum cares about the attachement/port residing on more than one host (the target host has to set them up before migration starts, and source host removes them only after migration ends).
i hope we are both understanding this the same way.

in any case, need to remember to test failure of live migration in its various steps cleans up any quantum related changes.

3.11.3 "In order to implement the above a REST client needs to be
implemented in the oVirt engine. "
did not understand this statement - please elaborate.
All interface with the Quantum server is done via REST. In order for
oVirt to be able to communicate, it will need to send REST messages to
the server and be able to parse the replies - this is what I meant by
the REST client.

4. host management
4.1 deployment
we do not deploy packages from engine to hosts, we can install them
from a repo configured to the host. but this is done today only as
part of initial bootstrap/installation of host.
also, it is not relevant for ovirt node which is 'firmware' like.
any reason to not require the 'plugin installation packages' at vdsm
rpm level for plugins we think are good enough to use (until then,
responsibility to deploy them is of admin)

Correct. I agree with you.

ok, so on ovirt-node all quantum agents will be pre-installed.
if a quantum agent is changed to work correctly with ovirt/vdsm, why not simply always require at rpm level to deploy it on the host?
then a regular host and ovirt-node are the same.
still need to define how vdsm reports which agents are supported by it (which is different than those which are installed, say for UCS?), which is different from those that are actively running since they are configured).
I was thinking that when VDSM dos the "handshake" with the engine that this information would be transferred.


(what are plugin level packages at host level - aren't these the agents?)
Each plugin has the relevant packages that should be installed.

4.2 plugin configuration
per DC? per cluster? per host? per plugin? please provide more details
here on configuration details expected and flow of when they are
configured and when they are expected to change?
I think that plugin per cluster is a good start. This could limit live
migration problems.

I agree...
question is this part of phases, or part of assumptions (end goal)
modeling should be aligned with end goal (i am not suggesting it can do in steps, just that end goal should be understood so it doesn't interrupt it. especially for the API part which requires to keep backward compatibility as much as possible).

I think this will merit a per 'ovirt supported' qunatum plugin to see
it works in a way we can use.

4.3 connectivity
again, this requires more details. if needed per plugin.
what is expected? how authentication/encryption happens? what iptables
rules need to change in engine/host if at all, etc.
I'm fine with this being 'direct access to db from hosts' for POC
level of patches, but not for something we actually merge/provide
support for later.
I am currently working on a blueprint to ensure better scaling of
quantum agents. This will be done by making use of the nova RPC library.
This supports Qpid, rabbit mq, kombu etc. These have an option of being
secure. Please clarify if this suffices?

having all work on same channel will solve most problems.
still need to map those that require data not in quantum for some reason to try and pass to qunatum the needed data as customer parameters rather than add a channel to engine if possible/makes sense.

do all of these require a broker to be deployed?
I assume all support certificate based authentication.

it should be the same one we'll pick for engine-vdsm (I heard fans of zeromq as well), but this doesn't have to be phase 1, or we might be lucky and agree on one already supported :)


5. VDSM
...


5.2 "The agent package can and may be received from the oVirt Engine
or can be downloaded via RPM's"
see 4.1 above - we don't deploy rpm's/code on the fly today.
When I was sitting with the guys from oVirt this is what I was told. I
guess that I misunderstood.

ok, I assume 4.1 resolves this and we simply deploy all those made to work with ovirt as part of vdsm rpm requirements?
I think that we need to discuss packaging - should the quantum agents be part of VDSM? I am not sure.

(implies each one should have a sub rpm, not sure if that is the case today)


5.3 "in addition to the treatment below VDSM should also maintain a
health check to the Quantum agent"
what if the restart didn't help? how should ovirt treat the host wrt
networking? to running VMs? to ability to live migrate VMs from the
host if needed?
If the agent is down then the oVirt engine should be notified to at
least events for the user.

ok.
I assume vdsm knows which agent to start and later monitor based on the host level config by admin of the fabric interface (whether the plugin type is cluster or interface level)
This is related to the issue above. If the user runs the RPM for VDSM then she/he can run the agent RPM - this will register the agent.


5.4 "Logical Network Management"
please see 3.11 above - we would want live migration to not require
additional information, so details should be available even if not
immediately acted upon.

5.5 "The tap device created uses an "ethernet" network device. This
means that the creation of the libvirt XML file is a bit different."
Yes, that is correct
5.5.1 does this impact stable device addresses somehow?
Not that I am aware of
5.5.2 how is live migration possible if the libvirt xml to a non
quantum host is different (or is this the binding part only)?
If the libvirt is "pacthed" when the migration takes place then this
should not be a problem.

does libvirt support something like that? the live migration happens directly between the two libvirt's. libvirt created an abstraction layer to their networking model to allow live migration between host where the physical layer is different so the logical layer remains the same during live migration. I just don't know it the 'ethernet' attribute you mentioned is in that physical mapping layer (which would be fine), or in the logical one (which won't live migrate).
please verify with a libvirt expert (dbp modeled that part iirc)
Thanks, I'll speak with Dan


5.6 I assume libvirtvm.py is part of vdsm.
is quantum.py part of quantum code base or vdsm codebase (it sounds
like it should be part of quantum code base)?
so how exactly the rpm for this would look like? deploy it to
/xxx/qunatum and have it add a symbolic link to it from vdsm paths?
In my opinion this should be part of VDSM. It would be ideal if each
plugin can bind load a driver - in Nova each driver is part of the nova
code. If the API is well defined then all vendors can provide their
drivers. This is open for discussion and we should try and understand
what the best way of doing this is. I like the NOVA model.

doesn't this mean we need to recode each driver for vdsm?
I thought it was part of making quantum a separate entity that it provides drivers as well (especially if the api for drivers is well defined)?
OpenStack nova has Quantum drivers. It would be best if we could use the same interface.


5.7 "When a communication channel is established between VDSM and the
oVirt engine. The VDSM host should notify the oVirt Engine of the
Quantum fabric that is supported."
how does vdsm goes about detecting an agent is installed exactly?
The user needs to install a package for the agent - this creates the
relevant configuartion files. These files can be used to detect the
running agent or via "ps"

1. I assume per the discussion in the deployment section that all relevant code is always deployed.
2. vdsm needs to report what is installed
3. vdsm needs to get the chosen technology at host level fabric interface definition to configure it to up state. 4. or, as part of engine "init_vds_on_up" flow, engine needs to pass to vdsm quantum relevant configuration if defined at cluster level and not relevant to interface level maybe)

(especially since deployment wise, most likely is all agents would be
deployed via rpm requirements?)
is any agent a service at host level rather than code only?
is there a scenario where a vdsm level plugin isn't required?

6. open issues
6.1 worth mentioning if any are in the works
No shortage :). But we have a good start. There are some blueprints in
the works which solve a large amount of problems

I'd add links to those that exists.
https://blueprints.launchpad.net/quantum/+spec/melange-integration
https://blueprints.launchpad.net/quantum/+spec/authorization-support-for-quantum (may be relevant for authentication)
https://blueprints.launchpad.net/quantum/+spec/provider-networks
https://blueprints.launchpad.net/quantum/+spec/quantum-usage
https://blueprints.launchpad.net/quantum/+spec/linuxbridge-portstats-support
https://blueprints.launchpad.net/quantum/+spec/network-statistics
https://blueprints.launchpad.net/quantum/+spec/scalable-agent-comms

6.2 specifying the vlan from ovirt engine is not a gap?
Yes it is. This is currently being addressed. This is not a reason to
stop us moving ahead.
6.3 ok, so this is the first time "no multiple plugins" is mentioned
("Non-uniform technology"). but it sounds like the approach in the
wiki assume it is/will be possible to have multiple technologies
(agents/plugins) going forward?
My take is that this can be hidden by oVirt engine. It is just an
implementation detail - this should not affect the user experience -
which at the end of the day is what counts

this is kind of a crucial part of the assumptions/modeling part for the engine.

6.4 need to discuss which of the open issues should block going
forward with merging of code, and expected timeframe for resolution of
some of them
Correct. In my opinion there are no major blocking issues at the moment.

again, apart of code level issues, we need to clear the modeling so code review can be done compared to some understanding of the plan.

in my view, clarifying these is blocking understanding the feature:
- defining the assumptions (even if some are not supported yet)
- modeling per the assumptions
- UI mockups per the modeling (since they make things visible)
- API modeling, since it is a pain on any one integrating with it when
  it changes
- db scheme changes
Quantum is evolving and changing. We will have to be flexible...

I agree solving these isn't blocking merging code:
- communication channels between quantum components
- deployment implications/flow

I do think these would block further consumption, so they should be covered/understood as to what the gaps are.

for communication part, i think the approach you suggested should be fine (assuming it will merge with the same brokering technology used by engine/vdsm)

but for deployment part, "how this will look like" should be cleared up in a bit more detail to have the gap list (on both engine and vdsm)


7. other questions/items (some mentioned above as well)
7.1 please add ui mockups, db scheme changes and REST api changes
OK
7.2 i'm missing the deployment flow:
OK
- user install engine. is quantum a separate install or bundled going
forward?
- any configuration items by user at this point? what are they?
- what rpms are deployed at host level? what configuration do they need
- communication channels and their security are a must to understand
7.3 upgrade path / compatibility levels this is relevant to?
7.4 monitoring flow - how is monitoring/state of a host is affected by
quantum service being down, communication problem from host to (where
actually?)

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to