Hi Gianluca,

I forgot to mention that you need to ensure that systemd knows that the new 
file exists. You should likely run `systemctl daemon-reload` after 
creating/modifying your custom systemd files. You can see that the After 
directive is combined from both files. Check it out by running `systemctl show 
vdsmd.service | grep After`

It makes sense to make further changes to ensure that NFS stops last, but I 
haven't looked into that yet.
:-)

Cheers,
Gervais



> On Oct 3, 2016, at 7:22 AM, Gianluca Cecchi <gianluca.cec...@gmail.com> wrote:
> 
> 
> Il 28/Set/2016 21:09, "Gervais de Montbrun" <gerv...@demontbrun.com 
> <mailto:gerv...@demontbrun.com>> ha scritto:
> >
> > Hi Gianluca,
> >
> > Instead of editing the system's built in systemd configuration, you can do 
> > the following...
> >
> > Create a file called /etc/systemd/system/ovirt-ha-broker.service
> >
> >> # My custom ovirt-ha-broker.service config that ensures NFS starts before 
> >> ovirt-ha-broker.service
> >> # thanks Gervais for this tip!  :-)
> >>
> >> .include /usr/lib/systemd/system/ovirt-ha-broker.service
> >>
> >> [Unit]
> >> After=nfs-server.service
> >
> >
> > Then disable and enable ovirt-ha-broker.service (systemctl disable 
> > ovirt-ha-broker.service ; systemctl enable ovirt-ha-broker.service) and you 
> > should see that it is using your customized systemd unit definition. You 
> > can see that systemd is using your file by running systemctl status 
> > ovirt-ha-broker.service. You'll see something like "Loaded: loaded 
> > (/etc/systemd/system/ovirt-ha-broker.service;" in the output.
> >
> > Your file will survive updates and therefore always wait for nfs to start 
> > prior to starting. You can do the same for your other customizations.
> >
> > Cheers,
> > Gervais
> >
> >
> >
> >> On Sep 28, 2016, at 1:31 PM, Gianluca Cecchi <gianluca.cec...@gmail.com 
> >> <mailto:gianluca.cec...@gmail.com>> wrote:
> >>
> >> On Sun, Sep 4, 2016 at 10:54 AM, Yedidyah Bar David <d...@redhat.com 
> >> <mailto:d...@redhat.com>> wrote:
> >>>
> >>> On Sat, Sep 3, 2016 at 1:18 PM, Gianluca Cecchi
> >>> <gianluca.cec...@gmail.com <mailto:gianluca.cec...@gmail.com>> wrote:
> >>> > Hello,
> >>> > how do the two modes apply in case of single host?
> >>> > During an upgrade phase, after having upgraded the self hosted engine 
> >>> > and
> >>> > leaving global maintenance and having checked all is ok, what is the 
> >>> > correct
> >>> > mode then to put host if I want finally to update it too?
> >>>
> >>> The docs say to put hosts to maintenance from the engine before upgrading 
> >>> them.
> >>>
> >>> This is (also) so that VMs on them are migrated away to other hosts.
> >>>
> >>> With a single host, you have no other hosts to migrate VMs to.
> >>>
> >>> So you should do something like this:
> >>>
> >>> 1. Set global maintenance (because you are going to take down the
> >>> engine and its vm)
> >>> 2. Shutdown all other VMs
> >>> 3. Shutdown engine vm from itself
> >>> At this point, you should be able to simply stop HA services. But it
> >>> might be cleaner to first set local maintenance. Not sure but perhaps
> >>> this might be required for vdsm. So:
> >>> 4. Set local maintenance
> >>> 5. Stop HA services. If setting local maintenance didn't work, perhaps
> >>> better stop also vdsm services. This stop should obviously happen
> >>> automatically by yum/rpm, but perhaps better do this manually to see
> >>> that it worked.
> >>> 6. yum (or dnf) update stuff.
> >>> 7. Start HA services
> >>> 8. Check status. I think you'll see that both local and global maint
> >>> are still set.
> >>> 9. Set maintenance to none
> >>> 10. Check status again - I think that after some time HA will decide
> >>> to start engine vm and should succeed.
> >>> 11. Start all other VMs.
> >>>
> >>> Didn't try this myself.
> >>>
> >>> Best,
> >>> --
> >>> Didi
> >>
> >>
> >> Hello Didi,
> >> I would like to leverage the update I have to do on 2 small different lab 
> >> environments to crosscheck the steps suggested.
> >> They are both single host environments with self hosted engine.
> >> One is 4.0.2 and the other is 4.0.3. Both on CentoS 7.2
> >> I plan to migrate to the just released 4.0.4
> >>
> >> One note: in both environments the storage is NFS and is provided by the 
> >> host itself, so a corner case (for all hosted_storage domain, main data 
> >> domain and iso storage domain).
> >> I customized the init scripts, basically for start phase of the server and 
> >> to keep in count of the NFS service, but probably something has to be done 
> >> for stop too?
> >>
> >> 1) In /usr/lib/systemd/system/ovirt-ha-broker.service
> >>
> >> added in section [Unit]
> >>
> >> After=nfs-server.service
> >>
> >> The file is overwritten at update so one has to keep in mind this
> >>
> >> 2) also in vdsmd.service changed 
> >> from:
> >> After=multipathd.service libvirtd.service iscsid.service rpcbind.service \
> >>       supervdsmd.service sanlock.service vdsm-network.service
> >>
> >> to:
> >> After=multipathd.service libvirtd.service iscsid.service rpcbind.service \
> >>       supervdsmd.service sanlock.service vdsm-network.service \
> >>       nfs-server.service
> >>
> >> Do you think any order setup I have to put in place related to NFS service 
> >> and oVirt services stop?
> >>
> >> _______________________________________________
> >> Users mailing list
> >> Users@ovirt.org <mailto:Users@ovirt.org>
> >> http://lists.ovirt.org/mailman/listinfo/users 
> >> <http://lists.ovirt.org/mailman/listinfo/users>
> >
> >
> 
> Nice! I'm going to try and see.
> Any particular dependency I should add for shutdown order due to the fact 
> that my host is also the NFS server providing data stores?
> Do I need to set up nfs stop only after a particular ovirt related service?
> Thanks,
> Gianluca

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to