Re: [ovirt-users] hosted-engine unknow stale-data

2018-01-21 Thread Artem Tambovskiy
Hello Kasturi,

Yes, I set global maintenance mode intentionally,
I'm run out of the ideas troubleshooting my cluster and decided to undeploy
the hosted engine from second host, clean the installation and add again to
the cluster.
Also I cleaned the metadata with *hosted-engine --clean-metadata
--host-id=2 --force-clean *But once I added the second host to the cluster
again it doesn't show the capability to run hosted engine. And doesn't even
appear in the output hosted-engine --vm-status
[root@ovirt1 ~]#hosted-engine --vm-status --== Host 1 status ==--
conf_on_shared_storage : True Status up-to-date : True Hostname :
ovirt1.telia.ru Host ID : 1 Engine status : {"health": "good", "vm": "up",
"detail": "up"} Score : 3400 stopped : False Local maintenance : False
crc32 : a23c7cbd local_conf_timestamp : 848931 Host timestamp : 848930
Extra metadata (valid at timestamp): metadata_parse_version=1
metadata_feature_version=1 timestamp=848930 (Mon Jan 22 09:53:29 2018)
host-id=1 score=3400 vm_conf_refresh_time=848931 (Mon Jan 22 09:53:29 2018)
conf_on_shared_storage=True maintenance=False state=GlobalMaintenance
stopped=False

On redeployed second host I see unknown-stale-data again, and second host
doesn't show up as a hosted-engine capable.
[root@ovirt2 ~]# hosted-engine --vm-status


--== Host 1 status ==--

conf_on_shared_storage : True
Status up-to-date  : False
Hostname   : ovirt1.telia.ru
Host ID: 1
Engine status  : unknown stale-data
Score  : 0
stopped: False
Local maintenance  : False
crc32  : 18765f68
local_conf_timestamp   : 848951
Host timestamp : 848951
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=848951 (Mon Jan 22 09:53:49 2018)
host-id=1
score=0
vm_conf_refresh_time=848951 (Mon Jan 22 09:53:50 2018)
conf_on_shared_storage=True
maintenance=False
state=ReinitializeFSM
stopped=False


Really strange situation ...

Regards,
Artem



On Mon, Jan 22, 2018 at 9:46 AM, Kasturi Narra  wrote:

> Hello Artem,
>
> Any reason why you chose hosted-engine undeploy action for the second
> host ? I see that the cluster is in global maintenance mode, was this
> intended ?
>
> command to clear the entries from hosted-engine --vm-status is "hosted-engine
> --clean-metadata --host-id= --force-clean"
>
> Hope this helps !!
>
> Thanks
> kasturi
>
>
> On Fri, Jan 19, 2018 at 12:07 AM, Artem Tambovskiy <
> artem.tambovs...@gmail.com> wrote:
>
>> Hi,
>>
>> Ok, i decided to remove second host from the cluster.
>> I reinstalled from webUI it with hosted-engine action UNDEPLOY, and
>> removed it from the cluster aftewards.
>> All VM's are fine hosted engine running ok,
>> But hosted-engine --vm-status still showing 2 hosts.
>>
>> How I can clean the traces of second host in a correct way?
>>
>>
>> --== Host 1 status ==--
>>
>> conf_on_shared_storage : True
>> Status up-to-date  : True
>> Hostname   : ovirt1.telia.ru
>> Host ID: 1
>> Engine status  : {"health": "good", "vm": "up",
>> "detail": "up"}
>> Score  : 3400
>> stopped: False
>> Local maintenance  : False
>> crc32  : 1b1b6f6d
>> local_conf_timestamp   : 545385
>> Host timestamp : 545385
>> Extra metadata (valid at timestamp):
>> metadata_parse_version=1
>> metadata_feature_version=1
>> timestamp=545385 (Thu Jan 18 21:34:25 2018)
>> host-id=1
>> score=3400
>> vm_conf_refresh_time=545385 (Thu Jan 18 21:34:25 2018)
>> conf_on_shared_storage=True
>> maintenance=False
>> state=GlobalMaintenance
>> stopped=False
>>
>>
>> --== Host 2 status ==--
>>
>> conf_on_shared_storage : True
>> Status up-to-date  : False
>> Hostname   : ovirt1.telia.ru
>> Host ID: 2
>> Engine status  : unknown stale-data
>> Score  : 0
>> stopped: True
>> Local maintenance  : False
>> crc32  : c7037c03
>> local_conf_timestamp   : 7530
>> Host timestamp : 7530
>> Extra metadata (valid at timestamp):
>> metadata_parse_version=1
>> metadata_feature_version=1
>> timestamp=7530 (Fri Jan 12 16:10:12 2018)
>> host-id=2
>> score=0
>> vm_conf_refresh_time=7530 (Fri Jan 12 16:10:12 2018)
>> conf_on_shared_storage=True
>> 

Re: [ovirt-users] oVirt 4.2 - Removed disk remains in VM OS with hooks?

2018-01-21 Thread Yaniv Kaul
On Mon, Jan 22, 2018 at 7:21 AM, Zip  wrote:

> I am having an issue where when I use the REST API to connect a snapshot
> from another VM to a Backup-Appliance-VM, after the clone when I remove the
> Disk and delete the Snapshot, the disk remains in the Backup-Appliance-VM
> as /dev/sdb ot /dev/vdb.
>
> If I reboot the Bakup-Applicance-VM the disk disappears.
>
> If I manually remove the disk by "echo 1 > /sys/block/sdb/device/delete”
> the disk will disappear, but if I rescan the scsi bus, it is found and
> shows up again in the VM OS, but the oVirt WebGUI does NOT show it as
> connected.
>

The first part is expected - the 2nd isn't.


>
> I am also not able to attach any other disks as it complains of :
>
> HotPlugDiskVDS failed: internal error: unable to execute QEMU command
> '__com.redhat_drive_add': Duplicate ID 'drive-scsi0-0-0-2' for drive
>
> I did see that others in the past have gotten around this issue by
> rebooting the Backup-Appliance-VM and then continuing on with the next VM
> backup and looping through backup-reboot-backup-reboot-etc.
>
> Anyone have an idea on how to solve this issue and remove the hooks from
> the guest OS?
>
> Steps to reproduce this issue:
>
>
>1. Create a backup appliance VM to be used for the backup script
>execution
>2. Currently I have the Vms set to virtio with threaded I/O enabled.
>Also tried virtio_scsi with same result.
>3. Using REST API – make snapshot of target VM
>4. Using REST API – fetch vm metadata
>5. Using REST API – attach the snapshot/disk to the Backup-Appliance-VM
>6. dd the drive to backup folder
>7. Using REST API – remove the disk from the Backup-Appliance-VM
>8. Using REST API – delete the snapshot
>9. ** Check the guest OS of the Backup-Appliance-VM and the mounted
>drive from the backup above still appears and behaves as mentioned in
>comments above.
>
>
>
There are many details missing, including versions of everything used, but
logs would be most helpful here.


> A second issue is that the above wont work when I have the Vms running on
> MPIO iSCSI storage, so for testing I have moved to NFS4. Anyone have ideas
> about either issue, I’d love to hear ;)
>

Same - logs would be helpful here.
Y.


>
> Thanks
>
> Irc.oftc.net #ovirt
> zipur
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] VirtIO-SCSI and viodiskcache custom property

2018-01-21 Thread Yaniv Kaul
On Mon, Jan 22, 2018 at 6:53 AM, Zip  wrote:

> Does it help if you edit the VM and under Resource Allocation – enable IO
> threads. ?
>

Confusingly, it has nothing to do with IO threads.
Y.


>
> Zip
>
>
>
> From:  on behalf of Yaniv Kaul 
> Date: Saturday, January 20, 2018 at 12:54 PM
> To: Matthias Leopold 
> Cc: Wolfgang Lendl , Ovirt Users <
> users@ovirt.org>
> Subject: Re: [ovirt-users] VirtIO-SCSI and viodiskcache custom property
>
>
>
> On Jan 19, 2018 3:29 PM, "Matthias Leopold"  ac.at> wrote:
>
> Hi,
>
> is there a reason why the viodiskcache custom property isn't honored when
> using VirtIO-SCSI?
>
> On a Cinder (Ceph) disk "viodiskcache=writeback" is ignored with
> VirtIO-SCSI and honored when using VirtIO.
>
> On an iSCSI disk "viodiskcache=writeback" is ignored with VirtIO-SCSI and
> the VM can't be started when using VirtIO with "unsupported configuration:
> native I/O needs either no disk cache or directsync cache mode, QEMU will
> fallback to aio=threads"
>
> We actually want to use "viodiskcache=writeback" with Cinder (Ceph) disks.
>
>
> That's because on block storage we use native io and not threads. I assume
> the hook needs to change to use native io in this case.
> Y.
>
>
> oVirt version: 4.1.8
>
> Thanks
> Matthias
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine unknow stale-data

2018-01-21 Thread Kasturi Narra
Hello Artem,

Any reason why you chose hosted-engine undeploy action for the second
host ? I see that the cluster is in global maintenance mode, was this
intended ?

command to clear the entries from hosted-engine --vm-status is "hosted-engine
--clean-metadata --host-id= --force-clean"

Hope this helps !!

Thanks
kasturi


On Fri, Jan 19, 2018 at 12:07 AM, Artem Tambovskiy <
artem.tambovs...@gmail.com> wrote:

> Hi,
>
> Ok, i decided to remove second host from the cluster.
> I reinstalled from webUI it with hosted-engine action UNDEPLOY, and
> removed it from the cluster aftewards.
> All VM's are fine hosted engine running ok,
> But hosted-engine --vm-status still showing 2 hosts.
>
> How I can clean the traces of second host in a correct way?
>
>
> --== Host 1 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : True
> Hostname   : ovirt1.telia.ru
> Host ID: 1
> Engine status  : {"health": "good", "vm": "up",
> "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 1b1b6f6d
> local_conf_timestamp   : 545385
> Host timestamp : 545385
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=545385 (Thu Jan 18 21:34:25 2018)
> host-id=1
> score=3400
> vm_conf_refresh_time=545385 (Thu Jan 18 21:34:25 2018)
> conf_on_shared_storage=True
> maintenance=False
> state=GlobalMaintenance
> stopped=False
>
>
> --== Host 2 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : False
> Hostname   : ovirt1.telia.ru
> Host ID: 2
> Engine status  : unknown stale-data
> Score  : 0
> stopped: True
> Local maintenance  : False
> crc32  : c7037c03
> local_conf_timestamp   : 7530
> Host timestamp : 7530
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=7530 (Fri Jan 12 16:10:12 2018)
> host-id=2
> score=0
> vm_conf_refresh_time=7530 (Fri Jan 12 16:10:12 2018)
> conf_on_shared_storage=True
> maintenance=False
> state=AgentStopped
> stopped=True
>
>
> !! Cluster is in GLOBAL MAINTENANCE mode !!
>
> Thank you in advance!
> Regards,
> Artem
>
>
> On Wed, Jan 17, 2018 at 6:47 PM, Artem Tambovskiy <
> artem.tambovs...@gmail.com> wrote:
>
>> Hello,
>>
>> Any further suggestions on how to fix the issue and make HA setup
>> working? Can the complete removal of second host (with complete removal
>> ovirt configuration files and packages) from cluster and adding it again
>> solve the issue? Or it might completly ruin the cluster?
>>
>> Regards,
>> Artem
>>
>> 16 янв. 2018 г. 17:00 пользователь "Artem Tambovskiy" <
>> artem.tambovs...@gmail.com> написал:
>>
>> Hi Martin,
>>>
>>> Thanks for feedback.
>>>
>>> All hosts and hosted-engine running 4.1.8 release.
>>> The strange thing : I can see that host ID is set to 1 on both hosts at
>>> /etc/ovirt-hosted-engine/hosted-engine.conf file.
>>> I have no idea how this happen, the only thing I have changed recently
>>> is that I have changed mnt_options in order to add backup-volfile-servers
>>> by using hosted-engine --set-shared-config command
>>>
>>> Both agent and broker are running on second host
>>>
>>> [root@ovirt2 ovirt-hosted-engine-ha]# ps -ef | grep ovirt-ha-
>>> vdsm  42331  1 26 14:40 ?00:31:35 /usr/bin/python
>>> /usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker --no-daemon
>>> vdsm  42332  1  0 14:40 ?00:00:16 /usr/bin/python
>>> /usr/share/ovirt-hosted-engine-ha/ovirt-ha-agent --no-daemon
>>>
>>> but I saw some tracebacks during the broker start
>>>
>>> [root@ovirt2 ovirt-hosted-engine-ha]# systemctl status ovirt-ha-broker
>>> -l
>>> ● ovirt-ha-broker.service - oVirt Hosted Engine High Availability
>>> Communications Broker
>>>Loaded: loaded (/usr/lib/systemd/system/ovirt-ha-broker.service;
>>> enabled; vendor preset: disabled)
>>>Active: active (running) since Tue 2018-01-16 14:40:15 MSK; 1h 58min
>>> ago
>>>  Main PID: 42331 (ovirt-ha-broker)
>>>CGroup: /system.slice/ovirt-ha-broker.service
>>>└─42331 /usr/bin/python 
>>> /usr/share/ovirt-hosted-engine-ha/ovirt-ha-broker
>>> --no-daemon
>>>
>>> Jan 16 14:40:15 ovirt2.telia.ru systemd[1]: Started oVirt Hosted Engine
>>> High Availability Communications Broker.
>>> Jan 16 14:40:15 ovirt2.telia.ru systemd[1]: Starting oVirt Hosted
>>> Engine High Availability Communications Broker...
>>> Jan 

[ovirt-users] oVirt 4.2 - Removed disk remains in VM OS with hooks?

2018-01-21 Thread Zip
I am having an issue where when I use the REST API to connect a snapshot
from another VM to a Backup-Appliance-VM, after the clone when I remove the
Disk and delete the Snapshot, the disk remains in the Backup-Appliance-VM as
/dev/sdb ot /dev/vdb.

If I reboot the Bakup-Applicance-VM the disk disappears.

If I manually remove the disk by "echo 1 > /sys/block/sdb/device/delete² the
disk will disappear, but if I rescan the scsi bus, it is found and shows up
again in the VM OS, but the oVirt WebGUI does NOT show it as connected.

I am also not able to attach any other disks as it complains of :

HotPlugDiskVDS failed: internal error: unable to execute QEMU command
'__com.redhat_drive_add': Duplicate ID 'drive-scsi0-0-0-2' for drive

I did see that others in the past have gotten around this issue by rebooting
the Backup-Appliance-VM and then continuing on with the next VM backup and
looping through backup-reboot-backup-reboot-etc.

Anyone have an idea on how to solve this issue and remove the hooks from the
guest OS?

Steps to reproduce this issue:

1. Create a backup appliance VM to be used for the backup script execution
2. Currently I have the Vms set to virtio with threaded I/O enabled. Also
tried virtio_scsi with same result.
3. Using REST API ­ make snapshot of target VM
4. Using REST API ­ fetch vm metadata
5. Using REST API ­ attach the snapshot/disk to the Backup-Appliance-VM
6. dd the drive to backup folder
7. Using REST API ­ remove the disk from the Backup-Appliance-VM
8. Using REST API ­ delete the snapshot
9. ** Check the guest OS of the Backup-Appliance-VM and the mounted drive
from the backup above still appears and behaves as mentioned in comments
above.

A second issue is that the above wont work when I have the Vms running on
MPIO iSCSI storage, so for testing I have moved to NFS4. Anyone have ideas
about either issue, I¹d love to hear ;)

Thanks

Irc.oftc.net #ovirt
zipur




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] VirtIO-SCSI and viodiskcache custom property

2018-01-21 Thread Zip
Does it help if you edit the VM and under Resource Allocation ­ enable IO
threads. ?

Zip



From:   on behalf of Yaniv Kaul 
Date:  Saturday, January 20, 2018 at 12:54 PM
To:  Matthias Leopold 
Cc:  Wolfgang Lendl , Ovirt Users

Subject:  Re: [ovirt-users] VirtIO-SCSI and viodiskcache custom property



On Jan 19, 2018 3:29 PM, "Matthias Leopold"
 wrote:
> Hi,
> 
> is there a reason why the viodiskcache custom property isn't honored when
> using VirtIO-SCSI?
> 
> On a Cinder (Ceph) disk "viodiskcache=writeback" is ignored with VirtIO-SCSI
> and honored when using VirtIO.
> 
> On an iSCSI disk "viodiskcache=writeback" is ignored with VirtIO-SCSI and the
> VM can't be started when using VirtIO with "unsupported configuration: native
> I/O needs either no disk cache or directsync cache mode, QEMU will fallback to
> aio=threads"
> 
> We actually want to use "viodiskcache=writeback" with Cinder (Ceph) disks.

That's because on block storage we use native io and not threads. I assume
the hook needs to change to use native io in this case.
Y. 

> 
> oVirt version: 4.1.8
> 
> Thanks
> Matthias
> 
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> 
> 



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] OVS not running / logwatch error after upgrade from 4.0.6 to 4.1.8

2018-01-21 Thread Derek Atkins
Hi,

I tried creating the directory but then it complained about missing log
files.

I think I'm just going to block out the logwatch file.

It's annoying that the default config bleats out.

Thanks for the feedback.

-derek

Darrell Budic  writes:

> OVS is an optional tech preview in 4.1.x, you don’t need it. It is annoying
> about the logwatch errors though…
>
> I think I created the directory to avoid the errors, I forgot exactly what it
> was, sorry.
>
> --
> From: Derek Atkins 
> Subject: [ovirt-users] OVS not running / logwatch error after upgrade from
> 4.0.6 to 4.1.8
> Date: January 19, 2018 at 10:44:56 AM CST
> To: users
>
> Hi,
> I recently upgraded my 1-host ovirt deployment from 4.0.6 to 4.1.8.
> Since then, the host has been reporting a cron.daily error:
>
> /etc/cron.daily/logrotate:
>
> logrotate_script: line 4: cd: /var/run/openvswitch: No such file or
> directory
>
> This isn't surprising, since:
>
> # systemctl status openvswitch
> ● openvswitch.service - Open vSwitch
>   Loaded: loaded (/usr/lib/systemd/system/openvswitch.service; disabled;
> vendor preset: disabled)
>   Active: inactive (dead)
>
> The host was just upgraded by "yum update".
> Was there anything special that needed to happen after the update?
> Do I *NEED* OVS running?
> The VMs all seem to be behaving properly.
>
> Thanks,
>
> -derek
>
> --
>   Derek Atkins 617-623-3745
>   de...@ihtfp.com www.ihtfp.com
>   Computer and Internet Security Consultant
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>

-- 
   Derek Atkins 617-623-3745
   de...@ihtfp.com www.ihtfp.com
   Computer and Internet Security Consultant
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Hosted-engine without mac-spoofing

2018-01-21 Thread Matteo Capuano
Hi folks,

i'm trying to install oVirt hyperconverged on Google Cloud Engine (GCE) to
create a lab environment.
Nested virtualization on GCE is in public beta, it seems to work fine but
there's a catch: i cannot enable mac-spoofing on the GCE instances' nics.
So here's my question: is there a way to install hosted-engine on a network
there you can't enabling mac-spoofing?

Thank you

Cheers

Matteo
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] New blog on ovirt.org - Build oVirt Reports Using Grafana

2018-01-21 Thread Yaniv Kaul
Grafana , The open platform for beautiful analytics
and monitoring, recently added support for PostgreSQL
.

It in now possible to connect Grafana to oVirt DWH
, in order to
visualize and monitor the oVirt environment.

Read more @ https://ovirt.org/blog/2018/01/ovirt-report-using-grafana/
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] New post on oVirt blog: Build oVirt Reports Using Grafana

2018-01-21 Thread Shirly Radco
Hello everyone,

A new oVirt blog post has been published, on how to build oVirt reports
using Grafana.
This allows connecting to oVirt DWH and create dashboards for System,
Hosts, VMs, Storage domains etc.

See https://ovirt.org/blog/2018/01/ovirt-report-using-grafana/ for the full
post.

For more information you can contact me.

Best regards,

--

SHIRLY RADCO

BI SeNIOR SOFTWARE ENGINEER

Red Hat Israel 

TRIED. TESTED. TRUSTED. 
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users