[ovirt-users] oVirt Ansible cluster deployment

2021-02-23 Thread Jeremey Wise
I am trying to eat own dog food.  Infrastructure as code...

So.. I have taken three physical servers.  Setup all the base packages, set
NICs, NTP, DNS, collected disk UUIDs etc..

The servers are prepped.

And one server has cockpit and ovirt-engine installed... and service
started.

#

---
# This task takes nodes prepaired and deploys cockpit, ovirt
deployment and scripts cluster initilization. ONLY RUN THIS ON FIRST
NODE OF CLUSTER
# v0.001 20210222
# file: /roles/task_hci_setup.yml
# Documentation
# Overview:
https://www.ovirt.org/documentation/gluster-hyperconverged/chap-Deploying_Hyperconverged.html

- name: Check if rpm ovirt-release44 is already installed
  yum:
list: ovirt-release44
  register: yum_list
- name: Conditionally do next thing
  debug:
msg: "ovirt-release44 Not installed"
  when:
yum_list.results | selectattr("yumstate", "match", "installed") | list
| length == 0

# Add oVirt Repo manual.
- name: Download oVirt Repo RPM
  get_url:
 url: https://resources.ovirt.org/pub/yum-repo/ovirt-release44.rpm
 dest: ~/ovirt-release44.rpm
 mode: '0440'
  when:
yum_list.results | selectattr("yumstate", "match", "installed") | list
| length == 0

- name: Install Python3 as this is needed for oVirt.
  yum:
name:
- python3
state: latest
  become: true

- name:
For some reason install of oVirt Repo fails as package so have to run as shell
  command: "rpm -ivh ~/ovirt-release44.rpm"
  when:
yum_list.results | selectattr("yumstate", "match", "installed") | list
| length == 0

- name: With new oVirt repo.. install rest of HCI dependancies
  yum:
name:
- cockpit-ovirt-dashboard
- vdsm-gluster
- ovirt-host
- ovirt-ansible-roles
- ovirt-ansible-infra
state: latest
  become: true

# Set variable for first node in cluster list to set as ignition master
- name: Only on first node thor deploy ovirt-engine-appliance
  yum:
name:
- ovirt-engine-appliance
state: latest
  become: true
  when: "'{{ hci_ignition_node }}' in inventory_hostname"

# on local ansible host install galaxy modules
# Broken: I need a means to check if installed. If so .. don't install again.
# - name: Install Galaxy Modules on local ansible host
#   local_action:
# module: shell
# _raw_params: ansible-galaxy install oVirt.ovirt-ansible-roles
#   # when: https://www.openvirtualization.pro/rhv-ovirt-ansible-getting-started/
#


What I am missing is the logic and though to get the last section of
playbook.. Which is ... what I would do in the wizard... ... but "step away
from the mouse".

# Ansible section:  from bastion node (laptop windows 10 WSL with ansible
and galaxy installed

# 1)  build playbook from my site variables... based on some ovirt master
sharing template of what that UI builds

# 2)   push the playbook that installs ansible on the first node.  Install
ansible as a first node...  Run ansible playbook from customized template
file

# 3)  Create vlans.

# 4) Setup new VM gluster shared volume from three other disk

# 5) Push a template VM into cluster, attached to new VLAN and attached to
new volume

#6) Deploy VM from this template

#7) Drink a cold beer



###

Question:
1) Can you do ansible deployment from "bastion host" or does the ansible
have to run on the node that is installed with the "ovirt-engine". If so..
how do you collect feedback etc..

2) I assume there is a template.  Maybe it is published but I could not
find it.  That I replace (or match) variables of my site that fill in that
file.

3) I installed galaxy oVirt.ovirt-ansible-roles  but.. how do I know it
is deployed on my bastion host?  I poked around in the ansible folder.. not
seeing anything.  And I assume this only is of help POST the above seven
steps being done.  But maybe this has deployment components and I am
missing documents.

Thanks,

-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/T7BMS44D6DZZVUCGYUJ7SO6LHXMTMOC2/


[ovirt-users] oVirt 4.4 - HA Proxy Fencing

2020-10-11 Thread Jeremey Wise
I have a pair of nodes which service DNS / NTP / FTP / AD /Kerberos / IPLB
etc..

ns01, ns02

These two "infrastructure VMs have HA Proxy and pacemaker and I have set to
have "HA" within ovirt and node affinity.

But.. within HAProxy, the nodes use to be able to call the STONITH function
of KVM to reset a node if / as needed.

Fence Device: "stonith:fence_virsh"
But with oVirt this function no longer works.

For VMs and other applications which need STONITH functions, how would this
be done with oVirt.

I would assume the call would be to oVirt-Engine now, but not seeing
documentation on HA-Proxy / Pacemaker to use oVirt.

Can someone point me in the right direction?


-- 
penguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/B4KLZLPAIZMUMUEIUOSPLEI23UAGZVJM/


[ovirt-users] Fencing with oVirt 4.4 apc master switch

2020-10-11 Thread Jeremey Wise
I use to have hand built CentOS+KVM + Gluster.

Moved to HCI oVirt controlled system and so far, been very happy with
stability and quality.

One feature that was working in old "hand build" version was fencing.

I have an old APC master switch AP9606.  I use to have and use it for
master switch functions of nodes but sense replacement with oVirt it has
not worked.   Here are setup notes and symptoms

IP: 172.16.100.2
Username: apc
Password: 
Port 8 (it is an eight port unit)
Slot:  


Error:
Test failed: WARNING:root:Parse error: Ignoring option 'agent' because it
does not have value ERROR:root:Unable to connect/login to fencing device
2020-10-11 11:28:50,885 ERROR: Unable to connect/login to fencing device


Questions:
1) Has anyone setup this with oVirt such that it uses apc master switch
2) Where is oVirt putting these logs / running this service.  I assumed
oVirt engine running the services and so tried to find details and logs
here.



-- 
penguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/GBD7MUJKXDZE56TYUCBJNDE7TWS7CHTZ/


[ovirt-users] CEPH - Opinions and ROI

2020-09-30 Thread Jeremey Wise
I have for many years used gluster because..well.  3 nodes.. and so long as
I can pull a drive out.. I can get my data.. and with three copies.. I have
much higher chance of getting it.

Downsides to gluster: Slower (its my home..meh... and I have SSD to avoid
MTBF issues ) and with VDO.. and thin provisioning.. not had issue.

BUT  gluster seems to be falling out of favor.  Especially as I move
towards OCP.

So..  CEPH.  I have one SSD in each of the three servers.  so I have some
space to play.

I googled around.. and find no clean deployment notes and guides on CEPH +
oVirt.

Comments or ideas..

-- 
p enguinpages.
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UTKROHYPKJOXJKAJPRL37IETMELMXCPD/


[ovirt-users] Re: Version 4.4.2.6-1.el8 -Console Error: java.lang.reflect.UndeclaredThrowableException

2020-09-30 Thread Jeremey Wise
Ok.. made a backup..  ran from ovirt-engin the command " taskcleaner.sh -R
-o -A"

This cleared out issue.. but scared me a bit as it just terminated the
connections / vm. I had to power vm off "hosted-engine --vm-shutdown"  and
then back on "hosted-engine --vm-start"

Off to find the next "feature"

:)


On Wed, Sep 30, 2020 at 4:02 PM Jeremey Wise  wrote:

>
>
> I found this note from RedHat on bugzilla
>
> https://bugzilla.redhat.com/show_bug.cgi?id=1827033
>
> Seems like this could be my issue
>
> This could be where had issue with rebuild of ovirt-engine.  And then had
> to scrape out the disk files to then import them back in. I then moved the
> "old files" into sub folder "junk"  ... so after reboot I could know nobody
> needed them..and delete old orphaned disk from previous oVirt-HCI deploy
> version.
>
> 
> [root@thor images]# tree
> .
> ├── 04002d39-4f60-4ed0-94db-74cb6b8de8d9
> │   ├── 3f68af97-44ac-4a14-926e-dca66f7ce544
> 
> ├── fff3c1de-e21e-4f03-8905-d587448f6543
> │   ├── 6f4343b6-dc3d-40e3-b872-6a954b8d1a7b
> │   ├── 6f4343b6-dc3d-40e3-b872-6a954b8d1a7b.lease
> │   └── 6f4343b6-dc3d-40e3-b872-6a954b8d1a7b.meta
> └── junk
> ├── 138a359c-13e6-4448-b543-533894e41fca
> │   ├── ece912a4-6756-4944-803c-c7ac58713ef4
> │   ├── ece912a4-6756-4944-803c-c7ac58713ef4.lease
> │   └── ece912a4-6756-4944-803c-c7ac58713ef4.meta
> 
> ├── ce4133ad-562f-4f23-add6-cd168a906267
> │   ├── a09c8a84-1904-4632-892e-beb55abc873a
> │   ├── a09c8a84-1904-4632-892e-beb55abc873a.lease
> │   └── a09c8a84-1904-4632-892e-beb55abc873a.meta
> └── f6679e35-fa56-4ed8-aa47-18492e00fd01
> ├── 3f2f0fc7-cb3c-4bbc-9f7b-4f196588c78c
> ├── 3f2f0fc7-cb3c-4bbc-9f7b-4f196588c78c.lease
> └── 3f2f0fc7-cb3c-4bbc-9f7b-4f196588c78c.meta
>
> 31 directories, 113 files
> 
>
> The bug notes running command "taskcleaner.sh -R -o -A"
>
> [root@ovirte01 dbutils]#
> /usr/share/ovirt-engine/setup/dbutils/taskcleaner.sh -R -o -A
>
>  This will remove all Tasks in async_tasks table, and all Job Steps and
> Compensation data!!!
> Caution, this operation should be used with care. Please contact support
> prior to running this command
> Are you sure you want to proceed? [y/n]
> n
> FATAL: Please contact support for further assistance.
> [root@ovirte01 dbutils]#
>
>
> I chickened out...   Asking for input on of this is a good idea
>
> On Wed, Sep 30, 2020 at 10:27 AM penguin pages 
> wrote:
>
>>
>> Got message this AM when tried to login to oVirt Engine which up till now
>> has been working fine.
>>
>> I can supply username and password and get portal to choose
>> "Administration Portal" or "VM Portal"
>>
>> I have tested both.. both have same response about
>> java.lang.reflect.UndeclaredThrowableException
>>
>> I restarted the engine
>> #
>> hosted-engine --set-maintenance --mode=global
>> hosted-engine --vm-shutdown
>> hosted-engine --vm-status
>> #make sure that the status is shutdown before restarting
>> hosted-engine --vm-start
>> hosted-engine --vm-status
>> #make sure the status is health before leaving maintenance mode
>> hosted-engine --set-maintenance --mode=none
>>
>> #
>> --== Host thor.penguinpages.local (id: 1) status ==--
>>
>> Host ID: 1
>> Host timestamp : 70359
>> Score  : 3400
>> Engine status  : {"vm": "down", "health": "bad",
>> "detail": "unknown", "reason": "vm not running on this host"}
>> Hostname   : thor.penguinpages.local
>> Local maintenance  : False
>> stopped: False
>> crc32  : 25adf6d0
>> conf_on_shared_storage : True
>> local_conf_timestamp   : 70359
>> Status up-to-date  : True
>> Extra metadata (valid at timestamp):
>> metadata_parse_version=1
>> metadata_feature_version=1
>> timestamp=70359 (Wed Sep 30 09:35:22 2020)
>> host-id=1
>> score=3400
>> vm_conf_refresh_time=70359 (Wed Sep 30 09:35:22 2020)
>> conf_on_shared_storage=True
>> maintenance=False
>> state=EngineDown
>> stopped=False
>>
>>
>> --== Host medusa.penguinpages.local (id: 3) status ==--
>>
>> Host ID 

[ovirt-users] Re: VM AutoStart

2020-09-30 Thread Jeremey Wise


Ya.. that is a lot easier.

Someone should put this in as a feature request.  I don't want HA (and have
errors on getting that to work)  I just want VMs to boot on initial cluster
start.

this was standard in libvirt :)   I am trying to convert to oVirt ways...

On Wed, Sep 30, 2020 at 4:10 PM Derek Atkins  wrote:

> HI,
>
> On Wed, September 30, 2020 3:50 pm, Jeremey Wise wrote:
> > As the three servers are Centos8 minimal installs. + oVirt HCI wizard to
> > keep them lean and mean... a couple questions
>
> Note that you run this on the Engine VM, not on a host.
>
> > 1) which version of python would I need for this (note in script about
> > python 2 but isn't that deprecated?)
> > [root@thor /]# yum install python
> > Last metadata expiration check: 2:29:38 ago on Wed 30 Sep 2020 01:18:32
> PM
> > EDT.
> > No match for argument: python
> > There are following alternatives for "python": python2, python36,
> python38
> > Error: Unable to find a match: python
>
> I am still running 4.3, so "python" is 2.7.
> I have not tested with python3..
>
> > 2)  When you have three nodes.. one is set to host the ovirt-engine
> > active,
> > and another as backup.  If this is added to rc.local.   Of the two nodes
> > hosting HA for oVirt-engine.. node which boots first will host (or so it
> > seems). I think if I add this to both those hosts .. it will not create
> > issues.  Any thoughts?
>
> Don't run it on a host, run it from within the Engine VM.
>
> The host(s) will figure out by themselves that they need to start the
> engine if one isn't running.  Then when the engine starts the script will
> run and start the VMs.
>
> -derek
>
> >
> > On Wed, Sep 30, 2020 at 3:23 PM Derek Atkins  wrote:
> >
> >> I run it out of rc.local:
> >>
> >> /usr/local/sbin/start_vms.py > /var/log/start_vms 2>&1 &
> >>
> >> The script is smart enough to wait for the engine to be fully active.
> >>
> >> -derek
> >>
> >> On Wed, September 30, 2020 3:11 pm, Jeremey Wise wrote:
> >> > i would like to eventually go ansible route..  and was starting down
> >> that
> >> > path but this is fabulous.
> >> >
> >> > I will modify and post how it went.
> >> >
> >> > One question:  How /where do you set this saved new and delicious
> >> script
> >> > so
> >> > once oVirt-engine comes up... it runs?
> >> >
> >> > Thanks
> >> >
> >> > On Wed, Sep 30, 2020 at 2:42 PM Derek Atkins  wrote:
> >> >
> >> >> Hi,
> >> >>
> >> >> I had a script based around ovirt-shell which I re-wrote as a script
> >> >> around the Python SDK4 which I run on my engine during the startup
> >> >> sequence.  The script will wait for the engine to come up and ensure
> >> the
> >> >> storage domains are up before it tries to start the VMs.  Then it
> >> will
> >> >> go
> >> >> ahead and start the VMs in the specified order with specified delay
> >> >> and/or
> >> >> wait-for-up signal between them.
> >> >>
> >> >> You can find my scripts at https://www.ihtfp.org/ovirt/
> >> >>
> >> >> Or you can go the ansible route :)
> >> >>
> >> >> Enjoy!
> >> >>
> >> >> -derek
> >> >>
> >> >> On Wed, September 30, 2020 11:21 am, Jeremey Wise wrote:
> >> >> > When I have to shut down cluster... ups runs out etc..  I need a
> >> >> sequence
> >> >> > set of just a small number of VMs to "autostart"
> >> >> >
> >> >> > Normally I just use DNS FQND to connect to oVirt engine but as two
> >> of
> >> >> my
> >> >> > VMs  are a DNS HA cluster..  as well as NTP / SMTP /DHCP etc...  I
> >> >> need
> >> >> > those two infrastructure VMs to be auto boot.
> >> >> >
> >> >> > I looked at HA settings for those VMs but it seems to be watching
> >> for
> >> >> > pause
> >> >> > /resume.. but it does not imply or state auto start on clean first
> >> >> boot.
> >> >> >
> >> >> > Options?
> >> >> >
> >> >> >
> >> >> > --
> >> >> > p enguinpages
> >> >&g

[ovirt-users] Re: Version 4.4.2.6-1.el8 -Console Error: java.lang.reflect.UndeclaredThrowableException

2020-09-30 Thread Jeremey Wise
I found this note from RedHat on bugzilla

https://bugzilla.redhat.com/show_bug.cgi?id=1827033

Seems like this could be my issue

This could be where had issue with rebuild of ovirt-engine.  And then had
to scrape out the disk files to then import them back in. I then moved the
"old files" into sub folder "junk"  ... so after reboot I could know nobody
needed them..and delete old orphaned disk from previous oVirt-HCI deploy
version.


[root@thor images]# tree
.
├── 04002d39-4f60-4ed0-94db-74cb6b8de8d9
│   ├── 3f68af97-44ac-4a14-926e-dca66f7ce544

├── fff3c1de-e21e-4f03-8905-d587448f6543
│   ├── 6f4343b6-dc3d-40e3-b872-6a954b8d1a7b
│   ├── 6f4343b6-dc3d-40e3-b872-6a954b8d1a7b.lease
│   └── 6f4343b6-dc3d-40e3-b872-6a954b8d1a7b.meta
└── junk
├── 138a359c-13e6-4448-b543-533894e41fca
│   ├── ece912a4-6756-4944-803c-c7ac58713ef4
│   ├── ece912a4-6756-4944-803c-c7ac58713ef4.lease
│   └── ece912a4-6756-4944-803c-c7ac58713ef4.meta

├── ce4133ad-562f-4f23-add6-cd168a906267
│   ├── a09c8a84-1904-4632-892e-beb55abc873a
│   ├── a09c8a84-1904-4632-892e-beb55abc873a.lease
│   └── a09c8a84-1904-4632-892e-beb55abc873a.meta
└── f6679e35-fa56-4ed8-aa47-18492e00fd01
├── 3f2f0fc7-cb3c-4bbc-9f7b-4f196588c78c
├── 3f2f0fc7-cb3c-4bbc-9f7b-4f196588c78c.lease
└── 3f2f0fc7-cb3c-4bbc-9f7b-4f196588c78c.meta

31 directories, 113 files


The bug notes running command "taskcleaner.sh -R -o -A"

[root@ovirte01 dbutils]#
/usr/share/ovirt-engine/setup/dbutils/taskcleaner.sh -R -o -A

 This will remove all Tasks in async_tasks table, and all Job Steps and
Compensation data!!!
Caution, this operation should be used with care. Please contact support
prior to running this command
Are you sure you want to proceed? [y/n]
n
FATAL: Please contact support for further assistance.
[root@ovirte01 dbutils]#


I chickened out...   Asking for input on of this is a good idea

On Wed, Sep 30, 2020 at 10:27 AM penguin pages 
wrote:

>
> Got message this AM when tried to login to oVirt Engine which up till now
> has been working fine.
>
> I can supply username and password and get portal to choose
> "Administration Portal" or "VM Portal"
>
> I have tested both.. both have same response about
> java.lang.reflect.UndeclaredThrowableException
>
> I restarted the engine
> #
> hosted-engine --set-maintenance --mode=global
> hosted-engine --vm-shutdown
> hosted-engine --vm-status
> #make sure that the status is shutdown before restarting
> hosted-engine --vm-start
> hosted-engine --vm-status
> #make sure the status is health before leaving maintenance mode
> hosted-engine --set-maintenance --mode=none
>
> #
> --== Host thor.penguinpages.local (id: 1) status ==--
>
> Host ID: 1
> Host timestamp : 70359
> Score  : 3400
> Engine status  : {"vm": "down", "health": "bad",
> "detail": "unknown", "reason": "vm not running on this host"}
> Hostname   : thor.penguinpages.local
> Local maintenance  : False
> stopped: False
> crc32  : 25adf6d0
> conf_on_shared_storage : True
> local_conf_timestamp   : 70359
> Status up-to-date  : True
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=70359 (Wed Sep 30 09:35:22 2020)
> host-id=1
> score=3400
> vm_conf_refresh_time=70359 (Wed Sep 30 09:35:22 2020)
> conf_on_shared_storage=True
> maintenance=False
> state=EngineDown
> stopped=False
>
>
> --== Host medusa.penguinpages.local (id: 3) status ==--
>
> Host ID: 3
> Host timestamp : 92582
> Score  : 3400
> Engine status  : {"vm": "up", "health": "good",
> "detail": "Up"}
> Hostname   : medusa.penguinpages.local
> Local maintenance  : False
> stopped: False
> crc32  : 623359d2
> conf_on_shared_storage : True
> local_conf_timestamp   : 92582
> Status up-to-date  : True
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=92582 (Wed Sep 30 09:35:25 2020)
> host-id=3
> score=3400
> vm_conf_refresh_time=92582 (Wed Sep 30 09:35:25 2020)
> conf_on_shared_storage=True
> maintenance=False
> state=EngineUp
> stopped=False
> ##
>
> I downloaded and installed key from portal.. thinking that may have been
> issue.. it was not.
>
> I googled around /searched forum and nothing jumped out.   (only hit I
> found in forum
> https://lists.ovirt.org/pipermail/users/2015-June/033421.html  but no
> note about 

[ovirt-users] Re: VM AutoStart

2020-09-30 Thread Jeremey Wise
As the three servers are Centos8 minimal installs. + oVirt HCI wizard to
keep them lean and mean... a couple questions

1) which version of python would I need for this (note in script about
python 2 but isn't that deprecated?)
[root@thor /]# yum install python
Last metadata expiration check: 2:29:38 ago on Wed 30 Sep 2020 01:18:32 PM
EDT.
No match for argument: python
There are following alternatives for "python": python2, python36, python38
Error: Unable to find a match: python

2)  When you have three nodes.. one is set to host the ovirt-engine active,
and another as backup.  If this is added to rc.local.   Of the two nodes
hosting HA for oVirt-engine.. node which boots first will host (or so it
seems). I think if I add this to both those hosts .. it will not create
issues.  Any thoughts?



On Wed, Sep 30, 2020 at 3:23 PM Derek Atkins  wrote:

> I run it out of rc.local:
>
> /usr/local/sbin/start_vms.py > /var/log/start_vms 2>&1 &
>
> The script is smart enough to wait for the engine to be fully active.
>
> -derek
>
> On Wed, September 30, 2020 3:11 pm, Jeremey Wise wrote:
> > i would like to eventually go ansible route..  and was starting down that
> > path but this is fabulous.
> >
> > I will modify and post how it went.
> >
> > One question:  How /where do you set this saved new and delicious script
> > so
> > once oVirt-engine comes up... it runs?
> >
> > Thanks
> >
> > On Wed, Sep 30, 2020 at 2:42 PM Derek Atkins  wrote:
> >
> >> Hi,
> >>
> >> I had a script based around ovirt-shell which I re-wrote as a script
> >> around the Python SDK4 which I run on my engine during the startup
> >> sequence.  The script will wait for the engine to come up and ensure the
> >> storage domains are up before it tries to start the VMs.  Then it will
> >> go
> >> ahead and start the VMs in the specified order with specified delay
> >> and/or
> >> wait-for-up signal between them.
> >>
> >> You can find my scripts at https://www.ihtfp.org/ovirt/
> >>
> >> Or you can go the ansible route :)
> >>
> >> Enjoy!
> >>
> >> -derek
> >>
> >> On Wed, September 30, 2020 11:21 am, Jeremey Wise wrote:
> >> > When I have to shut down cluster... ups runs out etc..  I need a
> >> sequence
> >> > set of just a small number of VMs to "autostart"
> >> >
> >> > Normally I just use DNS FQND to connect to oVirt engine but as two of
> >> my
> >> > VMs  are a DNS HA cluster..  as well as NTP / SMTP /DHCP etc...  I
> >> need
> >> > those two infrastructure VMs to be auto boot.
> >> >
> >> > I looked at HA settings for those VMs but it seems to be watching for
> >> > pause
> >> > /resume.. but it does not imply or state auto start on clean first
> >> boot.
> >> >
> >> > Options?
> >> >
> >> >
> >> > --
> >> > p enguinpages
> >> > ___
> >> > Users mailing list -- users@ovirt.org
> >> > To unsubscribe send an email to users-le...@ovirt.org
> >> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> >> > oVirt Code of Conduct:
> >> > https://www.ovirt.org/community/about/community-guidelines/
> >> > List Archives:
> >> >
> >>
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VAYHFFSANCBRN44ABBTXIYEAR3ZFCP6N/
> >> >
> >>
> >>
> >> --
> >>Derek Atkins 617-623-3745
> >>de...@ihtfp.com www.ihtfp.com
> >>Computer and Internet Security Consultant
> >>
> >>
> >
> > --
> > jeremey.w...@gmail.com
> >
>
>
> --
>Derek Atkins 617-623-3745
>de...@ihtfp.com www.ihtfp.com
>Computer and Internet Security Consultant
>
>

-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RDGSTXC5NEQD2NVRZHG4JP24EQDBRPSM/


[ovirt-users] Re: VM AutoStart

2020-09-30 Thread Jeremey Wise
i would like to eventually go ansible route..  and was starting down that
path but this is fabulous.

I will modify and post how it went.

One question:  How /where do you set this saved new and delicious script so
once oVirt-engine comes up... it runs?

Thanks

On Wed, Sep 30, 2020 at 2:42 PM Derek Atkins  wrote:

> Hi,
>
> I had a script based around ovirt-shell which I re-wrote as a script
> around the Python SDK4 which I run on my engine during the startup
> sequence.  The script will wait for the engine to come up and ensure the
> storage domains are up before it tries to start the VMs.  Then it will go
> ahead and start the VMs in the specified order with specified delay and/or
> wait-for-up signal between them.
>
> You can find my scripts at https://www.ihtfp.org/ovirt/
>
> Or you can go the ansible route :)
>
> Enjoy!
>
> -derek
>
> On Wed, September 30, 2020 11:21 am, Jeremey Wise wrote:
> > When I have to shut down cluster... ups runs out etc..  I need a sequence
> > set of just a small number of VMs to "autostart"
> >
> > Normally I just use DNS FQND to connect to oVirt engine but as two of my
> > VMs  are a DNS HA cluster..  as well as NTP / SMTP /DHCP etc...  I need
> > those two infrastructure VMs to be auto boot.
> >
> > I looked at HA settings for those VMs but it seems to be watching for
> > pause
> > /resume.. but it does not imply or state auto start on clean first boot.
> >
> > Options?
> >
> >
> > --
> > p enguinpages
> > ___
> > Users mailing list -- users@ovirt.org
> > To unsubscribe send an email to users-le...@ovirt.org
> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > oVirt Code of Conduct:
> > https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> >
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/VAYHFFSANCBRN44ABBTXIYEAR3ZFCP6N/
> >
>
>
> --
>Derek Atkins 617-623-3745
>de...@ihtfp.com www.ihtfp.com
>Computer and Internet Security Consultant
>
>

-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XGHXJVGACPIPIZB77KSXRFBF7S6VFEI3/


[ovirt-users] Power on VM - CLI / API

2020-09-30 Thread Jeremey Wise
Can anyone post link.  (with examples.. as most documentation for oVirt
lacks this)..  where I can power on a VM via CLI or API.

As of now I cannot login to oVirt-Engine.  No errors when I restart it..
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/45KKF5TN5PRQ3R7MDOWIQTSYZXZRVDIZ/

BUt... I need to get VMs booted.

I tried to follow:
http://ovirt.github.io/ovirt-engine-api-model/master/
and my server's API web portal
https://ovirte01.penguinpages.local/ovirt-engine/apidoc/#/documents/003_common_concepts

And.. even get POSTMAN (real newbie at that tool but ran into how to add
exported .pem key from portal to session issues)


# failed CLI example:   Power on VM "ns01"

###  DRAFT :: 2020-09-30

# Get key from oVirt engine and import.  Ex: from ovirte01  into server
'thor

curl -k '
https://ovirte01.penguinpages.local/ovirt-engine/services/pki-resource?resource=ca-certificate=X509-PEM-CA'
-o ovirt01_ca.pem

sudo cp ovirt01_ca.pem /etc/pki/ca-trust/source/anchors

sudo update-ca-trust extract



openssl s_client -connect ovirte01.penguinpages.local:443 -showcerts <
/dev/null

 

# Use key during GET list of VMs

?? 

 curl -X POST https://ovirte01.penguinpages.local/post -H
/ovirt-engine/api/vms/ns01/start HTTP/1.1



#

I just need to power on VM


-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/UN4AHVRITGBFUJBYATZA2DTUEIJEX6GL/


[ovirt-users] VM AutoStart

2020-09-30 Thread Jeremey Wise
When I have to shut down cluster... ups runs out etc..  I need a sequence
set of just a small number of VMs to "autostart"

Normally I just use DNS FQND to connect to oVirt engine but as two of my
VMs  are a DNS HA cluster..  as well as NTP / SMTP /DHCP etc...  I need
those two infrastructure VMs to be auto boot.

I looked at HA settings for those VMs but it seems to be watching for pause
/resume.. but it does not imply or state auto start on clean first boot.

Options?


-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VAYHFFSANCBRN44ABBTXIYEAR3ZFCP6N/


[ovirt-users] java.lang.reflect.UndeclaredThrowableException - oVirt engine UI

2020-09-30 Thread Jeremey Wise
I tried to post on website but .. it did not seem to work... so sorry if
this is double posting.

oVirt login this AM. accepted username and password but got java error.

Restarted oVirt engine
##

hosted-engine --set-maintenance --mode=global

hosted-engine --vm-shutdown

hosted-engine --vm-status

#make sure that the status is shutdown before restarting

hosted-engine --vm-start

hosted-engine --vm-status

#make sure the status is health before leaving maintenance mode

hosted-engine --set-maintenance --mode=none
##
[root@thor ~]# hosted-engine --vm-status


--== Host thor.penguinpages.local (id: 1) status ==--

Host ID: 1
Host timestamp : 65342
Score  : 3400
Engine status  : {"vm": "down", "health": "bad",
"detail": "unknown", "reason": "vm not running on this host"}
Hostname   : thor.penguinpages.local
Local maintenance  : False
stopped: False
crc32  : 824c29fd
conf_on_shared_storage : True
local_conf_timestamp   : 65342
Status up-to-date  : True
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=65342 (Wed Sep 30 08:11:45 2020)
host-id=1
score=3400
vm_conf_refresh_time=65342 (Wed Sep 30 08:11:45 2020)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False


--== Host medusa.penguinpages.local (id: 3) status ==--

Host ID: 3
Host timestamp : 87556
Score  : 3400
Engine status  : {"vm": "up", "health": "good",
"detail": "Up"}
Hostname   : medusa.penguinpages.local
Local maintenance  : False
stopped: False
crc32  : 63296a70
conf_on_shared_storage : True
local_conf_timestamp   : 87556
Status up-to-date  : True
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=87556 (Wed Sep 30 08:11:39 2020)
host-id=3
score=3400
vm_conf_refresh_time=87556 (Wed Sep 30 08:11:39 2020)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False
[root@thor ~]# yum update -y
Last metadata expiration check: 0:31:17 ago on Wed 30 Sep 2020 09:17:03 AM
EDT.
Dependencies resolved.
Nothing to do.
Complete!
[root@thor ~]#


Gogled around ..  just found this thread.
##
https://bugzilla.redhat.com/show_bug.cgi?id=1378045


# pgadmin connect to ovirte01.penguinpages.com as engine to db engine
select mac_addr from  vm_interface
"00:16:3e:57:0d:47"
"56:6f:86:41:00:01"
"56:6f:86:41:00:00"
"56:6f:86:41:00:02"
"56:6f:86:41:00:03"
"56:6f:86:41:00:04"
"56:6f:86:41:00:05"
"56:6f:86:41:00:15"

"56:6f:86:41:00:16"
"56:6f:86:41:00:17"
"56:6f:86:41:00:18"
"56:6f:86:41:00:19"


# Note one field is "null"

Question:
1) is this bad?
2) How do I fix?
3) Any idea on root cause?

-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/DJZ6RCDN6UB4VTACKZN6YVISKQGLCWPH/


[ovirt-users] Re: oVirt Change hosts to FQDN

2020-09-29 Thread Jeremey Wise
I think I found field to change host connection from IP to FQDN:

select * from vds_static

change "host_name"

Change IP to fqdn


To test .. I reboot cluster.   Getting a few other errors but ... I think
those are unrelated... will post if / as I learn more.. but so far this
seems to be working.


On Mon, Sep 28, 2020 at 10:56 PM Jeremey Wise 
wrote:

> I used pgadmin   Connected to oVirt-engine VM:
>
> username: engine
> password:  'cat /etc/ovirt-engine/engine.conf.d/10-setup-database.conf
> database: engine
>
> Schemas-> Tables -> 153 tables (which look like what we find in oVirt UI...
>
> Searched around.. no entry where it has 172.16.100.102  or 103  to reflect
> hosts in cluster that I could change to FQDN and restart engine and
>  fixes issue :)
>
> I will keep poking but if someone has done this before it would help.
>
>
>
> On Mon, Sep 28, 2020 at 8:42 PM penguin pages 
> wrote:
>
>> I saw note about holiday.. and I wish all well.   Just kind of stuck here
>> where I am afraid to move forward in building the stack with nodes left in
>> limbo with gluster / cluster.  I just need to repair the host set to
>> connect via IP vs DNS.
>>
>> Any ideas.. or is this a wipe and rebuild of engine again?
>> ___
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/privacy-policy.html
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/SNZ4LEKE5P6LB43BCIXAD5F7RR7IVIB6/
>>
>
>
> --
> jeremey.w...@gmail.com
>


-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/HY3V4GM7OQWBUOYVLFNKUHAPZJNXMGVO/


[ovirt-users] Re: oVirt Change hosts to FQDN

2020-09-28 Thread Jeremey Wise
I used pgadmin   Connected to oVirt-engine VM:

username: engine
password:  'cat /etc/ovirt-engine/engine.conf.d/10-setup-database.conf
database: engine

Schemas-> Tables -> 153 tables (which look like what we find in oVirt UI...

Searched around.. no entry where it has 172.16.100.102  or 103  to reflect
hosts in cluster that I could change to FQDN and restart engine and
 fixes issue :)

I will keep poking but if someone has done this before it would help.



On Mon, Sep 28, 2020 at 8:42 PM penguin pages 
wrote:

> I saw note about holiday.. and I wish all well.   Just kind of stuck here
> where I am afraid to move forward in building the stack with nodes left in
> limbo with gluster / cluster.  I just need to repair the host set to
> connect via IP vs DNS.
>
> Any ideas.. or is this a wipe and rebuild of engine again?
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/SNZ4LEKE5P6LB43BCIXAD5F7RR7IVIB6/
>


-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/YZGHQZ5DDZR6KVVBC7U54LES5NXFFQAB/


[ovirt-users] Re: oVirt Change hosts to FQDN

2020-09-27 Thread Jeremey Wise
when I re deployed ovirt engine after running
ovirt-hosted-engine-cleanup on all nodes.. it deployed oVirt on the first
node fine.. but when I tried to add the other two nodes it kept failing.  I
got it to success ONLY if I used IP vs DNS (post about error here:
https://lists.ovirt.org/archives/list/users@ovirt.org/thread/A6Z3MRFGFSEA7IOCE6WLPEPXE536Z6DR/#A6Z3MRFGFSEA7IOCE6WLPEPXE536Z6DR
)

But this seems to have been a bad idea.  I now need to correct this.

I can't send emails with images so I will post scrape:

##

Name
Comment
Hostname/IP
Cluster
Data Center
Status
Virtual Machines
Memory
CPU
Network
SPM

medusa
medusa host in three node HA cluster
172.16.100.103
Default_Cluster
Default_Datacenter
Up
0
6%
9%
0%
SPM

odin
odin host in three node HA cluster
172.16.100.102
Default_Cluster
Default_Datacenter
Up
1
8%
0%
0%
Normal


thor
thor host in three node HA cluster
thor.penguinpages.local
Default_Cluster
Default_Datacenter
Up
4
9%
2%
0%
Normal
##

[root@thor ~]# gluster pool list
UUIDHostname
 State
83c772aa-33cd-430f-9614-30a99534d10eodinst.penguinpages.local
Connected
977b2c1d-36a8-4852-b953-f75850ac5031medusast.penguinpages.local
Connected
7726b514-e7c3-4705-bbc9-5a90c8a966c9localhost
Connected

[root@thor ~]# gluster peer status
Number of Peers: 2

Hostname: odinst.penguinpages.local
Uuid: 83c772aa-33cd-430f-9614-30a99534d10e
State: Peer in Cluster (Connected)

Hostname: medusast.penguinpages.local
Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
State: Peer in Cluster (Connected)

[root@odin ~]# gluster peer status
Number of Peers: 2

Hostname: thorst.penguinpages.local
Uuid: 7726b514-e7c3-4705-bbc9-5a90c8a966c9
State: Peer in Cluster (Connected)

Hostname: medusast.penguinpages.local
Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
State: Peer in Cluster (Connected)
[root@medusa ~]# gluster peer status
Number of Peers: 2

Hostname: thorst.penguinpages.local
Uuid: 7726b514-e7c3-4705-bbc9-5a90c8a966c9
State: Peer in Cluster (Connected)

Hostname: odinst.penguinpages.local
Uuid: 83c772aa-33cd-430f-9614-30a99534d10e
State: Peer in Cluster (Connected)
[root@medusa ~]#

[root@thor ~]# cat /etc/hosts
# Version: 20190730a
127.0.0.1   localhost localhost.localdomain localhost4
localhost4.localdomain4
::1 localhost localhost.localdomain localhost6
localhost6.localdomain6

# Cluster node thor
172.16.100.91   thorm.penguinpages.local thorm
172.16.100.101  thor.penguinpages.local thor
172.16.101.101  thorst.penguinpages.local thorst

# Cluster node odin
172.16.100.92   odinm.penguinpages.local odinm
172.16.100.102  odin.penguinpages.local odin
172.16.101.102  odinst.penguinpages.local odinst

# Cluster node medusa
# 172.16.100.93   medusam.penguinpages.local medusam
172.16.100.103  medusa.penguinpages.local medusa
172.16.101.103  medusast.penguinpages.local medusast
172.16.100.31 ovirte01.penguinpages.local ovirte01
172.16.100.32 ovirte02.penguinpages.local ovirte02
172.16.100.33 ovirte03.penguinpages.local ovirte03
[root@thor ~]#


On Sun, Sep 27, 2020 at 1:54 PM Strahil Nikolov 
wrote:

> Hi Jeremey,
>
> I am not sure that I completely understand the problem.
>
> Can you provide the Host details page from UI and the output of:
> 'gluster pool list' & 'gluster peer status' from all nodes ?
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В събота, 26 септември 2020 г., 20:31:23 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
>
>
>
>
>
>
> I posted that I had wiped out the oVirt-engine..  running cleanup on all
> three nodes.  Done a re-deployment.   Then to add nodes back.. though all
> have entries for eachother in /etc/hosts and ssh works fine via short and
> long name.
>
> I added nodes back into cluster..  but had to do it via IP to get past
> error.
>
> Now, if I go to create a volume via the GUI in gluster I get:
> Error while executing action Create Gluster Volume: Volume create failed:
> rc=30800 out=() err=["Host 172_16_100_102 is not in 'Peer in Cluster'
> state"]
>
> Which seems to be related to using IP vs DNS to add gluster volumes
> https://bugzilla.redhat.com/show_bug.cgi?id=1055928
>
> Question:  how do i fix the hosts in cluster being defined by IP vs
> desired hostname?
>
>
>
> --
> jeremey.w...@gmail.com
> ___
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/I7Z7UQTQYDSJDZN5AHZCIYQECPUAIE66/
>


-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe sen

[ovirt-users] Re: oVirt Change hosts to FQDN

2020-09-27 Thread Jeremey Wise
Is their any way I can access the host oVirt-engine database and change
field of connection to host from IP to FQDN?

I tried putting server in maintenance mode.. and even remove one (telling
it to ignore gluster) but when I try to remove one node and then hope to
re-add to cluster via FQDN .. (so gluster commands stop failing) it keeps
saying it can't do that to a cluster  gluster node.



On Sat, Sep 26, 2020 at 1:41 PM Jeremey Wise  wrote:

> Another note of color to this.
>
> I can't repair a brick as in gluster it calls bricks by hostname.. and
> oVirt-engine now thinks of it by IP.
>
> Error while executing action Start Gluster Volume Reset Brick: Volume
> reset brick start failed: rc=-1 out=() err=['Pre Validation failed on
> thorst_penguinpages_local_ brick:
> 172_16_100_103:/gluster_bricks/vmstore/vmstore does not exist in volume:
> vmstore\nPre Validation failed on odinst_penguinpages_local_ brick:
> 172_16_100_103:/gluster_bricks/vmstore/vmstore does not exist in volume:
> vmstore']
>
>
>
> On Sat, Sep 26, 2020 at 1:27 PM Jeremey Wise 
> wrote:
>
>>
>> I posted that I had wiped out the oVirt-engine..  running cleanup on all
>> three nodes.  Done a re-deployment.   Then to add nodes back.. though all
>> have entries for eachother in /etc/hosts and ssh works fine via short and
>> long name.
>>
>> I added nodes back into cluster..  but had to do it via IP to get past
>> error.
>>
>> Now, if I go to create a volume via the GUI in gluster I get:
>> Error while executing action Create Gluster Volume: Volume create failed:
>> rc=30800 out=() err=["Host 172_16_100_102 is not in 'Peer in Cluster'
>> state"]
>>
>> Which seems to be related to using IP vs DNS to add gluster volumes
>> https://bugzilla.redhat.com/show_bug.cgi?id=1055928
>>
>> Question:  how do i fix the hosts in cluster being defined by IP vs
>> desired hostname?
>>
>>
>>
>> --
>> jeremey.w...@gmail.com
>>
>
>
> --
> jeremey.w...@gmail.com
>


-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/23TDGJRBRQTEDQAE653447F7SSUZGXOU/


[ovirt-users] Re: oVirt Change hosts to FQDN

2020-09-26 Thread Jeremey Wise
Another note of color to this.

I can't repair a brick as in gluster it calls bricks by hostname.. and
oVirt-engine now thinks of it by IP.

Error while executing action Start Gluster Volume Reset Brick: Volume reset
brick start failed: rc=-1 out=() err=['Pre Validation failed on
thorst_penguinpages_local_ brick:
172_16_100_103:/gluster_bricks/vmstore/vmstore does not exist in volume:
vmstore\nPre Validation failed on odinst_penguinpages_local_ brick:
172_16_100_103:/gluster_bricks/vmstore/vmstore does not exist in volume:
vmstore']



On Sat, Sep 26, 2020 at 1:27 PM Jeremey Wise  wrote:

>
> I posted that I had wiped out the oVirt-engine..  running cleanup on all
> three nodes.  Done a re-deployment.   Then to add nodes back.. though all
> have entries for eachother in /etc/hosts and ssh works fine via short and
> long name.
>
> I added nodes back into cluster..  but had to do it via IP to get past
> error.
>
> Now, if I go to create a volume via the GUI in gluster I get:
> Error while executing action Create Gluster Volume: Volume create failed:
> rc=30800 out=() err=["Host 172_16_100_102 is not in 'Peer in Cluster'
> state"]
>
> Which seems to be related to using IP vs DNS to add gluster volumes
> https://bugzilla.redhat.com/show_bug.cgi?id=1055928
>
> Question:  how do i fix the hosts in cluster being defined by IP vs
> desired hostname?
>
>
>
> --
> jeremey.w...@gmail.com
>


-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ALXIUFZKYTOX7LSW4CBINF4I2Z3WSPFV/


[ovirt-users] oVirt Change hosts to FQDN

2020-09-26 Thread Jeremey Wise
I posted that I had wiped out the oVirt-engine..  running cleanup on all
three nodes.  Done a re-deployment.   Then to add nodes back.. though all
have entries for eachother in /etc/hosts and ssh works fine via short and
long name.

I added nodes back into cluster..  but had to do it via IP to get past
error.

Now, if I go to create a volume via the GUI in gluster I get:
Error while executing action Create Gluster Volume: Volume create failed:
rc=30800 out=() err=["Host 172_16_100_102 is not in 'Peer in Cluster'
state"]

Which seems to be related to using IP vs DNS to add gluster volumes
https://bugzilla.redhat.com/show_bug.cgi?id=1055928

Question:  how do i fix the hosts in cluster being defined by IP vs desired
hostname?



-- 
jeremey.w...@gmail.com
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/I7Z7UQTQYDSJDZN5AHZCIYQECPUAIE66/


[ovirt-users] oVirt - Engine - VM Reconstitute

2020-09-25 Thread Jeremey Wise
As expected... this is a learning curve.  My three node cluster.. in an
attempt to learn how to do admin work on it,  debug it...  I have now
redeployed the engine and even added second one on a node in cluster.

But.

I now realize that my "production vms" are gone.

In the past, on a manual build with KVM + Gluster .. when I repaired a
damaged cluster I would just then browse to the xml file and import.

I think with oVirt, those days are gone.   As the PostGres engine knows the
links to the disk / thin provision volumes / network / VM definition files.

##

Question:
1) Can someone point me to the manual on how to re-constitute a VM and
bring it back into oVirt where all "oVirt-engines" were redeployed.  It is
only three or four VMs I typically care about (HA cluster and OCP ignition/
Ansible tower VM).
2) How do I make sure these core VMs are able to be reconstituted.  Can I
create a dedicated volume where the VMs are full provisioned, and the path
structure is "human understandable".
3) I know that you can backup the engine.  If I had been a smart person,
how does one backup and recover from this kind of situation.Does anyone
have any guides or good articles on this?

Thanks,



-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/74EYPATSQO46GX2KRFYNWD53ZWJ3JYSQ/


[ovirt-users] oVirt host "unregistered"

2020-09-24 Thread Jeremey Wise
Trying to get all the 3 node cluster back fully working... clearing out all
the errors.

I noted that the HCI wizard..  I thing SHOULD have deployed a hosted engine
on the nodes, but this is not the case .. Only thor... the first node in
cluster has hosted engine.

I tried to redeploy this via the Cockpit wizard to add engine to host.. but
I think this may not have been correct repair path.

Now node in cluster shows all bricks... green (so it detects after reboot
that host is back up and working.. but hosts lists it as "red triangle"
with error "unregistered"

I also just tried on third node to "edit" -> Host Engine -> and click drop
box and choose "deploy"

only log in event is "Host medusa configuration was updated by
admin@internal-authz.
9/24/208:49:24 PM"but nothing changes.

I then ran on odin (node with error)  ovirt-hosted-engine-cleanup but
no change

[root@odin ~]# ovirt-hosted-engine-cleanup
 This will de-configure the host to run ovirt-hosted-engine-setup from
scratch.
Caution, this operation should be used with care.

Are you sure you want to proceed? [y/n]
y
  -=== Destroy hosted-engine VM ===-
You must run deploy first
error: failed to get domain 'HostedEngine'

  -=== Stop HA services ===-
  -=== Shutdown sanlock ===-
shutdown force 1 wait 0
shutdown done 0
  -=== Disconnecting the hosted-engine storage domain ===-
You must run deploy first
  -=== De-configure VDSM networks ===-
ovirtmgmt
ovirtmgmt
 A previously configured management bridge has been found on the system,
this will try to de-configure it. Under certain circumstances you can loose
network connection.
Caution, this operation should be used with care.

Are you sure you want to proceed? [y/n]
y
  -=== Stop other services ===-
Warning: Stopping libvirtd.service, but it can still be activated by:
  libvirtd-ro.socket
  libvirtd.socket
  libvirtd-admin.socket
  -=== De-configure external daemons ===-
Removing database file /var/lib/vdsm/storage/managedvolume.db
  -=== Removing configuration files ===-
? /etc/init/libvirtd.conf already missing
- removing /etc/libvirt/nwfilter/vdsm-no-mac-spoofing.xml
? /etc/ovirt-hosted-engine/answers.conf already missing
? /etc/ovirt-hosted-engine/hosted-engine.conf already missing
- removing /etc/vdsm/vdsm.conf
- removing /etc/pki/vdsm/certs/cacert.pem
- removing /etc/pki/vdsm/certs/vdsmcert.pem
- removing /etc/pki/vdsm/keys/vdsmkey.pem
- removing /etc/pki/vdsm/libvirt-migrate/ca-cert.pem
- removing /etc/pki/vdsm/libvirt-migrate/server-cert.pem
- removing /etc/pki/vdsm/libvirt-migrate/server-key.pem
- removing /etc/pki/vdsm/libvirt-spice/ca-cert.pem
- removing /etc/pki/vdsm/libvirt-spice/server-cert.pem
- removing /etc/pki/vdsm/libvirt-spice/server-key.pem
- removing /etc/pki/vdsm/libvirt-vnc/ca-cert.pem
- removing /etc/pki/vdsm/libvirt-vnc/server-cert.pem
- removing /etc/pki/vdsm/libvirt-vnc/server-key.pem
- removing /etc/pki/CA/cacert.pem
- removing /etc/pki/libvirt/clientcert.pem
- removing /etc/pki/libvirt/private/clientkey.pem
? /etc/pki/ovirt-vmconsole/*.pem already missing
- removing /var/cache/libvirt/qemu
? /var/run/ovirt-hosted-engine-ha/* already missing
- removing /var/tmp/localvm69i1jxnd
- removing /var/tmp/localvmfyg59713
- removing /var/tmp/localvmmg5y6g52
  -=== Removing IP Rules ===-
[root@odin ~]#
[root@odin ~]#



Ideas on how to repair engine install issues on nodes?

-- 
penguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/A6Z3MRFGFSEA7IOCE6WLPEPXE536Z6DR/


[ovirt-users] Restart oVirt-Engine

2020-09-24 Thread Jeremey Wise
How ,without reboot of hosting system, do I restart the oVirt engine?

# I tried below but do not seem to effect the virtual machine
[root@thor iso]# systemctl restart ov
ovirt-ha-agent.service  ovirt-imageio.service
ovn-controller.service  ovs-delete-transient-ports.service
ovirt-ha-broker.service ovirt-vmconsole-host-sshd.service
ovsdb-server.serviceovs-vswitchd.service
[root@thor iso]#

# You cannot restart the VM " HostedEngine " as it responses:

Error while executing action:

HostedEngine:

   - Cannot restart VM. This VM is not managed by the engine.



Reason is I had to do some work on a node.  Reboot it..  it is back up..
network is all fine..  Cockpit working fine...  and gluster fine..  But
oVirt-Engine refuses to accept the node is up.


-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/EODP5D76JD6UB4FHMESIQZMO7XXE77CU/


[ovirt-users] Host has to be reinstalled

2020-09-23 Thread Jeremey Wise
Trying to repair / clean up HCI deployment so it is HA and ready for
"production".

I have gluster now showing three bricks  all green

Now I just have error on node.. and of course the node which is hosting the
ovirt-engine

# (as I can not send images to this forum... I will move to a breadcrumb
posting)
Compute -> Hosts -> "thor" (red exclamation)
"Host has to be reinstalled"

To fix gluster... i had to reinstall "vdsm-gluster"

But what package does this error need to be reviewed / fixed with?


-- 
penguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/E5BILIQ5L4OXQA7AEDTGBVQH6XSHP4M4/


[ovirt-users] Gluster Volumes - Correct Peer Connection

2020-09-23 Thread Jeremey Wise
I just noticed when HCI setup bult the gluster engine / data / vmstore
volumes... it did use correctly the definition of 10Gb "back end"
interfaces / hosts.

But..  oVirt Engine is NOT referencing this.
it lists bricks as 1Gb "managment / host" interfaces.  Is this a GUI
issue?  I doubt this and how do I correct it?
###  Data Volume Example
Name:
data
Volume ID:
0ae7b487-8b87-4192-bd30-621d445902fe
Volume Type:
Replicate
Replica Count:
3
Number of Bricks:
3
Transport Types:
TCP
Maximum no of snapshots:
256
Capacity:
999.51 GiB total, 269.02 GiB used, 730.49 GiB free, 297.91 GiB Guaranteed
free, 78 Deduplication/Compression savings (%)


medusa.penguinpages.local
medusa.penguinpages.local:/gluster_bricks/data/data
25%
OK
odin.penguinpages.local
odin.penguinpages.local:/gluster_bricks/data/data
25%
OK
thor.penguinpages.local
thor.penguinpages.local:/gluster_bricks/data/data
25%
OK


# I have storage back end of 172.16.101.x  which is 10Gb dedicated for
replication. Peers reflect this
[root@odin c4918f28-00ce-49f9-91c8-224796a158b9]# gluster peer status
Number of Peers: 2

Hostname: thorst.penguinpages.local
Uuid: 7726b514-e7c3-4705-bbc9-5a90c8a966c9
State: Peer in Cluster (Connected)

Hostname: medusast.penguinpages.local
Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
State: Peer in Cluster (Connected)
[root@odin c4918f28-00ce-49f9-91c8-224796a158b9]#



-- 
p enguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ULE66KK5UEGM5GTRG6IMWZLUEI6JLHVI/


[ovirt-users] ISO Repo

2020-09-23 Thread Jeremey Wise
I saw notes about oVirt 4.4 may no longer support ISO images... but there
are times like now I need to build based on specific ISO images.

I tried to do a cycle to create an image file 8GB  then do dd if=blah.iso
of=/

Created a new vm with this as boot disk and it fails to boot...  so.. back
to "create volume for iso images"

But when I do that I get error

New Domain -> "Domain function" =iso  Storage type = glusterFs
Use Managed Gluster volume -> Select already working gluster file space
"thor.penguinpages.local:/iso
VFS Type: glusterfs
mount
options: 
backup-volfile-servers=odin.penguinpages.local:medusa.penguinpages.local


Error:
Error while executing action: Cannot add Storage Connection. Performance
o_direct option is not enabled for storage domain.

Questions:
1) Why did the image and dd of iso to boot disk not work?
2) Any ideas about create of iso mount volume?

-- 
penguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WU2QJ3UYO7P2RSWAWEOXLYHGI2LMWKID/


[ovirt-users] Re: oVirt - Gluster Node Offline but Bricks Active

2020-09-23 Thread Jeremey Wise
in oVirt Engine I think I see some of the issue

When you go under volumes -> Data ->

[image: image.png]

It notes two servers..  when you choose "add brick" it says volume has 3
bricks but only two servers.

So I went back to my deployment notes and walked through setup

yum install https://resources.ovirt.org/pub/yum-repo/ovirt-release44.rpm -y

yum install -y cockpit-ovirt-dashboard vdsm-gluster ovirt-host

Last metadata expiration check: 1:59:46 ago on Wed 23 Sep 2020 06:10:46 AM
EDT.
Package cockpit-ovirt-dashboard-0.14.11-1.el8.noarch is already installed.
Package ovirt-host-4.4.1-4.el8.x86_64 is already installed.
Dependencies resolved.
=
 Package
 ArchitectureVersion
  Repository
   Size
=
Installing:
 vdsm-gluster   x86_64
 4.40.26.3-1.el8
  ovirt-4.4
  67 k
Installing dependencies:
 blivet-datanoarch
 1:3.1.0-21.el8_2
 AppStream
 238 k
 glusterfs-events   x86_64
 7.7-1.el8
  ovirt-4.4-centos-gluster7
  65 k
 glusterfs-geo-replication  x86_64
 7.7-1.el8
  ovirt-4.4-centos-gluster7
 212 k
 libblockdev-plugins-allx86_64
 2.19-12.el8
  AppStream
  62 k
 libblockdev-vdox86_64
 2.19-12.el8
  AppStream
  74 k
 python3-blivet noarch
 1:3.1.0-21.el8_2
 AppStream
 995 k
 python3-blockdev   x86_64
 2.19-12.el8
  AppStream
  79 k
 python3-bytesize   x86_64
 1.4-3.el8
  AppStream
  28 k
 python3-magic  noarch
 5.33-13.el8
  BaseOS
   45 k
 python3-pyparted   x86_64
 1:3.11.0-13.el8
  AppStream
 123 k


Dependencies resolved.
Nothing to do.
Complete!
[root@thor media]#



AKA.. something got removed from the node..


Rebooted.. as I am not sure which dependancies and services would need to
be restarted to get oVirt-engine to pick things up.


Host is now "green" .. now only errors are about gluster bricks..



On Tue, Sep 22, 2020 at 9:30 PM penguin pages 
wrote:

>
>
> eMail client with this forum is a bit .. I was told this web
> interface I could post images... as embedded ones in email get scraped
> out...  but not seeing how that is done. Seems to be txt only.
>
>
>
> 1) ..."I would give the engine a 'Windows'-style fix (a.k.a. reboot)"
> how does one restart just the oVirt-engine?
>
> 2) I now show in shell  3 nodes, each with the one brick for data,
> vmstore, engine (and an ISO one I am trying to make).. with one brick each
> and all online and replicating.   But the GUI shows thor (first server
> running engine) offline needing to be reloaded.  Now volumes show two
> bricks.. one online one offline.  And no option to start / force restart.
>
> 3) I have tried several times to try a graceful reboot to see if startup
> sequence was issue.   I tore down VLANs and bridges to make it flat 1 x 1Gb
> mgmt, 1 x 10Gb storage.   SSH between nodes is fine... copy test was
> great.   I don't think it is nodes.
>
> 4) To the question of "did I add third node later."  I would attach
> deployment guide I am building ... but can't do that in this forum.  but
> this is as simple as I can make it.  3 intel generic servers,  1 x boot
> drive , 1 x 512GB SSD,  2 x 1TB SSD in each.   wipe all data all
> configuration fresh Centos8 minimal install.. setup SSH setup basic
> networking... install cockpit.. run HCI wizard for all three nodes. That is
> all.
>
> Trying to learn and support concept of oVirt as a viable platform but
> still trying to work through learning how to root cause, kick tires, and
> debug / recover when things go down .. as they will.
>
> Help is appreciated.  The main concern I have is gap in what engine sees
> 

[ovirt-users] Re: oVirt - vdo: ERROR - Device /dev/sd excluded by a filter

2020-09-22 Thread Jeremey Wise
Correct..  on wwid


I do want to make clear here.  that to geta around the error you must ADD
(not remove ) drives to /etc/lvm/lvm.conf  so oVirt Gluster can complete
setup of drives.

[root@thor log]# cat /etc/lvm/lvm.conf |grep filter
# Broken for gluster in oVirt
#filter =
["a|^/dev/disk/by-id/lvm-pv-uuid-AAHPao-R62q-8aac-410x-ZdA7-UL4i-Bh2bwJ$|",
"a|^/dev/disk/by-id/lvm-pv-uuid-bSnFU3-jtUj-AGds-07sw-zdYC-52fM-mujuvC$|",
"r|.*|"]
# working for gluster wizard in oVirt
filter =
["a|^/dev/disk/by-id/lvm-pv-uuid-AAHPao-R62q-8aac-410x-ZdA7-UL4i-Bh2bwJ$|",
"a|^/dev/disk/by-id/lvm-pv-uuid-bSnFU3-jtUj-AGds-07sw-zdYC-52fM-mujuvC$|",
"a|^/dev/disk/by-id/wwn-0x5001b448b847be41$|", "r|.*|"]



On Tue, Sep 22, 2020 at 3:57 PM Strahil Nikolov 
wrote:

> Obtaining the wwid is not exactly correct.
> You can identify them via:
>
> multipath -v4 | grep 'got wwid of'
>
> Short example:
> [root@ovirt conf.d]# multipath -v4 | grep 'got wwid of'
> Sep 22 22:55:58 | nvme0n1: got wwid of
> 'nvme.1cc1-324a31313230303131343036-414441544120535838323030504e50-0001'
> Sep 22 22:55:58 | sda: got wwid of 'TOSHIBA-TR200_Z7KB600SK46S'
> Sep 22 22:55:58 | sdb: got wwid of 'ST500NM0011_Z1M00LM7'
> Sep 22 22:55:58 | sdc: got wwid of 'WDC_WD5003ABYX-01WERA0_WD-WMAYP2303189'
> Sep 22 22:55:58 | sdd: got wwid of 'WDC_WD15EADS-00P8B0_WD-WMAVU0115133'
>
> Of course if you are planing to use only gluster it could be far easier to
> set:
>
> [root@ovirt conf.d]# cat /etc/multipath/conf.d/blacklist.conf
> blacklist {
> devnode "*"
> }
>
>
>
> Best Regards,
> Strahil Nikolov
>
> В вторник, 22 септември 2020 г., 22:12:21 Гринуич+3, Nir Soffer <
> nsof...@redhat.com> написа:
>
>
>
>
>
> On Tue, Sep 22, 2020 at 1:50 AM Jeremey Wise 
> wrote:
> >
> >
> > Agree about an NVMe Card being put under mpath control.
>
> NVMe can be used via multipath, this is a new feature added in RHEL 8.1:
> https://bugzilla.redhat.com/1498546
>
> Of course when the NVMe device is local there is no point to use it
> via multipath.
> To avoid this, you need to blacklist the devices like this:
>
> 1. Find the device wwid
>
> For NVMe, you need the device ID_WWN:
>
> $ udevadm info -q property /dev/nvme0n1 | grep ID_WWN
> ID_WWN=eui.5cd2e42a81a11f69
>
> 2. Add local blacklist file:
>
> $ mkdir /etc/multipath/conf.d
> $ cat /etc/multipath/conf.d/local.conf
> blacklist {
> wwid "eui.5cd2e42a81a11f69"
> }
>
> 3. Reconfigure multipath
>
> $ multipathd reconfigure
>
> Gluster should do this for you automatically during installation, but
> it does not
> you can do this manually.
>
> > I have not even gotten to that volume / issue.  My guess is something
> weird in CentOS / 4.18.0-193.19.1.el8_2.x86_64  kernel with NVMe block
> devices.
> >
> > I will post once I cross bridge of getting standard SSD volumes working
> >
> > On Mon, Sep 21, 2020 at 4:12 PM Strahil Nikolov 
> wrote:
> >>
> >> Why is your NVME under multipath ? That doesn't make sense at all .
> >> I have modified my multipath.conf to block all local disks . Also
> ,don't forget the '# VDSM PRIVATE' line somewhere in the top of the file.
> >>
> >> Best Regards,
> >> Strahil Nikolov
> >>
> >>
> >>
> >>
> >>
> >>
> >> В понеделник, 21 септември 2020 г., 09:04:28 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> vdo: ERROR - Device /dev/sdc excluded by a filter
> >>
> >>
> >>
> >>
> >> Other server
> >> vdo: ERROR - Device
> /dev/mapper/nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001p1
> excluded by a filter.
> >>
> >>
> >> All systems when I go to create VDO volume on blank drives.. I get this
> filter error.  All disk outside of the HCI wizard setup are now blocked
> from creating new Gluster volume group.
> >>
> >> Here is what I see in /dev/lvm/lvm.conf |grep filter
> >> [root@odin ~]# cat /etc/lvm/lvm.conf |grep filter
> >> filter =
> ["a|^/dev/disk/by-id/lvm-pv-uuid-e1fvwo-kEfX-v3lT-SKBp-cgze-TwsO-PtyvmC$|",
> "a|^/dev/disk/by-id/lvm-pv-uuid-mr9awW-oQH5-F4IX-CbEO-RgJZ-x4jK-e4YZS1$|",
> "r|.*|"]
> >>
> >> [root@odin ~]# ls -al /dev/disk/by-id/
> >> total 0
> >> drwxr-xr-x. 2 root root 12

[ovirt-users] Re: oVirt - Gluster Node Offline but Bricks Active

2020-09-22 Thread Jeremey Wise
when I posted last..  in the tread I paste a roling restart.And...  now
it is replicating.

oVirt still showing wrong.  BUT..   I did my normal test from each of the
three nodes.

1) Mount Gluster file system with localhost as primary and other two as
tertiary to local mount (like a client would do)
2) run test file create Ex:   echo $HOSTNAME >>
/media/glustervolume/test.out
3) repeat from each node then read back that all are in sync.

I REALLY hate reboot (restart) as a fix.  I need to get better with root
cause of gluster issues if I am going to trust it.  Before when I manually
made the volumes and it was simply (vdo + gluster) then worst case was that
gluster would break... but I could always go into "brick" path and copy
data out.

Now with oVirt.. .and LVM and thin provisioning etc..   I am abstracted
from simple file recovery..  Without GLUSTER AND oVirt Engine up... all my
environment  and data is lost.  This means nodes moved more to "pets" then
cattle.

And with three nodes.. I can't afford to loose any pets.

I will post more when I get cluster settled and work on those wierd notes
about quorum volumes noted on two nodes when glusterd is restarted.

Thanks,

On Tue, Sep 22, 2020 at 8:44 AM Strahil Nikolov 
wrote:

> Replication issue could mean that one of the client (FUSE mounts) is not
> attached to all bricks.
>
> You can check the amount of clients via:
> gluster volume status all client-list
>
>
> As a prevention , just do a rolling restart:
> - set a host in maintenance and mark it to stop glusterd service (I'm
> reffering to the UI)
> - Activate the host , once it was moved to maintenance
>
> Wait for the host's HE score to recover (silver/gold crown in UI) and then
> proceed with the next one.
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
> В вторник, 22 септември 2020 г., 14:55:35 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
>
>
>
>
>
>
> I did.
>
> Here are all three nodes with restart. I find it odd ... their has been a
> set of messages at end (see below) which I don't know enough about what
> oVirt laid out to know if it is bad.
>
> ###
> [root@thor vmstore]# systemctl status glusterd
> ● glusterd.service - GlusterFS, a clustered file-system server
>Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
> vendor preset: disabled)
>   Drop-In: /etc/systemd/system/glusterd.service.d
>└─99-cpu.conf
>Active: active (running) since Mon 2020-09-21 20:32:26 EDT; 10h ago
>  Docs: man:glusterd(8)
>   Process: 2001 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
> --log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
>  Main PID: 2113 (glusterd)
> Tasks: 151 (limit: 1235410)
>Memory: 3.8G
>   CPU: 6min 46.050s
>CGroup: /glusterfs.slice/glusterd.service
>├─ 2113 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
> INFO
>├─ 2914 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
> -p /var/run/gluster/shd/data/data-shd.pid -l
> /var/log/glusterfs/glustershd.log -S
> /var/run/gluster/2f41374c2e36bf4d.socket --xlator-option
> *replicate*.node-uu>
>├─ 9342 /usr/sbin/glusterfsd -s thorst.penguinpages.local
> --volfile-id data.thorst.penguinpages.local.gluster_bricks-data-data -p
> /var/run/gluster/vols/data/thorst.penguinpages.local-gluster_bricks-data-data.pid
> -S /var/r>
>├─ 9433 /usr/sbin/glusterfsd -s thorst.penguinpages.local
> --volfile-id engine.thorst.penguinpages.local.gluster_bricks-engine-engine
> -p
> /var/run/gluster/vols/engine/thorst.penguinpages.local-gluster_bricks-engine-engine.p>
>├─ 9444 /usr/sbin/glusterfsd -s thorst.penguinpages.local
> --volfile-id
> vmstore.thorst.penguinpages.local.gluster_bricks-vmstore-vmstore -p
> /var/run/gluster/vols/vmstore/thorst.penguinpages.local-gluster_bricks-vmstore-vms>
>└─35639 /usr/sbin/glusterfsd -s thorst.penguinpages.local
> --volfile-id iso.thorst.penguinpages.local.gluster_bricks-iso-iso -p
> /var/run/gluster/vols/iso/thorst.penguinpages.local-gluster_bricks-iso-iso.pid
> -S /var/run/glu>
>
> Sep 21 20:32:24 thor.penguinpages.local systemd[1]: Starting GlusterFS, a
> clustered file-system server...
> Sep 21 20:32:26 thor.penguinpages.local systemd[1]: Started GlusterFS, a
> clustered file-system server.
> Sep 21 20:32:28 thor.penguinpages.local glusterd[2113]: [2020-09-22
> 00:32:28.605674] C [MSGID: 106003]
> [glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
> 0-management: Server quorum regained for volume data. Starting lo>
> Sep 21 20:32:28 thor.penguinpages.local glusterd[2113]: [2020-09-22
> 00:32:28.639490] C [MSGID: 106003]
> [

[ovirt-users] Re: oVirt - Gluster Node Offline but Bricks Active

2020-09-22 Thread Jeremey Wise
6.101.101:24007 has not responded in the
last 30 seconds, di>
Sep 21 20:33:24 medusa.penguinpages.local glustershd[7863]: [2020-09-22
00:33:24.821774] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
4-vmstore-client-0: server 172.16.101.101:24007 has not responded in the
last 30 seconds, d>
Sep 21 20:33:36 medusa.penguinpages.local glustershd[7863]: [2020-09-22
00:33:36.819762] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
3-iso-client-0: server 172.16.101.101:24007 has not responded in the last
42 seconds, disco>
[root@medusa sw2_usb_A2]# systemctl restart glusterd
[root@medusa sw2_usb_A2]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
   └─99-cpu.conf
   Active: active (running) since Tue 2020-09-22 07:51:46 EDT; 2s ago
 Docs: man:glusterd(8)
  Process: 80099 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 80100 (glusterd)
Tasks: 146 (limit: 409064)
   Memory: 207.7M
  CPU: 2.705s
   CGroup: /glusterfs.slice/glusterd.service
   ├─ 7819 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id data.medusast.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/medusast.penguinpages.local-gluster_bricks-data-data.pid
-S >
   ├─ 7830 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id
engine.medusast.penguinpages.local.gluster_bricks-engine-engine -p
/var/run/gluster/vols/engine/medusast.penguinpages.local-gluster_bricks-engine-en>
   ├─ 7841 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id
vmstore.medusast.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/medusast.penguinpages.local-gluster_bricks-vmsto>
   ├─21228 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id iso.medusast.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/medusast.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/r>
   ├─80100 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO
   └─80152 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
-p /var/run/gluster/shd/data/data-shd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/gluster/709d753e1e04185a.socket --xlator-option
*replicate*.node-uu>

Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: glusterd.service:
Found left-over process 7863 (glusterfs) in control group while starting
unit. Ignoring.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: glusterd.service:
Found left-over process 21228 (glusterfsd) in control group while starting
unit. Ignoring.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: Starting GlusterFS, a
clustered file-system server...
Sep 22 07:51:46 medusa.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
Sep 22 07:51:46 medusa.penguinpages.local glusterd[80100]: [2020-09-22
11:51:46.789628] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume data. Starting>
Sep 22 07:51:46 medusa.penguinpages.local glusterd[80100]: [2020-09-22
11:51:46.807618] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume engine. Starti>
Sep 22 07:51:46 medusa.penguinpages.local glusterd[80100]: [2020-09-22
11:51:46.825589] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume vmstore. Start>
[root@medusa sw2_usb_A2]# ls /gluster_bricks/vmstore/vmstore/
example.log  f118dcae-6162-4e9a-89e4-f30ffcfb9ccf  isos  media
 ns01_20200910.tgz  ns02_20200910.tgz  ns02.qcow2  ns02_var.qcow2  qemu


As for files... there is replication issues.  Not really sure how bricks
show ok but it is not replicating



On Tue, Sep 22, 2020 at 2:38 AM Strahil Nikolov 
wrote:

> Have you restarted glusterd.service on the affected node.
> glusterd is just management layer and it won't affect the brick processes.
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В вторник, 22 септември 2020 г., 01:43:36 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
>
>
>
>
>
>
> Start is not an option.
>
&

[ovirt-users] Re: oVirt - KVM QCow2 Import

2020-09-21 Thread Jeremey Wise
Well.. to know how to do it with Curl is helpful.. but I think I did

[root@odin ~]#  curl -s -k --user admin@internal:blahblah
https://ovirte01.penguinpages.local/ovirt-engine/api/storagedomains/ |grep
''
data
hosted_storage
ovirt-image-repository

What I guess I did is translated that field --sd-name my-storage-domain \
  to " volume" name... My question is .. where do those fields come from?
And which would you typically place all your VMs into?
[image: image.png]



I just took a guess..  and figured "data" sounded like a good place to
stick raw images to build into VM...

[root@medusa thorst.penguinpages.local:_vmstore]# python3
/usr/share/doc/python3-ovirt-engine-sdk4/examples/upload_disk.py
--engine-url https://ovirte01.penguinpages.local/ --username admin@internal
--password-file
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/.ovirt.password
--cafile
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/.ovirte01_pki-resource.cer
--sd-name data --disk-sparse
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/ns02.qcow2
Checking image...
Image format: qcow2
Disk format: cow
Disk content type: data
Disk provisioned size: 21474836480
Disk initial size: 11574706176
Disk name: ns02.qcow2
Disk backup: False
Connecting...
Creating disk...
Disk ID: 9ccb26cf-dd4a-4c9a-830c-ee084074d7a1
Creating image transfer...
Transfer ID: 3a382f0b-1e7d-4397-ab16-4def0e9fe890
Transfer host name: medusa
Uploading image...
[ 100.00% ] 20.00 GiB, 249.86 seconds, 81.97 MiB/s
Finalizing image transfer...
Upload completed successfully
[root@medusa thorst.penguinpages.local:_vmstore]# python3
/usr/share/doc/python3-ovirt-engine-sdk4/examples/upload_disk.py
--engine-url https://ovirte01.penguinpages.local/ --username admin@internal
--password-file
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/.ovirt.password
--cafile
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/.ovirte01_pki-resource.cer
--sd-name data --disk-sparse
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/ns02_v^C
[root@medusa thorst.penguinpages.local:_vmstore]# ls
example.log  f118dcae-6162-4e9a-89e4-f30ffcfb9ccf  ns02_20200910.tgz
 ns02.qcow2  ns02_var.qcow2
[root@medusa thorst.penguinpages.local:_vmstore]# python3
/usr/share/doc/python3-ovirt-engine-sdk4/examples/upload_disk.py
--engine-url https://ovirte01.penguinpages.local/ --username admin@internal
--password-file
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/.ovirt.password
--cafile
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/.ovirte01_pki-resource.cer
--sd-name data --disk-sparse
/rhev/data-center/mnt/glusterSD/thorst.penguinpages.local:_vmstore/ns02_var.qcow2
Checking image...
Image format: qcow2
Disk format: cow
Disk content type: data
Disk provisioned size: 107374182400
Disk initial size: 107390828544
Disk name: ns02_var.qcow2
Disk backup: False
Connecting...
Creating disk...
Disk ID: 26def4e7-1153-417c-88c1-fd3dfe2b0fb9
Creating image transfer...
Transfer ID: 41518eac-8881-453e-acc0-45391fd23bc7
Transfer host name: medusa
Uploading image...
[  16.50% ] 16.50 GiB, 556.42 seconds, 30.37 MiB/s

Now with those ID numbers and that it kept its name (very helpful)... I am
able to re-constitute the VM
[image: image.png]

VM boots fine.  Fixing VLANs and manual macs on vNICs.. but this process
worked fine.

Thanks for input.   Would be nice to have a GUI "upload" via http into
system :)







On Mon, Sep 21, 2020 at 2:19 PM Nir Soffer  wrote:

> On Mon, Sep 21, 2020 at 8:37 PM penguin pages 
> wrote:
> >
> >
> > I pasted old / file path not right example above.. But here is a cleaner
> version with error i am trying to root cause
> >
> > [root@odin vmstore]# python3
> /usr/share/doc/python3-ovirt-engine-sdk4/examples/upload_disk.py
> --engine-url https://ovirte01.penguinpages.local/ --username
> admin@internal --password-file
> /gluster_bricks/vmstore/vmstore/.ovirt.password --cafile
> /gluster_bricks/vmstore/vmstore/.ovirte01_pki-resource.cer --sd-name
> vmstore --disk-sparse /gluster_bricks/vmstore/vmstore/ns01.qcow2
> > Checking image...
> > Image format: qcow2
> > Disk format: cow
> > Disk content type: data
> > Disk provisioned size: 21474836480
> > Disk initial size: 431751168
> > Disk name: ns01.qcow2
> > Disk backup: False
> > Connecting...
> > Creating disk...
> > Traceback (most recent call last):
> >   File
> "/usr/share/doc/python3-ovirt-engine-sdk4/examples/upload_disk.py", line
> 262, in 
> > name=args.sd_name
> >   File "/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py", line
> 7697, in add
> > return self._internal_add(disk, headers, query, wait)
> >   File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line
> 232, in _internal_add
> > return future.wait() if wait else future
> >   File "/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py", line
> 55, in wait
> > return 

[ovirt-users] Re: oVirt - vdo: ERROR - Device /dev/sd excluded by a filter

2020-09-21 Thread Jeremey Wise
Agree about an NVMe Card being put under mpath control.

I have not even gotten to that volume / issue.   My guess is something
weird in CentOS / 4.18.0-193.19.1.el8_2.x86_64  kernel with NVMe block
devices.

I will post once I cross bridge of getting standard SSD volumes working

On Mon, Sep 21, 2020 at 4:12 PM Strahil Nikolov 
wrote:

> Why is your NVME under multipath ? That doesn't make sense at all .
> I have modified my multipath.conf to block all local disks . Also ,don't
> forget the '# VDSM PRIVATE' line somewhere in the top of the file.
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В понеделник, 21 септември 2020 г., 09:04:28 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
>
>
>
>
>
>
>
>
>
>
> vdo: ERROR - Device /dev/sdc excluded by a filter
>
>
>
>
> Other server
> vdo: ERROR - Device
> /dev/mapper/nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001p1
> excluded by a filter.
>
>
> All systems when I go to create VDO volume on blank drives.. I get this
> filter error.  All disk outside of the HCI wizard setup are now blocked
> from creating new Gluster volume group.
>
> Here is what I see in /dev/lvm/lvm.conf |grep filter
> [root@odin ~]# cat /etc/lvm/lvm.conf |grep filter
> filter =
> ["a|^/dev/disk/by-id/lvm-pv-uuid-e1fvwo-kEfX-v3lT-SKBp-cgze-TwsO-PtyvmC$|",
> "a|^/dev/disk/by-id/lvm-pv-uuid-mr9awW-oQH5-F4IX-CbEO-RgJZ-x4jK-e4YZS1$|",
> "r|.*|"]
>
> [root@odin ~]# ls -al /dev/disk/by-id/
> total 0
> drwxr-xr-x. 2 root root 1220 Sep 18 14:32 .
> drwxr-xr-x. 6 root root  120 Sep 18 14:32 ..
> lrwxrwxrwx. 1 root root9 Sep 18 22:40
> ata-INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN -> ../../sda
> lrwxrwxrwx. 1 root root   10 Sep 18 22:40
> ata-INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN-part1 -> ../../sda1
> lrwxrwxrwx. 1 root root   10 Sep 18 22:40
> ata-INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN-part2 -> ../../sda2
> lrwxrwxrwx. 1 root root9 Sep 18 14:32
> ata-Micron_1100_MTFDDAV512TBN_17401F699137 -> ../../sdb
> lrwxrwxrwx. 1 root root9 Sep 18 22:40
> ata-WDC_WDS100T2B0B-00YS70_183533804564 -> ../../sdc
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40 dm-name-cl-home -> ../../dm-2
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40 dm-name-cl-root -> ../../dm-0
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40 dm-name-cl-swap -> ../../dm-1
> lrwxrwxrwx. 1 root root   11 Sep 18 16:40
> dm-name-gluster_vg_sdb-gluster_lv_data -> ../../dm-11
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40
> dm-name-gluster_vg_sdb-gluster_lv_engine -> ../../dm-6
> lrwxrwxrwx. 1 root root   11 Sep 18 16:40
> dm-name-gluster_vg_sdb-gluster_lv_vmstore -> ../../dm-12
> lrwxrwxrwx. 1 root root   10 Sep 18 23:35
> dm-name-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
> -> ../../dm-3
> lrwxrwxrwx. 1 root root   10 Sep 18 23:49
> dm-name-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001p1
> -> ../../dm-4
> lrwxrwxrwx. 1 root root   10 Sep 18 14:32 dm-name-vdo_sdb -> ../../dm-5
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40
> dm-uuid-LVM-GpvYIuypEfrR7nEDn5uHPenKwjrsn4ADc49gc6PWLRBCoJ2B3JC9tDJejyx5eDPT
> -> ../../dm-1
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40
> dm-uuid-LVM-GpvYIuypEfrR7nEDn5uHPenKwjrsn4ADOMNJfgcat9ZLOpcNO7FyG8ixcl5s93TU
> -> ../../dm-2
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40
> dm-uuid-LVM-GpvYIuypEfrR7nEDn5uHPenKwjrsn4ADzqPGk0yTQ19FIqgoAfsCxWg7cDMtl71r
> -> ../../dm-0
> lrwxrwxrwx. 1 root root   10 Sep 18 16:40
> dm-uuid-LVM-ikNfztYY7KGT1SI2WYXPz4DhM2cyTelOq6Om5comvRFWJDbtVZAKtE5YGl4jciP9
> -> ../../dm-6
> lrwxrwxrwx. 1 root root   11 Sep 18 16:40
> dm-uuid-LVM-ikNfztYY7KGT1SI2WYXPz4DhM2cyTelOqVheASEgerWSEIkjM1BR3us3D9ekHt0L
> -> ../../dm-11
> lrwxrwxrwx. 1 root root   11 Sep 18 16:40
> dm-uuid-LVM-ikNfztYY7KGT1SI2WYXPz4DhM2cyTelOQz6vXuivIfup6cquKAjPof8wIGOSe4Vz
> -> ../../dm-12
> lrwxrwxrwx. 1 root root   10 Sep 18 23:35
> dm-uuid-mpath-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
> -> ../../dm-3
> lrwxrwxrwx. 1 root root   10 Sep 18 23:49
> dm-uuid-part1-mpath-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
> -> ../../dm-4
> lrwxrwxrwx. 1 root root   10 Sep 18 14:32
> dm-uuid-VDO-472035cc-8d2b-40ac-afe9-fa60b62a887f -> ../../dm-5
> lrwxrwxrwx. 1 root root   10 Sep 18 14:32
> lvm-pv-uuid-e1fvwo-kEfX-v3lT-SKBp-cgze-TwsO-PtyvmC -> ../../dm-5
> lrwxrwxrwx. 1 root root   10 Sep 18 22:40
> lvm-pv-uuid-mr9awW-oQH5-F4IX-CbEO-RgJZ-x4jK-e4YZS1 -> ../../sda2
> lrwxrwxrwx

[ovirt-users] Re: oVirt - Gluster Node Offline but Bricks Active

2020-09-21 Thread Jeremey Wise
Start is not an option.

It notes two bricks.  but command line denotes three bricks and all present

[root@odin thorst.penguinpages.local:_vmstore]# gluster volume status data
Status of volume: data
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick thorst.penguinpages.local:/gluster_br
icks/data/data  49152 0  Y
33123
Brick odinst.penguinpages.local:/gluster_br
icks/data/data  49152 0  Y
2970
Brick medusast.penguinpages.local:/gluster_
bricks/data/data49152 0  Y
2646
Self-heal Daemon on localhost   N/A   N/AY
3004
Self-heal Daemon on thorst.penguinpages.loc
al  N/A   N/AY
33230
Self-heal Daemon on medusast.penguinpages.l
ocalN/A   N/AY
2475

Task Status of Volume data
--
There are no active volume tasks

[root@odin thorst.penguinpages.local:_vmstore]# gluster peer status
Number of Peers: 2

Hostname: thorst.penguinpages.local
Uuid: 7726b514-e7c3-4705-bbc9-5a90c8a966c9
State: Peer in Cluster (Connected)

Hostname: medusast.penguinpages.local
Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
State: Peer in Cluster (Connected)
[root@odin thorst.penguinpages.local:_vmstore]#




On Mon, Sep 21, 2020 at 4:32 PM Strahil Nikolov 
wrote:

> Just select the volume and press "start" . It will automatically mark
> "force start" and will fix itself.
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В понеделник, 21 септември 2020 г., 20:53:15 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
>
>
>
>
>
>
> oVirt engine shows  one of the gluster servers having an issue.  I did a
> graceful shutdown of all three nodes over weekend as I have to move around
> some power connections in prep for UPS.
>
> Came back up.. but
>
>
>
> And this is reflected in 2 bricks online (should be three for each volume)
>
>
> Command line shows gluster should be happy.
>
> [root@thor engine]# gluster peer status
> Number of Peers: 2
>
> Hostname: odinst.penguinpages.local
> Uuid: 83c772aa-33cd-430f-9614-30a99534d10e
> State: Peer in Cluster (Connected)
>
> Hostname: medusast.penguinpages.local
> Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
> State: Peer in Cluster (Connected)
> [root@thor engine]#
>
> # All bricks showing online
> [root@thor engine]# gluster volume status
> Status of volume: data
> Gluster process TCP Port  RDMA Port  Online
>  Pid
>
> --
> Brick thorst.penguinpages.local:/gluster_br
> icks/data/data  49152 0  Y
> 11001
> Brick odinst.penguinpages.local:/gluster_br
> icks/data/data  49152 0  Y
> 2970
> Brick medusast.penguinpages.local:/gluster_
> bricks/data/data49152 0  Y
> 2646
> Self-heal Daemon on localhost   N/A   N/AY
> 50560
> Self-heal Daemon on odinst.penguinpages.loc
> al  N/A   N/AY
> 3004
> Self-heal Daemon on medusast.penguinpages.l
> ocalN/A   N/AY
> 2475
>
> Task Status of Volume data
>
> --
> There are no active volume tasks
>
> Status of volume: engine
> Gluster process TCP Port  RDMA Port  Online
>  Pid
>
> --
> Brick thorst.penguinpages.local:/gluster_br
> icks/engine/engine  49153 0  Y
> 11012
> Brick odinst.penguinpages.local:/gluster_br
> icks/engine/engine  49153 0  Y
> 2982
> Brick medusast.penguinpages.local:/gluster_
> bricks/engine/engine49153 0  Y
> 2657
> Self-heal Daemon on localhost   N/A   N/AY
> 50560
> Self-heal Daemon on odinst.penguinpages.loc
> al  N/A   N/AY
> 3004
> Self-heal Daemon on medusast.penguinpages.l
> ocalN/A   N/AY
> 2475
>
> Task Status of Volume engine
>
> --
> There are no active volum

[ovirt-users] oVirt - Gluster Node Offline but Bricks Active

2020-09-21 Thread Jeremey Wise
oVirt engine shows  one of the gluster servers having an issue.  I did a
graceful shutdown of all three nodes over weekend as I have to move around
some power connections in prep for UPS.

Came back up.. but

[image: image.png]

And this is reflected in 2 bricks online (should be three for each volume)
[image: image.png]

Command line shows gluster should be happy.

[root@thor engine]# gluster peer status
Number of Peers: 2

Hostname: odinst.penguinpages.local
Uuid: 83c772aa-33cd-430f-9614-30a99534d10e
State: Peer in Cluster (Connected)

Hostname: medusast.penguinpages.local
Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
State: Peer in Cluster (Connected)
[root@thor engine]#

# All bricks showing online
[root@thor engine]# gluster volume status
Status of volume: data
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick thorst.penguinpages.local:/gluster_br
icks/data/data  49152 0  Y
11001
Brick odinst.penguinpages.local:/gluster_br
icks/data/data  49152 0  Y
2970
Brick medusast.penguinpages.local:/gluster_
bricks/data/data49152 0  Y
2646
Self-heal Daemon on localhost   N/A   N/AY
50560
Self-heal Daemon on odinst.penguinpages.loc
al  N/A   N/AY
3004
Self-heal Daemon on medusast.penguinpages.l
ocalN/A   N/AY
2475

Task Status of Volume data
--
There are no active volume tasks

Status of volume: engine
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick thorst.penguinpages.local:/gluster_br
icks/engine/engine  49153 0  Y
11012
Brick odinst.penguinpages.local:/gluster_br
icks/engine/engine  49153 0  Y
2982
Brick medusast.penguinpages.local:/gluster_
bricks/engine/engine49153 0  Y
2657
Self-heal Daemon on localhost   N/A   N/AY
50560
Self-heal Daemon on odinst.penguinpages.loc
al  N/A   N/AY
3004
Self-heal Daemon on medusast.penguinpages.l
ocalN/A   N/AY
2475

Task Status of Volume engine
--
There are no active volume tasks

Status of volume: iso
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick thorst.penguinpages.local:/gluster_br
icks/iso/iso49156 49157  Y
151426
Brick odinst.penguinpages.local:/gluster_br
icks/iso/iso49156 49157  Y
69225
Brick medusast.penguinpages.local:/gluster_
bricks/iso/iso  49156 49157  Y
45018
Self-heal Daemon on localhost   N/A   N/AY
50560
Self-heal Daemon on odinst.penguinpages.loc
al  N/A   N/AY
3004
Self-heal Daemon on medusast.penguinpages.l
ocalN/A   N/AY
2475

Task Status of Volume iso
--
There are no active volume tasks

Status of volume: vmstore
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick thorst.penguinpages.local:/gluster_br
icks/vmstore/vmstore49154 0  Y
11023
Brick odinst.penguinpages.local:/gluster_br
icks/vmstore/vmstore49154 0  Y
2993
Brick medusast.penguinpages.local:/gluster_
bricks/vmstore/vmstore  49154 0  Y
2668
Self-heal Daemon on localhost   N/A   N/AY
50560
Self-heal Daemon on medusast.penguinpages.l
ocalN/A   N/AY
2475
Self-heal Daemon on odinst.penguinpages.loc
al  N/A   N/AY
3004

Task Status of Volume vmstore
--
There are no active volume tasks

[root@thor engine]# gluster volume heal
data engine   iso  vmstore
[root@thor engine]# gluster volume heal data info
Brick thorst.penguinpages.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0

Brick odinst.penguinpages.local:/gluster_bricks/data/data
Status: Connected
Number of entries: 0

Brick 

[ovirt-users] Re: oVirt - KVM QCow2 Import

2020-09-21 Thread Jeremey Wise
Ugh.. this is bad.

On the hypervisor where the files are located ...

My customers send me tar files with VMs all the time.   And I send them.
This will make it much more difficult, if I can't import xml / qcow2



This cluster .. is my home cluster and so..  three servers.. and they were
CentOs 7 + VDO + Gluster...  I use to have link qemu directory from all
three on gluster so if one server died .. or I messed it up and it was
needing repair... I could still start up and run the VMs.

Old cluster notes:

 Optional: Redirect Default KVM VM Storage location.  Ex:  /data/gv0/vms
on thor

# <<<<< Broken with HCI.. not sure process here yet….. hold off till oVirt
HCI engine issues worked out on how it enables new VM definitions to be
shared if one or more nodes goes down  2020-09-17 >>>>

#  Pool default XML configuration edited.

virsh pool-edit default



  default

  d3ae9e9a-8bc8-4a17-8476-3fe3334204f3

  37734498304

  27749486592

  9985011712

  

  

  

#   /var/lib/libvirt/images

/data/gv0/vms



  0711

  0

  0



  





#  For now each KVM host has shared folder linked.  Not sure how with out
restart of libvirtd to get peers to easily see configuration file. Can run
import command but need to test.

# To enable multiple KVM nodes in a shared environment to be able to take
over the roles of peers in the event of one failing the XML files stored in
/etc/libvirt/qemu/  need to be on a shared device.

# Ex:  Move medusa /etc/libvirt/qemu/   to be on gluster share volume space
/data/gv0/vms/medusa

systemctl stop libvirtd

mkdir -p /media/vmstore/qemu

mv -f /etc/libvirt/qemu/* /media/vmstore/qemu

ln -s /media/vmstore/qemu /etc/libvirt/qemu



systemctl daemon-reload

systemctl start libvirt-guests.service

systemctl enable libvirt-guests.service

systemctl status libvirt-guests.service



As I tried to use setup of engine it became apparent, my manual use of
libvirtd setup was NOT going to be any way helpful with oVirt way of using
it...  Ok... I can learn new things..


I had to backup and remove all data (see other post about errors for HCI
wizard failing if it detected existing VDO volume)...  So I moved my four
or so important VMs off to an external mount.

I now need a way to bring them back.  I really can't spend weeks rebuilding
those infrastructure VMs.  And I don't have a fourth server to rebuild
hosting KVM system to import and then with oVirt to LibVirt connection..
slurp vm out.
Plus.. that means anytime someone sends me a tar of qcow2 and xml..  I have
to re-host to export..  :P



On Mon, Sep 21, 2020 at 8:18 AM Nir Soffer  wrote:

> On Mon, Sep 21, 2020 at 9:11 AM Jeremey Wise 
> wrote:
> >
> >
> > I rebuilt my lab environment.   And their are four or five VMs that
> really would help if I did not have to rebuild.
> >
> > oVirt as I am now finding when it creates infrastructure, sets it out
> such that I cannot just use older  means of placing .qcow2 files in folders
> and .xml files in other folders and they show up on services restarting.
> >
> > How do I import VMs from files?
>
> You did not share the oVirt version, so I'm assuming 4.4.
>
> The simplest way is to upload the qcow2 images to oVirt, and create a new
> VM with the new disk.
>
> On the hypervisor where the files are located, install the required
> packages:
>
> dnf install ovirt-imageio-client python3-ovirt-engine-sdk4
>
> And upload the image:
>
> python3
> /usr/share/doc/python3-ovirt-engine-sdk4/examples/upload_disk.py \
> --engine-url https://my.engine/ \
> --username admin@internal \
> --password-file /path/to/password/file \
> --cafile /path/to/cafile \
> --sd-name my-storage-domain \
> --disk-sparse \
> /path/to/image.qcow2
>
> This will upload the file in qcow2 format to whatever type of storage you
> have. You can change the format if you like using --disk-format. See --help
> for all the options.
>
> We also support importing from libvirt, but for this you need to have the
> vm
> defined in libvirt. If you don't have this, It will probably be easier to
> upload
> the images and create a new vm in oVirt.
>
> Nir
>
> > I found this article but implies VM is running:
> https://www.ovirt.org/develop/release-management/features/virt/KvmToOvirt.html
> >
> https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.1/html/administration_guide/sect-adding_external_providers#Adding_KVM_as_an_External_Provider
> >
> > I need a way to import a file.  Even if it means temporarily hosting on
> "KVM on one of the hosts to then bring in once it is up.
> >
> >
> > Thanks
> > --
> >
> > penguinpages
> > ___
&g

[ovirt-users] Re: oVirt - vdo: ERROR - Device /dev/sd excluded by a filter

2020-09-21 Thread Jeremey Wise
Old System
Three servers..  Centos 7 -> Lay down VDO (dedup / compression) add those
VDO volumes as bricks to gluster.

New cluster (remove boot drives and run wipe of all data drives)

Goal: use first 512GB Drives to ignite the cluster and get things on feet
and stage infrastructure things.  Then use one of the 1TB drives in each
server for my "production" volume.  And second 1TB drive in each server as
staging.  I want to be able to "learn" and not loose days / weeks of
data... so disk level rather give up capacity for sake of "oh.. well ..
that messed up.. rebuild.

After minimal install.  Setup of network..  run HCI wizard.

It failed various times along build... lack SELInux permissive, .. did not
wipe 1TB drives with hope of importing old Gluster file system / VDO
voluemes to import my five or six custom and important VMs. (OCP cluster
bootstrap environment, Plex servers, DNS / DHCP / Proxy HA cluster nodes
et)

Gave up on too many HCI failures about disk.. so wiped drives (will use
external NAS to repopulate important VMs back (or so is plan... see other
posting on no import of qcow2 images / xml :P )

Ran into next batch of issues about use of true device ID  ... as name too
long... but /dev/sd?  makes me nervious as I have seen many systems with
issues when they use this old and should be depricated means to address
disk ID:  use UUID  or raw ID...
"/dev/disk/by-id/ata-Samsung_SSD_850_PRO_512GB_S250NXAGA15787L

Started getting errors about HCI failing with "excluded by filter" errors.

wiped drives ( gdisk /dev/sd?  => x => z => y => y)

filters errors I could not fiture out what they were.. .. error of "filter
exists"  to me meant ..  you have one.. remove it so I can remove drive.

Did full dd if=/dev/zero of=dev/sd? ..  still same issue
filtered in multipath just for grins still same issue.

Posted to forums.. nobody had ideas
https://forums.centos.org/viewtopic.php?f=54=75687   Posted to slack
gluster channel.. they looked at it and could not figure out...

Wiped systems.. started over.   This time the HCI wizard deployed.

My guess... is once I polished setup to make sure wizard did not attempt
before SELinux set to permissive (vs disable)  drives all wiped (even
though they SHOULD just be ignored..  I I think VDO scanned and saw VDO
definition on drive so freeked some ansible wizard script out).

Now cluster is up..  but then went to add "production"  gluster +VDO and
"staging"  gluster + vdo volumes... and having issues.

Sorry for long back story but I think this will add color to issues.

My Thoughts as to root issues
1) HCI wizard has issues just using drives told, and ignoring other data
drives in system ... VDO as example I saw notes about failed attempt ...
but it should not have touched that volume... just used one it needed and
igored rest.
2) HCI wizard bug of ignoring user set /dev/sd?  for each server again, was
another failure attempt where clean up may not have run. (noted this in
posting about manual edit .. and apply button :P to ingest)
3) HCI wizard bug of name I was using of device ID vs /sd?  which is IMAO
... bad form.. but name too long.. again. another cleanup where things may
not have fully cleaned.. or I forgot to click clean ...  where system was
left in non-pristine state
2) HCI wizard does NOT clean itself up properly if it fails ... or when I
ran clean up, maybe it did not complete and I closed wizard which then
created this orphaned state.
3) HCI Setup and post setup needs to add filtering


  With a perfect and pristine process  .. it ran.   But only when all other
learning and requirements to get it just right were setup first.  oVirt HCI
is S very close to being a great platform , well thought out and
production class.  Just needs some more nerds beating on it to find these
cracks, and get the GUI and setup polished.

My $0.002


On Mon, Sep 21, 2020 at 8:06 AM Nir Soffer  wrote:

> On Mon, Sep 21, 2020 at 9:02 AM Jeremey Wise 
> wrote:
> >
> >
> >
> >
> >
> >
> > vdo: ERROR - Device /dev/sdc excluded by a filter
> >
> >
> >
> >
> >
> > Other server
> >
> > vdo: ERROR - Device
> /dev/mapper/nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001p1
> excluded by a filter.
> >
> >
> >
> > All systems when I go to create VDO volume on blank drives.. I get this
> filter error.  All disk outside of the HCI wizard setup are now blocked
> from creating new Gluster volume group.
> >
> > Here is what I see in /dev/lvm/lvm.conf |grep filter
> > [root@odin ~]# cat /etc/lvm/lvm.conf |grep filter
> > filter =
> ["a|^/dev/disk/by-id/lvm-pv-uuid-e1fvwo-kEfX-v3lT-SKBp-cgze-TwsO-PtyvmC$|",
> "a|^/dev/disk/by-id/lvm-pv-uuid-mr9awW-oQH5-F4IX-CbEO-RgJZ-x4jK-e4YZS1$|

[ovirt-users] oVirt - KVM QCow2 Import

2020-09-21 Thread Jeremey Wise
I rebuilt my lab environment.   And their are four or five VMs that really
would help if I did not have to rebuild.

oVirt as I am now finding when it creates infrastructure, sets it out such
that I cannot just use older  means of placing .qcow2 files in folders and
.xml files in other folders and they show up on services restarting.

How do I import VMs from files?

I found this article but implies VM is running:
https://www.ovirt.org/develop/release-management/features/virt/KvmToOvirt.html
https://access.redhat.com/documentation/en-us/red_hat_virtualization/4.1/html/administration_guide/sect-adding_external_providers#Adding_KVM_as_an_External_Provider

I need a way to import a file.  Even if it means temporarily hosting on
"KVM on one of the hosts to then bring in once it is up.


Thanks
-- 

penguinpages
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6LSE4MNEBGODIRPVAQCUNBO2KGCCQTM5/


[ovirt-users] oVirt - vdo: ERROR - Device /dev/sd excluded by a filter

2020-09-21 Thread Jeremey Wise
[image: image.png]

vdo: ERROR - Device /dev/sdc excluded by a filter

[image: image.png]


Other server
vdo: ERROR - Device
/dev/mapper/nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001p1
excluded by a filter.


All systems when I go to create VDO volume on blank drives.. I get this
filter error.  All disk outside of the HCI wizard setup are now blocked
from creating new Gluster volume group.

Here is what I see in /dev/lvm/lvm.conf |grep filter
[root@odin ~]# cat /etc/lvm/lvm.conf |grep filter
filter =
["a|^/dev/disk/by-id/lvm-pv-uuid-e1fvwo-kEfX-v3lT-SKBp-cgze-TwsO-PtyvmC$|",
"a|^/dev/disk/by-id/lvm-pv-uuid-mr9awW-oQH5-F4IX-CbEO-RgJZ-x4jK-e4YZS1$|",
"r|.*|"]

[root@odin ~]# ls -al /dev/disk/by-id/
total 0
drwxr-xr-x. 2 root root 1220 Sep 18 14:32 .
drwxr-xr-x. 6 root root  120 Sep 18 14:32 ..
lrwxrwxrwx. 1 root root9 Sep 18 22:40
ata-INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN -> ../../sda
lrwxrwxrwx. 1 root root   10 Sep 18 22:40
ata-INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN-part1 -> ../../sda1
lrwxrwxrwx. 1 root root   10 Sep 18 22:40
ata-INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN-part2 -> ../../sda2
lrwxrwxrwx. 1 root root9 Sep 18 14:32
ata-Micron_1100_MTFDDAV512TBN_17401F699137 -> ../../sdb
lrwxrwxrwx. 1 root root9 Sep 18 22:40
ata-WDC_WDS100T2B0B-00YS70_183533804564 -> ../../sdc
lrwxrwxrwx. 1 root root   10 Sep 18 16:40 dm-name-cl-home -> ../../dm-2
lrwxrwxrwx. 1 root root   10 Sep 18 16:40 dm-name-cl-root -> ../../dm-0
lrwxrwxrwx. 1 root root   10 Sep 18 16:40 dm-name-cl-swap -> ../../dm-1
lrwxrwxrwx. 1 root root   11 Sep 18 16:40
dm-name-gluster_vg_sdb-gluster_lv_data -> ../../dm-11
lrwxrwxrwx. 1 root root   10 Sep 18 16:40
dm-name-gluster_vg_sdb-gluster_lv_engine -> ../../dm-6
lrwxrwxrwx. 1 root root   11 Sep 18 16:40
dm-name-gluster_vg_sdb-gluster_lv_vmstore -> ../../dm-12
lrwxrwxrwx. 1 root root   10 Sep 18 23:35
dm-name-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
-> ../../dm-3
lrwxrwxrwx. 1 root root   10 Sep 18 23:49
dm-name-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001p1
-> ../../dm-4
lrwxrwxrwx. 1 root root   10 Sep 18 14:32 dm-name-vdo_sdb -> ../../dm-5
lrwxrwxrwx. 1 root root   10 Sep 18 16:40
dm-uuid-LVM-GpvYIuypEfrR7nEDn5uHPenKwjrsn4ADc49gc6PWLRBCoJ2B3JC9tDJejyx5eDPT
-> ../../dm-1
lrwxrwxrwx. 1 root root   10 Sep 18 16:40
dm-uuid-LVM-GpvYIuypEfrR7nEDn5uHPenKwjrsn4ADOMNJfgcat9ZLOpcNO7FyG8ixcl5s93TU
-> ../../dm-2
lrwxrwxrwx. 1 root root   10 Sep 18 16:40
dm-uuid-LVM-GpvYIuypEfrR7nEDn5uHPenKwjrsn4ADzqPGk0yTQ19FIqgoAfsCxWg7cDMtl71r
-> ../../dm-0
lrwxrwxrwx. 1 root root   10 Sep 18 16:40
dm-uuid-LVM-ikNfztYY7KGT1SI2WYXPz4DhM2cyTelOq6Om5comvRFWJDbtVZAKtE5YGl4jciP9
-> ../../dm-6
lrwxrwxrwx. 1 root root   11 Sep 18 16:40
dm-uuid-LVM-ikNfztYY7KGT1SI2WYXPz4DhM2cyTelOqVheASEgerWSEIkjM1BR3us3D9ekHt0L
-> ../../dm-11
lrwxrwxrwx. 1 root root   11 Sep 18 16:40
dm-uuid-LVM-ikNfztYY7KGT1SI2WYXPz4DhM2cyTelOQz6vXuivIfup6cquKAjPof8wIGOSe4Vz
-> ../../dm-12
lrwxrwxrwx. 1 root root   10 Sep 18 23:35
dm-uuid-mpath-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
-> ../../dm-3
lrwxrwxrwx. 1 root root   10 Sep 18 23:49
dm-uuid-part1-mpath-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
-> ../../dm-4
lrwxrwxrwx. 1 root root   10 Sep 18 14:32
dm-uuid-VDO-472035cc-8d2b-40ac-afe9-fa60b62a887f -> ../../dm-5
lrwxrwxrwx. 1 root root   10 Sep 18 14:32
lvm-pv-uuid-e1fvwo-kEfX-v3lT-SKBp-cgze-TwsO-PtyvmC -> ../../dm-5
lrwxrwxrwx. 1 root root   10 Sep 18 22:40
lvm-pv-uuid-mr9awW-oQH5-F4IX-CbEO-RgJZ-x4jK-e4YZS1 -> ../../sda2
lrwxrwxrwx. 1 root root   13 Sep 18 14:32
nvme-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001
-> ../../nvme0n1
lrwxrwxrwx. 1 root root   15 Sep 18 14:32
nvme-nvme.126f-4141303030303030303030303030303032343538-53504343204d2e32205043496520535344-0001-part1
-> ../../nvme0n1p1
lrwxrwxrwx. 1 root root   13 Sep 18 14:32
nvme-SPCC_M.2_PCIe_SSD_AA002458 -> ../../nvme0n1
lrwxrwxrwx. 1 root root   15 Sep 18 14:32
nvme-SPCC_M.2_PCIe_SSD_AA002458-part1 -> ../../nvme0n1p1
lrwxrwxrwx. 1 root root9 Sep 18 22:40
scsi-0ATA_INTEL_SSDSC2BB08_BTWL40350DXP080KGN -> ../../sda
lrwxrwxrwx. 1 root root   10 Sep 18 22:40
scsi-0ATA_INTEL_SSDSC2BB08_BTWL40350DXP080KGN-part1 -> ../../sda1
lrwxrwxrwx. 1 root root   10 Sep 18 22:40
scsi-0ATA_INTEL_SSDSC2BB08_BTWL40350DXP080KGN-part2 -> ../../sda2
lrwxrwxrwx. 1 root root9 Sep 18 14:32
scsi-0ATA_Micron_1100_MTFD_17401F699137 -> ../../sdb
lrwxrwxrwx. 1 root root9 Sep 18 22:40
scsi-0ATA_WDC_WDS100T2B0B-_183533804564 -> ../../sdc
lrwxrwxrwx. 1 root root9 Sep 18 22:40
scsi-1ATA_INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN -> ../../sda
lrwxrwxrwx. 1 root root   10 Sep 18 22:40
scsi-1ATA_INTEL_SSDSC2BB080G4_BTWL40350DXP080KGN-part1 -> ../../sda1
lrwxrwxrwx. 1 root root   10 

[ovirt-users] Gluster Name too long

2020-09-12 Thread Jeremey Wise
Deployment on three node cluster using oVirt HCI wizard.

I think this is a bug where it needs to do either a pre-flight name length
validation, or increase valid field length.


I avoid using /dev/sd#   as those can change.  And the wizard allows for
this change to a more explicit devices Ex:
/dev/mapper/Samsung_SSD_850_PRO_512GB_S250NXAGA15787L


Error:
TASK [gluster.infra/roles/backend_setup : Create a LV thinpool for similar
device types] ***
task path:
/etc/ansible/roles/gluster.infra/roles/backend_setup/tasks/thin_pool_create.yml:239
failed: [thorst.penguinpages.local] (item={'vgname':
'gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L', 'thinpoolname':
'gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L',
'poolmetadatasize': '3G'}) => {"ansible_loop_var": "item", "changed":
false, "err": "  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
is too long.\n  Full LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tmeta\"
is too long.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tmeta\"
length 130 is not supported.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
length 130 is not supported.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tmeta\"
length 130 is not supported.\n  Internal error: LV name
\"gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L/gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L_tdata\"
length 130 is not supported.\n", "item": {"poolmetadatasize": "3G",
"thinpoolname":
"gluster_thinpool_gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L",
"vgname": "gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L"}, "msg":
"Creating logical volume 'None' failed", "rc": 5}
failed: [medusast.penguinpages.local] (item={'vgname':
'gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306', 'thinpoolname':
'gluster_thinpool_gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306',
'poolmetadatasize': '3G'}) => {"ansible_loop_var": "item", "changed":
false, "err": "  Internal error: LV name
\"gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306/gluster_thinpool_gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306\"
length 130 is not supported.\n", "item": {"poolmetadatasize": "3G",
"thinpoolname":
"gluster_thinpool_gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306",
"vgname": "gluster_vg_SAMSUNG_SSD_PM851_mSATA_512GB_S1EWNYAF609306"},
"msg": "Creating logical volume 'None' failed", "rc": 5}
changed: [odinst.penguinpages.local] => (item={'vgname':
'gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137', 'thinpoolname':
'gluster_thinpool_gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137',
'poolmetadatasize': '3G'}) => {"ansible_loop_var": "item", "changed": true,
"item": {"poolmetadatasize": "3G", "thinpoolname":
"gluster_thinpool_gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137",
"vgname": "gluster_vg_Micron_1100_MTFDDAV512TBN_17401F699137"}, "msg": ""}


I will revert back to  /dev/sd# for now... but this should be cleaned up.

Attached is YAML file for deployment of cluster

-- 
p enguinpages
hc_nodes:
  hosts:
thorst.penguinpages.local:
  gluster_infra_volume_groups:
- vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  pvname: /dev/mapper/vdo_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  gluster_infra_mount_devices:
- path: /gluster_bricks/engine
  lvname: gluster_lv_engine
  vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
- path: /gluster_bricks/data
  lvname: gluster_lv_data
  vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
- path: /gluster_bricks/vmstore
  lvname: gluster_lv_vmstore
  vgname: gluster_vg_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  gluster_infra_vdo:
- name: vdo_Samsung_SSD_850_PRO_512GB_S250NXAGA15787L
  device: