Re: [ovirt-users] Ovirt Cluster Setup

2018-02-20 Thread Kasturi Narra
Hello sakhi,

  Can you please let us know what is the script it is failing
at ?

Thanks
kasturi

On Tue, Feb 20, 2018 at 1:05 PM, Sakhi Hadebe  wrote:

> I have 3 Dell R515 servers all installed with centOS 7, and trying to
> setup an oVirt Cluster.
>
> Disks configurations:
> 2 x 1TB - Raid1 - OS Deployment
> 6 x 1TB - Raid 6 - Storage
>
> ​Memory is 128GB
>
> I am following this documentation https://www.
> ovirt.org/blog/2017/04/up-and-running-with-ovirt-4.1-and-gluster-storage/
> and I am getting the issue below:
>
> PLAY [gluster_servers] **
> ***
>
> TASK [Run a shell script] **
> 
> fatal: [ovirt2.sanren.ac.za]: FAILED! => {"msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt3.sanren.ac.za]: FAILED! => {"msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt1.sanren.ac.za]: FAILED! => {"msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> to retry, use: --limit @/tmp/tmpxFXyGG/run-script.retry
>
> PLAY RECAP 
> *
> ovirt1.sanren.ac.za: ok=0changed=0unreachable=0
> failed=1
> ovirt2.sanren.ac.za: ok=0changed=0unreachable=0
> failed=1
> ovirt3.sanren.ac.za: ok=0changed=0unreachable=0
> failed=1
>
> *Error: Ansible(>= 2.2) is not installed.*
> *Some of the features might not work if not installed.*
>
>
> ​I have installed ansible2.4 in all the servers, but the error persists.
>
> Is there anything I can do to get rid of this error?
> --
> Regards,
> Sakhi Hadebe
>
> Engineer: South African National Research Network (SANReN)Competency Area, 
> Meraka, CSIR
>
> Tel:   +27 12 841 2308 <+27128414213>
> Fax:   +27 12 841 4223 <+27128414223>
> Cell:  +27 71 331 9622 <+27823034657>
> Email: sa...@sanren.ac.za 
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Fail to add a second network

2018-01-23 Thread Kasturi Narra
Hi Carl,

 Can you try doing a refresh capabilities on the host ? Are you seeing
the same issue even after that ?

Thanks
kasturi

On Tue, Jan 23, 2018 at 11:05 PM, carl langlois <crl.langl...@gmail.com>
wrote:

> Hi Kasturi
>
> Thanks for the help. When i am able to add a new network but when i try to
> sync the network it always stay out of sync and if i ssh to that host i do
> not see the bridge but the network card is up and a ip is assign to it.
>
> Thanks
>
> Carl
>
> On Tue, Jan 23, 2018 at 2:38 AM, Kasturi Narra <kna...@redhat.com> wrote:
>
>> Hi Carl,
>>
>>  Below are the steps to configure back end network which is glusternw.
>>
>> 1) create a new network called 'glusternw'
>> 2) Now go to clusters  tab and select Logical networks tab.
>> 3) You should see the newly created network there.
>> 4) select the network and click on 'Manage Networks'
>> 5) In Manage networks window for the newly created network select both
>> 'Migration' & glusternw and click ok.
>> 6) Now go to the hosts tab click on one host and select 'Network
>> Interfaces' tab.
>> 7) select the interface to which you would like to assign the backend
>> network, drag and drop glusternw to that interface and click ok
>> 8) You will see that network is now out of sync.
>> 9) Please click refresh capabilities which will bring the network online.
>>
>> Hope this helps !!
>>
>> Thanks
>> kasturi
>>
>> On Tue, Jan 23, 2018 at 5:33 AM, carl langlois <crl.langl...@gmail.com>
>> wrote:
>>
>>> HI all,
>>>
>>> I am trying to add second network for back-end traffic. My setup is a 3
>>> hosts running with gluster storage in 3+1 replicat. Now i am trying to add
>>> a second network with the UI and when i try to had a the host interface to
>>> the new network it is always failing and i lost communication with this
>>> host.  I am not sure what are the step to add a second network with already
>>> 3 Host in the cluster.
>>>
>>> My setup us 4.2.
>>>
>>> Any inputs would be appreciated.
>>>
>>> Thanks
>>>
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Fail to add a second network

2018-01-22 Thread Kasturi Narra
Hi Carl,

 Below are the steps to configure back end network which is glusternw.

1) create a new network called 'glusternw'
2) Now go to clusters  tab and select Logical networks tab.
3) You should see the newly created network there.
4) select the network and click on 'Manage Networks'
5) In Manage networks window for the newly created network select both
'Migration' & glusternw and click ok.
6) Now go to the hosts tab click on one host and select 'Network
Interfaces' tab.
7) select the interface to which you would like to assign the backend
network, drag and drop glusternw to that interface and click ok
8) You will see that network is now out of sync.
9) Please click refresh capabilities which will bring the network online.

Hope this helps !!

Thanks
kasturi

On Tue, Jan 23, 2018 at 5:33 AM, carl langlois 
wrote:

> HI all,
>
> I am trying to add second network for back-end traffic. My setup is a 3
> hosts running with gluster storage in 3+1 replicat. Now i am trying to add
> a second network with the UI and when i try to had a the host interface to
> the new network it is always failing and i lost communication with this
> host.  I am not sure what are the step to add a second network with already
> 3 Host in the cluster.
>
> My setup us 4.2.
>
> Any inputs would be appreciated.
>
> Thanks
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Questions about converged infrastructure setup and glusterFS sizing/performance

2018-01-22 Thread Kasturi Narra
Hello Jayme,

   Please find the responses inline.

On Fri, Jan 19, 2018 at 7:44 PM, Jayme  wrote:

> I am attempting to narrow down choices for storage in a new oVirt build
> that will eventually be used for a mix of dev and production servers.
>
> My current space usage excluding backups sits at about only 1TB so I
> figure 3-5 TB would be more than enough for VM storage only + some room to
> grow.  There will be around 24 linux VMs total but 80% of them are VERY low
> usage and low spec servers.
>
> I've been considering a 3 host hyperconverged oVirt setup, replica 3
> arbiter 1 setup with a disaster recovery plan to replicate the gluster
> volume to a separate server.  I would of course do additional incremental
> backups to an alternate server as well probably with rsync or some other
> method.
>
> Some questions:
>
> 1. Is it recommended to use SSDs for glusterFS or can the performance of
> regular server/sas drives be sufficient enough performance.  If using SSDs
> is it recommended to use enterprise SSDs are consumer SSDs good enough due
> to the redundancy of glusterFS?   I would love to hear of any use cases
> from any of you regarding hardware specs you used in hyperconverged setups
> and what level of performance you are seeing.
>

You can use SSD's if you would like to. But you could use regular server/
sas drives too.

>
> 2. Is it recommended to RAID the drives that form the gluster bricks?  If
> so what raid level?
>

RAID level can be Raid5/6.

>
> 3. How do I calculate how much space will be usable in a replicate 3
> arbiter 1 configuration?  Will it be 75% of total drive capacity minus what
> I lose from raid (if I raid the drives)?
>
Each replica subvolume is defined to have 1 arbiter out of the 3 bricks.
The arbiter bricks are taken from the end of each replica subvolume. Since
the arbiter brick does not store file data, its disk usage will be
considerably less than the other bricks of the replica. The sizing of the
brick will depend on how many files you plan to store in the volume. A good
estimate will be 4kb times the number of files in the replica. In the other
two nodes if the size of the brick is 1TB, then total capacity will be 1 TB
.

>
> 4. For replication of the gluster volume, is it possible for me to
> replicate the entire volume to a single drive/raid array in an alternate
> server or does the replicated volume need to match the configuration of the
> main glusterFS volume (i.e. same amount of drives/configuration etc).
>

you could replicate the entire volume to another single volume but please
make sure that size of the volume is more so that no data is lost.

>
> 5. Has the meltdown bug caused or expected to cause major issues with
> oVirt hyperconverged setup due to performance loss from the patches.  I've
> been reading articles suggesting up to 30% performance loss on some
> converged/storage setups due to how CPU intensive converged setups are.
>
> Thanks in advance!
>
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine unknow stale-data

2018-01-21 Thread Kasturi Narra
:11:09 AM Artem Tambovskiy <
>>>> artem.tambovs...@gmail.com>
>>>> > wrote:
>>>> >>
>>>> >> Hello,
>>>> >>
>>>> >> Yes, I followed exactly the same procedure while reinstalling the
>>>> hosts
>>>> >> (the only difference that I have SSH key configured instead of the
>>>> >> password).
>>>> >>
>>>> >> Just reinstalled the second host one more time, after 20 min the host
>>>> >> still haven't reached active score of 3400 (Hosted Engine HA:Not
>>>> Active) and
>>>> >> I still don't see crown icon for this host.
>>>> >>
>>>> >> hosted-engine --vm-status  from ovirt1 host
>>>> >>
>>>> >> [root@ovirt1 ~]# hosted-engine --vm-status
>>>> >>
>>>> >>
>>>> >> --== Host 1 status ==--
>>>> >>
>>>> >> conf_on_shared_storage : True
>>>> >> Status up-to-date  : True
>>>> >> Hostname   : ovirt1.telia.ru
>>>> >> Host ID: 1
>>>> >> Engine status  : {"health": "good", "vm": "up",
>>>> >> "detail": "up"}
>>>> >> Score  : 3400
>>>> >> stopped: False
>>>> >> Local maintenance  : False
>>>> >> crc32  : 3f94156a
>>>> >> local_conf_timestamp   : 349144
>>>> >> Host timestamp : 349144
>>>> >> Extra metadata (valid at timestamp):
>>>> >> metadata_parse_version=1
>>>> >> metadata_feature_version=1
>>>> >> timestamp=349144 (Tue Jan 16 15:03:45 2018)
>>>> >> host-id=1
>>>> >> score=3400
>>>> >> vm_conf_refresh_time=349144 (Tue Jan 16 15:03:45 2018)
>>>> >> conf_on_shared_storage=True
>>>> >> maintenance=False
>>>> >> state=EngineUp
>>>> >> stopped=False
>>>> >>
>>>> >>
>>>> >> --== Host 2 status ==--
>>>> >>
>>>> >> conf_on_shared_storage : True
>>>> >> Status up-to-date  : False
>>>> >> Hostname   : ovirt1.telia.ru
>>>> >> Host ID: 2
>>>> >> Engine status  : unknown stale-data
>>>> >> Score  : 0
>>>> >> stopped: True
>>>> >> Local maintenance  : False
>>>> >> crc32  : c7037c03
>>>> >> local_conf_timestamp   : 7530
>>>> >> Host timestamp : 7530
>>>> >> Extra metadata (valid at timestamp):
>>>> >> metadata_parse_version=1
>>>> >> metadata_feature_version=1
>>>> >> timestamp=7530 (Fri Jan 12 16:10:12 2018)
>>>> >> host-id=2
>>>> >> score=0
>>>> >> vm_conf_refresh_time=7530 (Fri Jan 12 16:10:12 2018)
>>>> >> conf_on_shared_storage=True
>>>> >> maintenance=False
>>>> >> state=AgentStopped
>>>> >> stopped=True
>>>> >>
>>>> >>
>>>> >> hosted-engine --vm-status output from ovirt2 host
>>>> >>
>>>> >> [root@ovirt2 ovirt-hosted-engine-ha]# hosted-engine --vm-status
>>>> >>
>>>> >>
>>>> >> --== Host 1 status ==--
>>>> >>
>>>> >> conf_on_shared_storage : True
>>>> >> Status up-to-date  : False
>>>> >> Hostname   : ovirt1.telia.ru
>>>> >> Host ID: 1
>>>> >> Engine status  : unknown stale-data
>>>> >> Score  : 3400
>>>> >> stopped

Re: [ovirt-users] gluster volume permission denied

2018-01-17 Thread Kasturi Narra
Hello Bill,

   can you attach vdsm logs during the time period where migration
failed ? It would help us to see why the migration failed.

Thanks
kasturi

On Wed, Jan 17, 2018 at 5:24 AM, Bill James  wrote:

> I have one node in our cluster that has problems when migrating a VM if
> the VM is hosted on a gluster volume.
> If the VM is hosted on a NFS volume it migrates fine.
>
> /var/log/messages:
> Jan 16 14:36:09 ovirt1 libvirtd: 2018-01-16 22:36:09.769+: 9700: error
> : qemuProcessReportLogError:1862
>  : internal error: process exited while connecting to monitor:
> 2018-01-16T22:36:09.587405Z qemu-kvm: -drive
>  file=/rhev/data-center/8b6303b3-79c6-4633-ae21-71b15ed00675
> /67b4d9aa-f174-436a-b5a1-ec7cee5f2edb/images/60
> 3271e4-3089-4779-aa1c-faf280fabd17/7c7f7726-4ab0-4047-9b91-
> e0817a139e90,format=raw,if=none,id=drive-virtio-
> disk0,serial=603271e4-3089-4779-aa1c-faf280fabd17,cache=none
> ,werror=stop,rerror=stop,aio=threads: Could not
>  open '/rhev/data-center/8b6303b3-79c6-4633-ae21-71b15ed00675/67b4
> d9aa-f174-436a-b5a1-ec7cee5f2edb/images/603271e4-3089-4779-
> aa1c-faf280fabd17/7c7f7726-4ab0-4047-9b91-e0817a139e90': Permission denied
> (libvirt/csapi1.test.j2noc.com.log says same thing)
>
> [root@ovirt1 test log]# ls -hl /rhev/data-center/8b6303b3-79c
> 6-4633-ae21-71b15ed00675/67b4d9aa-f174-436a-b5a1-ec7cee5f2ed
> b/images/603271e4-3089-4779-aa1c-faf280fabd17/7c7f7726-
> 4ab0-4047-9b91-e0817a139e90
> -rw-rw 1 vdsm kvm 20G Jan 16 15:32 /rhev/data-center/8b6303b3-79c
> 6-4633-ae21-71b15ed00675/67b4d9aa-f174-436a-b5a1-ec7cee5f2ed
> b/images/603271e4-3089-4779-aa1c-faf280fabd17/7c7f7726-
> 4ab0-4047-9b91-e0817a139e90
>
> lrwxrwxrwx 1 vdsm kvm  98 Jan 10 21:21 67b4d9aa-f174-436a-b5a1-ec7cee5f2edb
> -> /rhev/data-center/mnt/glusterSD/ovirt3-ks.test.j2noc.com:_
> gv3/67b4d9aa-f174-436a-b5a1-ec7cee5f2edb
>
>
> engine.log:
> 2018-01-16 14:36:18,846-08 WARN [org.ovirt.engine.core.dal.dbb
> roker.auditloghandling.AuditLogDirector] (o
> rg.ovirt.thread.pool-6-thread-34) [78d8b38c] EVENT_ID:
> VM_MIGRATION_TRYING_RERUN(128), Correlation ID: e39
> 9da53-cd31-401c-bf63-16b118b8884c, Job ID: 
> 4db6bd87-0da4-4358-acae-dc453c813969,
> Call Stack: null, Custom
> ID: null, Custom Event ID: -1, Message: Failed to migrate VM
> csapi1.test.j2noc.com to Host ovirt1.test.j2n
> oc.com . Trying to migrate to another Host.
>
>
> ovirt-engine-tools-4.1.8.2-1.el7.centos.noarch
> glusterfs-3.8.15-2.el7.x86_64
> vdsm-4.19.43-1.el7.centos.x86_64
> libvirt-daemon-3.2.0-14.el7_4.7.x86_64
>
> Let me know what logs would be useful in troubleshooting this.
>
> Thanks!
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted-engine unknow stale-data

2018-01-15 Thread Kasturi Narra
Hello Artem,

Can you check if glusterd service is running on host1 and all the
peers are in connected state ? If yes, can you restart ovirt-ha-agent and
broker services and check if things are working fine ?

Thanks
kasturi

On Sat, Jan 13, 2018 at 12:33 AM, Artem Tambovskiy <
artem.tambovs...@gmail.com> wrote:

> Explored logs on both hosts.
> broker.log shows no errors.
>
> agent.log looking not good:
>
> on host1 (which running hosted engine) :
>
> MainThread::ERROR::2018-01-12 21:51:03,883::agent::205::
> ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Traceback (most
> recent call last):
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 191, in _run_agent
> return action(he)
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 64, in action_proper
> return he.start_monitoring()
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 411, in start_monitoring
> self._initialize_sanlock()
>   File 
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/hosted_engine.py",
> line 749, in _initialize_sanlock
> "Failed to initialize sanlock, the number of errors has"
> SanlockInitializationError: Failed to initialize sanlock, the number of
> errors has exceeded the limit
>
> MainThread::ERROR::2018-01-12 21:51:03,884::agent::206::
> ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Trying to restart
> agent
> MainThread::WARNING::2018-01-12 21:51:08,889::agent::209::
> ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent) Restarting agent,
> attempt '1'
> MainThread::INFO::2018-01-12 21:51:08,919::hosted_engine::
> 242::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_get_hostname)
> Found certificate common name: ovirt1.telia.ru
> MainThread::INFO::2018-01-12 21:51:08,921::hosted_engine::
> 604::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_vdsm) Initializing VDSM
> MainThread::INFO::2018-01-12 21:51:11,398::hosted_engine::
> 630::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_storage_images) Connecting the storage
> MainThread::INFO::2018-01-12 21:51:11,399::storage_server::
> 220::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(validate_storage_server)
> Validating storage server
> MainThread::INFO::2018-01-12 21:51:13,725::storage_server::
> 239::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Connecting storage server
> MainThread::INFO::2018-01-12 21:51:18,390::storage_server::
> 246::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Connecting storage server
> MainThread::INFO::2018-01-12 21:51:18,423::storage_server::
> 253::ovirt_hosted_engine_ha.lib.storage_server.StorageServer::(connect_storage_server)
> Refreshing the storage domain
> MainThread::INFO::2018-01-12 21:51:18,689::hosted_engine::
> 663::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_storage_images) Preparing images
> MainThread::INFO::2018-01-12 21:51:18,690::image::126::
> ovirt_hosted_engine_ha.lib.image.Image::(prepare_images) Preparing images
> MainThread::INFO::2018-01-12 21:51:21,895::hosted_engine::
> 666::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_storage_images) Refreshing vm.conf
> MainThread::INFO::2018-01-12 21:51:21,895::config::493::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.config::(refresh_vm_conf)
> Reloading vm.conf from the shared storage domain
> MainThread::INFO::2018-01-12 21:51:21,896::config::416::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.
> config::(_get_vm_conf_content_from_ovf_store) Trying to get a fresher
> copy of vm configuration from the OVF_STORE
> MainThread::INFO::2018-01-12 21:51:21,896::ovf_store::132::
> ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(getEngineVMOVF)
> Extracting Engine VM OVF from the OVF_STORE
> MainThread::INFO::2018-01-12 21:51:21,897::ovf_store::134::
> ovirt_hosted_engine_ha.lib.ovf.ovf_store.OVFStore::(getEngineVMOVF)
> OVF_STORE volume path: /var/run/vdsm/storage/4a7f8717-9bb0-4d80-8016-
> 498fa4b88162/5cabd8e1-5f4b-469e-becc-227469e03f5c/8048cbd7-77e2-4805-9af4-
> d109fa36dfcf
> MainThread::INFO::2018-01-12 21:51:21,915::config::435::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.
> config::(_get_vm_conf_content_from_ovf_store) Found an OVF for HE VM,
> trying to convert
> MainThread::INFO::2018-01-12 21:51:21,918::config::440::
> ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine.
> config::(_get_vm_conf_content_from_ovf_store) Got vm.conf from OVF_STORE
> MainThread::INFO::2018-01-12 21:51:21,919::hosted_engine::
> 509::ovirt_hosted_engine_ha.agent.hosted_engine.
> HostedEngine::(_initialize_broker) Initializing ha-broker connection
> MainThread::INFO::2018-01-12 21:51:21,919::brokerlink::130:
> 

Re: [ovirt-users] (no subject)

2018-01-15 Thread Kasturi Narra
Hello,

Can you attach ovirt-ha-agent and ovirt-ha-broker logs ?

Thanks
kasturi

On Fri, Jan 12, 2018 at 9:38 PM, Artem Tambovskiy <
artem.tambovs...@gmail.com> wrote:

> Trying to fix one thing I broke another :(
>
> I fixed mnt_options for hosted engine storage domain and installed latest
> security patches to my hosts and hosted engine. All VM's up and running,
> but  hosted_engine --vm-status reports about issues:
>
> [root@ovirt1 ~]# hosted-engine --vm-status
>
>
> --== Host 1 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : False
> Hostname   : ovirt2
> Host ID: 1
> Engine status  : unknown stale-data
> Score  : 0
> stopped: False
> Local maintenance  : False
> crc32  : 193164b8
> local_conf_timestamp   : 8350
> Host timestamp : 8350
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=8350 (Fri Jan 12 19:03:54 2018)
> host-id=1
> score=0
> vm_conf_refresh_time=8350 (Fri Jan 12 19:03:54 2018)
> conf_on_shared_storage=True
> maintenance=False
> state=EngineUnexpectedlyDown
> stopped=False
> timeout=Thu Jan  1 05:24:43 1970
>
>
> --== Host 2 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : False
> Hostname   : ovirt1.telia.ru
> Host ID: 2
> Engine status  : unknown stale-data
> Score  : 0
> stopped: True
> Local maintenance  : False
> crc32  : c7037c03
> local_conf_timestamp   : 7530
> Host timestamp : 7530
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=7530 (Fri Jan 12 16:10:12 2018)
> host-id=2
> score=0
> vm_conf_refresh_time=7530 (Fri Jan 12 16:10:12 2018)
> conf_on_shared_storage=True
> maintenance=False
> state=AgentStopped
> stopped=True
> [root@ovirt1 ~]#
>
>
>
> from second host situation looks a bit different:
>
>
> [root@ovirt2 ~]# hosted-engine --vm-status
>
>
> --== Host 1 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : True
> Hostname   : ovirt2
> Host ID: 1
> Engine status  : {"reason": "vm not running on this
> host", "health": "bad", "vm": "down", "detail": "unknown"}
> Score  : 0
> stopped: False
> Local maintenance  : False
> crc32  : 78eabdb6
> local_conf_timestamp   : 8403
> Host timestamp : 8402
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=8402 (Fri Jan 12 19:04:47 2018)
> host-id=1
> score=0
> vm_conf_refresh_time=8403 (Fri Jan 12 19:04:47 2018)
> conf_on_shared_storage=True
> maintenance=False
> state=EngineUnexpectedlyDown
> stopped=False
> timeout=Thu Jan  1 05:24:43 1970
>
>
> --== Host 2 status ==--
>
> conf_on_shared_storage : True
> Status up-to-date  : False
> Hostname   : ovirt1.telia.ru
> Host ID: 2
> Engine status  : unknown stale-data
> Score  : 0
> stopped: True
> Local maintenance  : False
> crc32  : c7037c03
> local_conf_timestamp   : 7530
> Host timestamp : 7530
> Extra metadata (valid at timestamp):
> metadata_parse_version=1
> metadata_feature_version=1
> timestamp=7530 (Fri Jan 12 16:10:12 2018)
> host-id=2
> score=0
> vm_conf_refresh_time=7530 (Fri Jan 12 16:10:12 2018)
> conf_on_shared_storage=True
> maintenance=False
> state=AgentStopped
> stopped=True
>
>
> WebGUI shows that engine running on host ovirt1.
> Gluster looks fine
> [root@ovirt1 ~]# gluster volume status engine
> Status of volume: engine
> Gluster process TCP Port  RDMA Port  Online
>  Pid
> 
> --
> Brick ovirt1.telia.ru:/oVirt/engine 49169 0  Y
> 3244
> Brick ovirt2.telia.ru:/oVirt/engine 49179 0  Y
> 20372
> Brick ovirt3.telia.ru:/oVirt/engine 49206 0  Y

Re: [ovirt-users] Rebuilding my infra..

2018-01-09 Thread Kasturi Narra
Hi carl,

  During deployment via cockpit+gdeploy plugin when you input the host,
host in the third text box will be considered as Arbiter host.

Thanks
kasturi



On Tue, Jan 9, 2018 at 11:26 PM, carl langlois 
wrote:

> Some question about the arbiter box.
>
> 1- Lets says i am using a old box for the arbiter with minimum spec. Do i
> need to mark it as a arbiter box? Does the installer scripts need to know
> that it is a arbiter box?
> 2- When i get my hand on new host for running VM. Do i need still need the
> arbiter box?
>
> Thanks
>
> On Tue, Jan 9, 2018 at 2:28 AM, Yedidyah Bar David 
> wrote:
>
>> On Mon, Jan 8, 2018 at 10:57 PM, Karli Sjöberg 
>> wrote:
>>
>>>
>>>
>>> Den 8 jan. 2018 21:48 skrev Vinícius Ferrão :
>>>
>>> If I’m not wrong GlusterFS in oVirt requires 3 hosts.
>>>
>>>
>>> Wasn't there a Ridning Hood not long ago that said they'd change it down
>>> to 1 with 4.2, just to have something to get you started? To lower the bar
>>> for POC systems, like all-in-one?
>>>
>>
>> There is this:
>>
>> https://bugzilla.redhat.com/show_bug.cgi?id=1494112
>>
>> One dependent bug is in POST. Although I was asked to
>> review the patch, I don't know much about the general
>> plan, including whether that's enough or more bugs/fixes
>> will be needed. Adding Sahina.
>>
>> Best regards,
>>
>>
>>>
>>> /K
>>>
>>>
>>> Here’s the RHHI guide, it’s pretty much the same for oVirt:
>>> https://access.redhat.com/documentation/en-us/red_hat_hyperc
>>> onverged_infrastructure/1.1/html/deploying_red_hat_hyperconv
>>> erged_infrastructure/
>>>
>>> > On 8 Jan 2018, at 18:10, carl langlois wrote:
>>> >
>>> > Hi all
>>> >
>>> > After screwing my infra with the update to 4.2 (probably a bad
>>> manipulation), i am planning a rebuild of the entire infra. First i want to
>>> replace my NFS storage with a glusterfs storage. All documentation tell me
>>> that i need 3 hosts.. but for the moment i only have 2 but planning to had
>>> more later.
>>> >
>>> > So does it make sense to start with 2 hosts and use glusterfs as the
>>> storage domain(lets says with a replicate of two with all its limitations).
>>> > If it make sense,
>>> > 1- what is the best way to do it.
>>> > 2- how hard will it be to had the 3rd host when available and make it
>>> replica 2+arbiter.
>>> >
>>> > Also in a setup where i have 3 hosts (replica 2+arbiter) does all the
>>> 3 hosts can run users vm?
>>> >
>>> > Thanks for your inputs.
>>> >
>>> > Carl
>>> >
>>> > ___
>>> > Users mailing list
>>> > Users@ovirt.org
>>> > http://lists.ovirt.org/mailman/listinfo/users
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>>
>> --
>> Didi
>>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Two node configuration.

2017-12-20 Thread Kasturi Narra
Hello Jarek,

 As of today we cannot have gdeploy to work with different devices
on different nodes for deploying HC. Currently device name has to be same
on data and arbiter nodes.

Hope this helps !!

Thanks
kasturi

On Wed, Dec 20, 2017 at 2:29 PM, Jarek  wrote:

> One thing:
>
> I have two nodes with sda3 partitions for gluster deploy and one node with
> for example vdd (arbiter).
> How can I play with gdeploy config to change device name for arbiter node
> (without udevs)?
> --
> *From: *"Jaroslaw Augustynowicz" 
> *To: *"Sandro Bonazzola" 
> *Cc: *"users" 
> *Sent: *Friday, December 15, 2017 4:44:50 PM
>
> *Subject: *Re: [ovirt-users] Two node configuration.
>
> Yes, sure :)
>
> Dnia 15 grudnia 2017 16:43:10 CET, Sandro Bonazzola 
> napisał(a):
>>
>>
>>
>> 2017-12-15 16:37 GMT+01:00 Jarosław Augustynowicz :
>>
>>> It can be good idea:
>>> 2 physical nodes with local disks (raid6) as gluster bricks + 2 kvm vms
>>> with pcs and drbd for arbiter.
>>>
>>
>> Just be sure to not run the kvm vms arbiters on the same real hosts or it
>> may become a mess.
>>
>>
>>> I'll test it.
>>>
>>> Dnia 15 grudnia 2017 13:05:08 CET, Sandro Bonazzola 
>>> napisał(a):



 2017-12-15 12:53 GMT+01:00 Jarek :

> Yes, I checked it but it seems I still need three nodes - 2 for
> storage and one smaller for arbiter.
> Is it safe to deploy it only on two nodes?
>

 the host arbiter can be even just a Raspberry PI as far as I can tell.
 Adding Sahina.



> Am I wrong?
>
> --
> *From: *"Sandro Bonazzola" 
> *To: *"Jaroslaw Augustynowicz" 
> *Cc: *"users" 
> *Sent: *Friday, December 15, 2017 12:03:24 PM
> *Subject: *Re: [ovirt-users] Two node configuration.
>
>
>
> 2017-12-15 8:55 GMT+01:00 Jarek :
>
>> Hello, currently I'm using kvms with pcs on vms... is there any
>> ovirt solution for ha with two nodes (storage on local disks) without 
>> pcs&
>> drbd for storage? I know about gluster storage but it needs third host;/
>>
>
> Did you check https://access.redhat.com/documentation/en-us/red_
> hat_gluster_storage/3.3/html/administration_guide/creating_
> arbitrated_replicated_volumes ?
>
>
>
>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
>
> --
>
> SANDRO BONAZZOLA
>
> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R
>
> Red Hat EMEA 
> 
> TRIED. TESTED. TRUSTED. 
>
>
>


 --

 SANDRO BONAZZOLA

 ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R

 Red Hat EMEA 
 
 TRIED. TESTED. TRUSTED. 


>>
>>
>> --
>>
>> SANDRO BONAZZOLA
>>
>> ASSOCIATE MANAGER, SOFTWARE ENGINEERING, EMEA ENG VIRTUALIZATION R
>>
>> Red Hat EMEA 
>> 
>> TRIED. TESTED. TRUSTED. 
>>
>>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt Node 4.1 question, writing files to /root and RPMs

2017-12-19 Thread Kasturi Narra
Hello Matt,

   All the partitions will be persisted when gluster is installed on
the ovirt node since gluster recommends user not to create bricks in root
directory. If the gluster bricks are created in root partition then once
the update of the node is done, you will not be able to see any of the
bricks.

Hope this helps !!!

Thanks
kasturi.

On Tue, Dec 19, 2017 at 4:28 AM, Matt Simonsen  wrote:

> On 12/15/2017 03:06 AM, Simone Tiraboschi wrote:
>
> On Fri, Dec 15, 2017 at 4:45 AM, Donny Davis  wrote:
>
>> have you gotten an image update yet?
>>
>> On Thu, Dec 14, 2017 at 8:08 PM, Matt Simonsen  wrote:
>>
>>> Hello all,
>>>
>>> I read at https://www.ovirt.org/develop/projects/node/troubleshooting/
>>> that "Changes made from the command line are done at your own risk. Making
>>> changes has the potential to leave your system in an unusable state." It
>>> seems clear that RPMs should not be installed.
>>>
>>
> That document mainly refers to vintage node.
> In Next Generation Node now we have rpm persistence; please check
> https://www.ovirt.org/develop/release-management/features/
> node/node-next-persistence/
>
>
>
>
> I'm sure glad we tested!
>
> On one Node image we had images locally stored in /exports and shared out
> via NFS. After an upgrade & reboot, images are gone.
>
> If we "Convert to local storage" will the data persist?  I am planning to
> test, but want to be sure how this is designed.
>
> I assume during a Gluster installation something is also updated in oVirt
> Node to allow for the Gluster partition to persist?
>
> At this point I'm thinking I should manually install via CentOS7 to ensure
> folders and partitions are persistent. Is there any downside to installing
> over CentOS7?
>
> Thanks
> Matt
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt with Hosted Engine and 4 Hosts

2017-12-19 Thread Kasturi Narra
Hello Gabriel,

   Any specific reason  you are looking to expand storage for
engine volume?

  If you are running a HC setup recommended way of upgrading the
setup is to add three more hosts to the existing cluster and create volumes
out of the bricks carved from those additional hosts .

We do not recommend adding bricks to existing volumes in the HC
setup as it might lead to some issues.

Hope this helps !!!

Thanks
kasturi

On Tue, Dec 19, 2017 at 3:05 PM, Gabriel Stein 
wrote:

> Hi all,
>
> I know, that oVirt with Hosted Engine requires  a replica 3 glusterfs
> volume, but I'm using 4 nodes and  I would like to know if is possible to
> add it to the glusterfs engine volume.
>
> Or should I always use 3, 5 and 7 hosts with hosted-engine?
>
> What I'm doing:
>
> gluster volume add-brick engine server1:/gluster/engine/brick1
> server2:/gluster/engine/brick1 server3:/gluster/engine/brick1
> server4:/gluster/engine/brick1
>
> Error:
> volume add-brick: failed: Incorrect number of bricks supplied 4 with count
> 3
>
> If I'm using replica 3 will not be possible, but how can I upgrade it?
> Adding 2 more nodes and upgrade to replica 5?
>
> Thanks in Advance!
>
> Best Regards
>
> Gabriel Stein
>
> Gabriel Stein
> --
> Gabriel Ferraz Stein
> Tel.: +49 (0)  170 2881531
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Host ovritnode1 installation failed. Command returned failure code 1 during SSH session 'root@X.X.X.X'.

2017-12-07 Thread Kasturi Narra
Hello,

 Looks like there is a problem with the repo which is present in
your system. Can you please disable the repo and try installing the host
again? That should solve the problem.

Thanks
kasturi

On Thu, Dec 7, 2017 at 1:53 PM, M.I.S <1312121...@qq.com> wrote:

> hi,
>I encountered a problem.
>I am getting an error when adding a host to ovirt-engine:Host
> ovritnode1 installation failed. Command returned failure code 1 during SSH
> session 'root@192.168.1.152'.
>PS: all user and password is correct.
> I checked engine log, the information is as follows:
>
> 2017-12-06 18:58:52,995-05 INFO  [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [617b92e4]
> EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Correlation ID: 617b92e4, Call
> Stack: null, Custom ID: null, Custom Event ID: -1, Message: Installing Host
> ovirtnode1. Stage: Environment setup.
>
> 2017-12-06 18:58:53,029-05 INFO  [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [617b92e4]
> EVENT_ID: VDS_INSTALL_IN_PROGRESS(509), Correlation ID: 617b92e4, Call
> Stack: null, Custom ID: null, Custom Event ID: -1, Message: Installing Host
> ovirtnode1. Stage: Environment packages setup.
>
> 2017-12-06 18:59:22,974-05 ERROR [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [617b92e4]
> EVENT_ID: VDS_INSTALL_IN_PROGRESS_ERROR(511), Correlation ID: 617b92e4,
> Call Stack: null, Custom ID: null, Custom Event ID: -1, Message: *Failed
> to install Host ovirtnode1. Yum Cannot queue package iproute: Cannot
> retrieve metalink for repository: ovirt-4.1-epel/x86_64. Please verify its
> path and try again.*
>
> 2017-12-06 18:59:22,999-05 ERROR [org.ovirt.engine.core.dal.
> dbbroker.auditloghandling.AuditLogDirector] (VdsDeploy) [617b92e4]
> EVENT_ID: VDS_INSTALL_IN_PROGRESS_ERROR(511), Correlation ID: 617b92e4,
> Call Stack: null, Custom ID: null, Custom Event ID: -1, Message: *Failed
> to install Host ovirtnode1. Failed to execute stage 'Environment packages
> setup': Cannot retrieve metalink for repository: ovirt-4.1-epel/x86_64.
> Please verify its path and try again.*
>   How to solve this problem, please help analyze, thank you!
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fwd: Update gluster HCI from 4.1.3 to 4.1.7

2017-11-28 Thread Kasturi Narra
Hello,
I have an environment with 3 hosts and gluster HCI on 4.1.3.
I'm following this link to take it to 4.1.7
https://www.ovirt.org/documentation/how-to/hosted-
engine/#upgrade-hosted-engine

The hosts and engine were at 7.3 prior of beginning the update.
All went ok for the engine that now is on 7.4 (not rebooted yet)
Points 4. 5. 6. for the first updated host were substituted by rebooting it

I'm at point:

7. Exit the global maintenance mode: in a few minutes the engine VM should
migrate to the fresh upgraded host cause it will get an higher score

One note: actually exiting from global maintenance doesn't imply that the
host previously put into maintenance exiting from it, correct?

[kasturi] - you are right. Global maintenance main use is to allow
administrator start / stop / modify the engine vm with out any worry of
interference from the HA agents .

So in my workflow, before point 7., actually I have selected the host and
activated it.

Currently the situation is this
- engine running on ovirt02
- update happened on ovirt03

[kasturi] - looks fine.

Then after exiting from global maintenance I don't see the engine vm
migrating to it.
[kasturi] - which is expected.

And in fact (see below) the score of ovirt02 is the same (3400) as the one
of ovirt03, so it seems it is correct that engine remains there...?

[kasturi] - yes, it remains there. For a properly active host which has
everything configured properly  the score will be 3400. When global
maintenance is enabled on the cluster, active score becomes 'Global
Maintenance Enabled' (can be viewed from general tab of host ) and once
exited score on all the hosts becomes 3400 and hosted engine will be
running on the same host where it was running before enabling global
maintenance.

Which kind of messages should I see inside logs of engine/hosts?

[root@ovirt01 ~]# rpm -q vdsm
vdsm-4.19.20-1.el7.centos.x86_64

[root@ovirt02 ~]# rpm -q vdsm
vdsm-4.19.20-1.el7.centos.x86_64
[root@ovirt02 ~]#

[root@ovirt03 ~]# rpm -q vdsm
vdsm-4.19.37-1.el7.centos.x86_64

from host ovirt01:

[root@ovirt01 ~]# hosted-engine --vm-status


--== Host 1 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : ovirt01.localdomain.local
Host ID: 1
Engine status  : {"reason": "vm not running on this
host", "health": "bad", "vm": "down", "detail": "unknown"}
Score  : 3352
stopped: False
Local maintenance  : False
crc32  : 256f2128
local_conf_timestamp   : 12251210
Host timestamp : 12251178
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=12251178 (Tue Nov 28 10:11:20 2017)
host-id=1
score=3352
vm_conf_refresh_time=12251210 (Tue Nov 28 10:11:52 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False


--== Host 2 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : 192.168.150.103
Host ID: 2
Engine status  : {"health": "good", "vm": "up",
"detail": "up"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 9b8c8a6c
local_conf_timestamp   : 12219386
Host timestamp : 12219357
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=12219357 (Tue Nov 28 10:11:23 2017)
host-id=2
score=3400
vm_conf_refresh_time=12219386 (Tue Nov 28 10:11:52 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineUp
stopped=False


--== Host 3 status ==--

conf_on_shared_storage : True
Status up-to-date  : True
Hostname   : ovirt03.localdomain.local
Host ID: 3
Engine status  : {"reason": "vm not running on this
host", "health": "bad", "vm": "down", "detail": "unknown"}
Score  : 3400
stopped: False
Local maintenance  : False
crc32  : 9f6399ef
local_conf_timestamp   : 2136
Host timestamp : 2136
Extra metadata (valid at timestamp):
metadata_parse_version=1
metadata_feature_version=1
timestamp=2136 (Tue Nov 28 10:11:56 2017)
host-id=3
score=3400
vm_conf_refresh_time=2136 (Tue Nov 28 10:11:56 2017)
conf_on_shared_storage=True
maintenance=False
state=EngineDown
stopped=False
[root@ovirt01 ~]#

Can I manually migrate engine vm to ovirt03?

[kasturi]

yes, definitely. You should be able to migrate.
hosted-engine --vm-status looks fine


On ovirt03:

[root@ovirt03 ~]# gluster volume info engine

Volume Name: engine
Type: Replicate

Re: [ovirt-users] ovirt 4.1, hosted engine and hyperconvergence

2017-11-26 Thread Kasturi Narra
Hello Gabriel,

Can you copy paste the contents of centos-base and ovirt repos ? this
is simply a result of mixed repos pkgs. Remove the -ev package and install
the -rhev version, that should take care of it.

Thanks
kasturi

On Thu, Nov 23, 2017 at 1:25 PM, Gabriel Stein 
wrote:

> Hi!
>
> I'm having problems to add new hosts on ovirt web interface and deploying
> the hosted engine.
>
> Everything works until yum try to install the needed packages.
>
>
> Error:
>
> Failed to install Host a.b.c. Yum ['qemu-kvm-tools-ev conflicts with
> 10:qemu-kvm-tools-1.5.3-141.el7_4.2.x86_64', 'qemu-kvm-tools conflicts
> with 10:qemu-kvm-tools-ev-2.9.0-16.el7_4.8.1.x86_64'].
>
> I already tried:
>
> - delete all repos and leave just ovirt and centos-base(doesn't works)
> - yum update(no updates)
>
> Some Data:
>
> CentOS Linux release 7.4.1708 (Core)
> oVirt Engine Version: 4.1.7.6-1.el7.centos
>
> Any Hints?
>
> Thanks!
>
> Best Regards,
>
> Gabriel
>
>
> Gabriel Stein
> --
> Gabriel Ferraz Stein
> Tel.: +49 (0)  170 2881531
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Huge Glsuter Issues - oVirt 4.1.7

2017-11-24 Thread Kasturi Narra
Hi Florian,

   Are you seeing these issues with gfapi or fuse access as well ?

Thanks
kasturi

On Fri, Nov 24, 2017 at 3:06 AM, Florian Nolden  wrote:

> I have the same issue when I run backup tasks during the night.
>
> I have a Gluster setup with a 1TB SSD on each of the tree nodes. Maybe its
> related to bug: https://bugzilla.redhat.com/show_bug.cgi?id=1430847
>
> sanlock.log:
> 2017-11-23 00:46:42 3410597 [1114]: s15 check_our_lease warning 60
> last_success 3410537
> 2017-11-23 00:46:43 3410598 [1114]: s15 check_our_lease warning 61
> last_success 3410537
> 2017-11-23 00:46:44 3410599 [1114]: s15 check_our_lease warning 62
> last_success 3410537
> 2017-11-23 00:46:45 3410600 [1114]: s15 check_our_lease warning 63
> last_success 3410537
> 2017-11-23 00:46:46 3410601 [1114]: s15 check_our_lease warning 64
> last_success 3410537
> 2017-11-23 00:46:47 3410602 [1114]: s15 check_our_lease warning 65
> last_success 3410537
> 2017-11-23 00:46:48 3410603 [1114]: s15 check_our_lease warning 66
> last_success 3410537
> 2017-11-23 00:46:49 3410603 [28384]: s15 delta_renew long write time 46 sec
> 2017-11-23 00:46:49 3410603 [28384]: s15 renewed 3410557 delta_length 46
> too long
> 2017-11-23 02:48:04 3417878 [28384]: s15 delta_renew long write time 10 sec
> 2017-11-23 02:57:23 3418438 [28384]: s15 delta_renew long write time 34 sec
> 2017-11-23 02:57:23 3418438 [28384]: s15 renewed 3418404 delta_length 34
> too long
>
>
> vdsm.log | grep "WARN"
> 017-11-23 00:20:05,544+0100 WARN  (jsonrpc/0) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') monitor became unresponsive
> (command timeout, age=63.719997) (vm:5109)
> 2017-11-23 00:20:06,840+0100 WARN  (check/loop) [storage.check] Checker
> u'/rhev/data-center/mnt/glusterSD/x-c01-n03:_fastIO/
> f0e21aae-1237-4dd3-88ec-81254d29c372/dom_md/metadata' is blocked for
> 10.00 seconds (check:279)
> 2017-11-23 00:20:13,853+0100 WARN  (periodic/170)
> [virt.periodic.VmDispatcher] could not run  'vdsm.virt.periodic.UpdateVolumes'>
> on [u'e1f26ea9-9294-4d9c-8f70-d59f96dec5f7'] (periodic:308)
> 2017-11-23 00:20:15,031+0100 WARN  (jsonrpc/2) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') monitor became unresponsive
> (command timeout, age=73.21) (vm:5109)
> 2017-11-23 00:20:20,586+0100 WARN  (jsonrpc/4) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') monitor became unresponsive
> (command timeout, age=78.759998) (vm:5109)
> 2017-11-23 00:21:06,849+0100 WARN  (check/loop) [storage.check] Checker
> u'/rhev/data-center/mnt/glusterSD/x-c01-n03:_fastIO/
> f0e21aae-1237-4dd3-88ec-81254d29c372/dom_md/metadata' is blocked for
> 10.01 seconds (check:279)
> 2017-11-23 00:21:13,847+0100 WARN  (periodic/167)
> [virt.periodic.VmDispatcher] could not run  'vdsm.virt.periodic.UpdateVolumes'>
> on [u'd8f22423-9fe3-4c06-97dc-5c9e9f5b33c8'] (periodic:308)
> 2017-11-23 00:22:13,854+0100 WARN  (periodic/172)
> [virt.periodic.VmDispatcher] could not run  'vdsm.virt.periodic.UpdateVolumes'>
> on [u'd8f22423-9fe3-4c06-97dc-5c9e9f5b33c8'] (periodic:308)
> 2017-11-23 00:22:16,846+0100 WARN  (check/loop) [storage.check] Checker
> u'/rhev/data-center/mnt/glusterSD/x-c01-n03:_fastIO/
> f0e21aae-1237-4dd3-88ec-81254d29c372/dom_md/metadata' is blocked for 9.99
> seconds (check:279)
> 2017-11-23 00:23:06,040+0100 WARN  (jsonrpc/6) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') monitor became unresponsive
> (command timeout, age=64.219997) (vm:5109)
> 2017-11-23 00:23:06,850+0100 WARN  (check/loop) [storage.check] Checker
> u'/rhev/data-center/mnt/glusterSD/x-c01-n03:_fastIO/
> f0e21aae-1237-4dd3-88ec-81254d29c372/dom_md/metadata' is blocked for 9.98
> seconds (check:279)
> 2017-11-23 00:23:13,845+0100 WARN  (periodic/169)
> [virt.periodic.VmDispatcher] could not run  'vdsm.virt.periodic.UpdateVolumes'>
> on [u'5ef506de-44b9-4ced-9b7f-b90ee098f4f7'] (periodic:308)
> 2017-11-23 00:23:16,855+0100 WARN  (jsonrpc/7) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') monitor became unresponsive
> (command timeout, age=75.030003) (vm:5109)
> 2017-11-23 00:23:21,082+0100 WARN  (jsonrpc/1) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') monitor became unresponsive
> (command timeout, age=79.259998) (vm:5109)
> 2017-11-23 00:25:31,488+0100 WARN  (libvirt/events) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') unknown eventid 8 args
> ('/rhev/data-center/0001-0001-0001-0001-0370/
> f0e21aae-1237-4dd3-88ec-81254d29c372/images/1a1b9620-
> 52fc-4008-9047-15cd725f8bd8/90b
> 913ba-e03f-46c5-bccf-bae011fcdd55', 4, 3, 8) (clientIF:549)
> 2017-11-23 00:25:32,372+0100 WARN  (libvirt/events) [virt.vm]
> (vmId='0a83954f-56d1-42d0-88b9-825435055fd0') unknown eventid 8 args
> ('/rhev/data-center/0001-0001-0001-0001-0370/
> f0e21aae-1237-4dd3-88ec-81254d29c372/images/1a1b9620-
> 52fc-4008-9047-15cd725f8bd8/90b
> 913ba-e03f-46c5-bccf-bae011fcdd55', 4, 0, 8) (clientIF:549)
> 2017-11-23 

Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2

2017-11-15 Thread Kasturi Narra
Hi Misak,

  I do not think we will be supporting any in the near future. Enabling
gfapi means that access mechanism of drive changes and i think we will need
to stop the vms.

Thanks
kasturi

On Wed, Nov 15, 2017 at 12:27 PM, Misak Khachatryan <kmi...@gmail.com>
wrote:

> Hi,
>
> will it be a more clean approach? I can't tolerate full stop of all
> VMs just to enable it, seems too disastrous for real production
> environment. Will it be some migration mechanisms in future?
>
> Best regards,
> Misak Khachatryan
>
>
> On Fri, Nov 10, 2017 at 12:35 AM, Darrell Budic <bu...@onholyground.com>
> wrote:
> > You do need to stop the VMs and restart them, not just issue a reboot. I
> > havn’t tried under 4.2 yet, but it works in 4.1.6 that way for me.
> >
> > 
> > From: Alessandro De Salvo <alessandro.desa...@roma1.infn.it>
> > Subject: Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2
> > Date: November 9, 2017 at 2:35:01 AM CST
> > To: users@ovirt.org
> >
> >
> > Hi again,
> >
> > OK, tried to stop all the vms, except the engine, set engine-config -s
> > LibgfApiSupported=true (for 4.2 only) and restarted the engine.
> >
> > When I tried restarting the VMs they are still not using gfapi, so it
> does
> > not seem to help.
> >
> > Cheers,
> >
> >
> > Alessandro
> >
> >
> >
> > Il 09/11/17 09:12, Alessandro De Salvo ha scritto:
> >
> > Hi,
> > where should I enable gfapi via the UI?
> > The only command I tried was engine-config -s LibgfApiSupported=true but
> the
> > result is what is shown in my output below, so it’s set to true for
> v4.2. Is
> > it enough?
> > I’ll try restarting the engine. Is it really needed to stop all the VMs
> and
> > restart them all? Of course this is a test setup and I can do it, but for
> > production clusters in the future it may be a problem.
> > Thanks,
> >
> >Alessandro
> >
> > Il giorno 09 nov 2017, alle ore 07:23, Kasturi Narra <kna...@redhat.com>
> ha
> > scritto:
> >
> > Hi ,
> >
> > The procedure to enable gfapi is below.
> >
> > 1) stop all the vms running
> > 2) Enable gfapi via UI or using engine-config command
> > 3) Restart ovirt-engine service
> > 4) start the vms.
> >
> > Hope you have not missed any !!
> >
> > Thanks
> > kasturi
> >
> > On Wed, Nov 8, 2017 at 11:58 PM, Alessandro De Salvo
> > <alessandro.desa...@roma1.infn.it> wrote:
> >>
> >> Hi,
> >>
> >> I'm using the latest 4.2 beta release and want to try the gfapi access,
> >> but I'm currently failing to use it.
> >>
> >> My test setup has an external glusterfs cluster v3.12, not managed by
> >> oVirt.
> >>
> >> The compatibility flag is correctly showing gfapi should be enabled with
> >> 4.2:
> >>
> >> # engine-config -g LibgfApiSupported
> >> LibgfApiSupported: false version: 3.6
> >> LibgfApiSupported: false version: 4.0
> >> LibgfApiSupported: false version: 4.1
> >> LibgfApiSupported: true version: 4.2
> >>
> >> The data center and cluster have the 4.2 compatibility flags as well.
> >>
> >> However, when starting a VM with a disk on gluster I can still see the
> >> disk is mounted via fuse.
> >>
> >> Any clue of what I'm still missing?
> >>
> >> Thanks,
> >>
> >>
> >>Alessandro
> >>
> >> ___
> >> Users mailing list
> >> Users@ovirt.org
> >> http://lists.ovirt.org/mailman/listinfo/users
> >
> >
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
> >
> >
> > ___
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> >
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] cannot remove gluster brick

2017-11-15 Thread Kasturi Narra
Hello Rudi,

 Removing a brick from a replica 3 volume means that you are
reducing the replica count from 3 to 2. You are seeing the first error
because when you are trying to remove a brick from replica 3 volume you do
not need to migrate data as the same data is present in other two replicate
sets. What you could simply do is 'gluster volume remove-brick data replica
2 srv1:/gluster/data/brick1 force' which removes it from the volume.

Hope this helps !!!

Thanks
kasturi



On Wed, Nov 15, 2017 at 5:13 PM, Rudi Ahlers  wrote:

> Hi,
>
> I am trying to remove a brick, from a server which is no longer part of
> the gluster pool, but I keep running into errors for which I cannot find
> answers on google.
>
> [root@virt2 ~]# gluster peer status
> Number of Peers: 3
>
> Hostname: srv1
> Uuid: 2bed7e51-430f-49f5-afbc-06f8cec9baeb
> State: Peer in Cluster (Disconnected)
>
> Hostname: srv3
> Uuid: 0e78793c-deca-4e3b-a36f-2333c8f91825
> State: Peer in Cluster (Connected)
>
> Hostname: srv4
> Uuid: 1a6eedc6-59eb-4329-b091-2b9bc6f0834f
> State: Peer in Cluster (Connected)
> [root@virt2 ~]#
>
>
>
>
> [root@virt2 ~]# gluster volume info data
>
> Volume Name: data
> Type: Replicate
> Volume ID: d09e4534-8bc0-4b30-be89-bc1ec2b439c7
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x 3 = 3
> Transport-type: tcp
> Bricks:
> Brick1: srv1:/gluster/data/brick1
> Brick2: srv2:/gluster/data/brick1
> Brick3: srv3:/gluster/data/brick1
> Options Reconfigured:
> nfs.disable: on
> transport.address-family: inet
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> performance.low-prio-threads: 32
> network.remote-dio: enable
> cluster.eager-lock: enable
> cluster.quorum-type: auto
> cluster.server-quorum-type: server
> cluster.data-self-heal-algorithm: full
> cluster.locking-scheme: granular
> cluster.shd-max-threads: 8
> cluster.shd-wait-qlength: 1
> features.shard: on
> user.cifs: off
> storage.owner-uid: 36
> storage.owner-gid: 36
> features.shard-block-size: 512MB
>
>
>
> [root@virt2 ~]# gluster volume remove-brick data replica 2
> srv1:/gluster/data/brick1 start
> volume remove-brick start: failed: Migration of data is not needed when
> reducing replica count. Use the 'force' option
>
>
> [root@virt2 ~]# gluster volume remove-brick data replica 2
> srv1:/gluster/data/brick1 commit
> Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y
> volume remove-brick commit: failed: Brick srv1:/gluster/data/brick1 is not
> decommissioned. Use start or force option
>
>
>
> The server virt1 is not part of the cluster anymore.
>
> --
> Kind Regards
> Rudi Ahlers
> Website: http://www.rudiahlers.co.za
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2

2017-11-14 Thread Kasturi Narra
Hi Bryan,

 In your output if you see the -drive file=gluster:///
this means that vm disk drives are being accessed using libgfapi.

If it is fuse then you would have seen something like
"file=/rhev/data-center/59f2df7e-0388-00ea-02c2-017b/67d7d3cc-df3f-4d07-b6f3-944982c5677c/images/8e6f96d3-2ed4-4c56-87d1-3a994284e683/9bf5a54e-d72d-4b1f-8ab1-0a84eb987fdd"

Thanks
kasturi

On Tue, Nov 14, 2017 at 8:37 PM, Bryan Sockel <bryan.soc...@altn.com> wrote:

> Hrm, not sure what i am doing wrong then, does not seem to be working for
> me.  I am not using the hosted engine, but a direct install on a physical
> server.  I thought i have enabled support for libgfapi with this command:
>
> # engine-config -g LibgfApiSupported
> LibgfApiSupported: false version: 3.6
> LibgfApiSupported: false version: 4.0
> LibgfApiSupported: true version: 4.1
>
> restarted the engine, shutdown the vm completely and started it back up a
> short time later.
>
> I am using this command to check:
>  ps ax | grep qemu | grep 'file=gluster\|file=/rhev'
>
> Output is
>  file=gluster://10.20.102.181/gl-vm12/....
>
> Thanks
> Bryan
>
> -Original Message-
> From: Kasturi Narra <kna...@redhat.com>
> To: Bryan Sockel <bryan.soc...@altn.com>
> Cc: Alessandro De Salvo <alessandro.desa...@roma1.infn.it>, users <
> users@ovirt.org>
> Date: Tue, 14 Nov 2017 12:56:49 +0530
> Subject: Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2
>
> yes, it  does work with 4.1.7.6 version
>
> On Tue, Nov 14, 2017 at 4:49 AM, Bryan Sockel <bryan.soc...@altn.com>
> wrote:
>>
>> Is libgfapi supposed to be working in 4.1.7.6?
>> Bryan
>>
>> -Original Message-
>> From: Alessandro De Salvo <alessandro.desa...@roma1.infn.it>
>> To: users@ovirt.org
>> Date: Thu, 9 Nov 2017 09:35:01 +0100
>> Subject: Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2
>>
>> Hi again,
>> OK, tried to stop all the vms, except the engine, set engine-config -s
>> LibgfApiSupported=true (for 4.2 only) and restarted the engine.
>> When I tried restarting the VMs they are still not using gfapi, so it
>> does not seem to help.
>> Cheers,
>>
>> Alessandro
>>
>>
>> Il 09/11/17 09:12, Alessandro De Salvo ha scritto:
>>
>>
>> Hi,
>> where should I enable gfapi via the UI?
>> The only command I tried was engine-config -s LibgfApiSupported=true but
>> the result is what is shown in my output below, so it’s set to true for
>> v4.2. Is it enough?
>> I’ll try restarting the engine. Is it really needed to stop all the VMs
>> and restart them all? Of course this is a test setup and I can do it, but
>> for production clusters in the future it may be a problem.
>> Thanks,
>>
>>Alessandro
>>
>> Il giorno 09 nov 2017, alle ore 07:23, Kasturi Narra <kna...@redhat.com>
>> ha scritto:
>>
>>
>> Hi ,
>>
>> The procedure to enable gfapi is below.
>>
>> 1) stop all the vms running
>> 2) Enable gfapi via UI or using engine-config command
>> 3) Restart ovirt-engine service
>> 4) start the vms.
>>
>> Hope you have not missed any !!
>>
>> Thanks
>> kasturi
>>
>> On Wed, Nov 8, 2017 at 11:58 PM, Alessandro De Salvo <
>> alessandro.desa...@roma1.infn.it> wrote:
>>>
>>> Hi,
>>>
>>> I'm using the latest 4.2 beta release and want to try the gfapi access,
>>> but I'm currently failing to use it.
>>>
>>> My test setup has an external glusterfs cluster v3.12, not managed by
>>> oVirt.
>>>
>>> The compatibility flag is correctly showing gfapi should be enabled with
>>> 4.2:
>>>
>>> # engine-config -g LibgfApiSupported
>>> LibgfApiSupported: false version: 3.6
>>> LibgfApiSupported: false version: 4.0
>>> LibgfApiSupported: false version: 4.1
>>> LibgfApiSupported: true version: 4.2
>>>
>>> The data center and cluster have the 4.2 compatibility flags as well.
>>>
>>> However, when starting a VM with a disk on gluster I can still see the
>>> disk is mounted via fuse.
>>>
>>> Any clue of what I'm still missing?
>>>
>>> Thanks,
>>>
>>>
>>>Alessandro
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
>> ___
>> Users mailing 
>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2

2017-11-13 Thread Kasturi Narra
yes, it  does work with 4.1.7.6 version

On Tue, Nov 14, 2017 at 4:49 AM, Bryan Sockel <bryan.soc...@altn.com> wrote:

> Is libgfapi supposed to be working in 4.1.7.6?
> Bryan
>
> -Original Message-
> From: Alessandro De Salvo <alessandro.desa...@roma1.infn.it>
> To: users@ovirt.org
> Date: Thu, 9 Nov 2017 09:35:01 +0100
> Subject: Re: [ovirt-users] Enabling libgfapi disk access with oVirt 4.2
>
> Hi again,
> OK, tried to stop all the vms, except the engine, set engine-config -s
> LibgfApiSupported=true (for 4.2 only) and restarted the engine.
> When I tried restarting the VMs they are still not using gfapi, so it does
> not seem to help.
> Cheers,
>
> Alessandro
>
>
> Il 09/11/17 09:12, Alessandro De Salvo ha scritto:
>
>
> Hi,
> where should I enable gfapi via the UI?
> The only command I tried was engine-config -s LibgfApiSupported=true but
> the result is what is shown in my output below, so it’s set to true for
> v4.2. Is it enough?
> I’ll try restarting the engine. Is it really needed to stop all the VMs
> and restart them all? Of course this is a test setup and I can do it, but
> for production clusters in the future it may be a problem.
> Thanks,
>
>Alessandro
>
> Il giorno 09 nov 2017, alle ore 07:23, Kasturi Narra <kna...@redhat.com>
> ha scritto:
>
>
> Hi ,
>
> The procedure to enable gfapi is below.
>
> 1) stop all the vms running
> 2) Enable gfapi via UI or using engine-config command
> 3) Restart ovirt-engine service
> 4) start the vms.
>
> Hope you have not missed any !!
>
> Thanks
> kasturi
>
> On Wed, Nov 8, 2017 at 11:58 PM, Alessandro De Salvo <
> alessandro.desa...@roma1.infn.it> wrote:
>>
>> Hi,
>>
>> I'm using the latest 4.2 beta release and want to try the gfapi access,
>> but I'm currently failing to use it.
>>
>> My test setup has an external glusterfs cluster v3.12, not managed by
>> oVirt.
>>
>> The compatibility flag is correctly showing gfapi should be enabled with
>> 4.2:
>>
>> # engine-config -g LibgfApiSupported
>> LibgfApiSupported: false version: 3.6
>> LibgfApiSupported: false version: 4.0
>> LibgfApiSupported: false version: 4.1
>> LibgfApiSupported: true version: 4.2
>>
>> The data center and cluster have the 4.2 compatibility flags as well.
>>
>> However, when starting a VM with a disk on gluster I can still see the
>> disk is mounted via fuse.
>>
>> Any clue of what I'm still missing?
>>
>> Thanks,
>>
>>
>>Alessandro
>>
>> ___
>> Users mailing list
>> Users@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
> ___
> Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine

2017-11-13 Thread Kasturi Narra
unreachable=0
>  failed=1
>
> Error: Ansible(>= 2.2) is not installed.
> Some of the features might not work if not installed.
>
>
> [root@ovirt2 ~]# yum info ansible
>
> Loaded plugins: fastestmirror, imgbased-persist
>
> Loading mirror speeds from cached hostfile
>
>  * epel: mirror01.idc.hinet.net
>
>  * ovirt-4.1: ftp.nluug.nl
>
>  * ovirt-4.1-epel: mirror01.idc.hinet.net
>
> Installed Packages
>
> Name: *ansible*
>
> Arch: noarch
>
> Version : 2.4.0.0
>
> Release : 5.el7
>
> Size: 38 M
>
> Repo: installed
>
> Summary : SSH-based configuration management, deployment, and task
> execution system
>
> URL : http://ansible.com
>
> License : GPLv3+
>
> Description :
>
> : Ansible is a radically simple model-driven configuration
> management,
>
> : multi-node deployment, and remote task execution system.
> Ansible works
>
> : over SSH and does not require any software or daemons to be
> installed
>
> : on remote nodes. Extension modules can be written in any
> language and
>
> : are transferred to managed machines automatically.
>
>
> root@ovirt2 ~]# yum info gdeploy
>
> Loaded plugins: fastestmirror, imgbased-persist
>
> Loading mirror speeds from cached hostfile
>
>  * epel: mirror01.idc.hinet.net
>
>  * ovirt-4.1: mirror.isoc.org.il
>
>  * ovirt-4.1-epel: mirror01.idc.hinet.net
>
> Installed Packages
>
> Name: gdeploy
>
> Arch: noarch
>
> Version : 2.0.2
>
> Release : 7
>
> Size: 2.7 M
>
> Repo: installed
>
> Summary : Tool to deploy and manage GlusterFS cluster
>
> URL : http://www.redhat.com/storage
>
> License : GPLv3
>
> Description : gdeploy is an Ansible based deployment tool. Initially
> gdeploy was written to
>
> : install GlusterFS clusters, eventually it grew out to do lot
> of other things. On
>
> : a given set of hosts, gdeploy can create physical volumes,
> volume groups, and
>
> : logical volumes, install packages, subscribe to RHN
> channels, run shell
>
> : commands, create GlusterFS volumes and lot more.
>
> :
>
> : See http://gdeploy.readthedocs.io/en/latest/ for more
> details
>
>
>
> Thanks Again,
>
> hk
>
>
> On Mon, Nov 13, 2017 at 3:17 PM, Kasturi Narra <kna...@redhat.com> wrote:
>
>> Hello,
>>
>> Can you please let me know which is the script it is failing  and
>> ansible and gdeploy versions?
>>
>> Thanks
>> kasturi
>>
>> On Mon, Nov 13, 2017 at 2:54 PM, Open tech <opentech...@gmail.com> wrote:
>>
>>> Hi All,
>>>I am new to Ovirt. I am hitting the exact same error while trying a
>>> new install in a nested virtualization setup on esxi 6.5.
>>> I am following this tutorial as well. Have three nodes on esxi with dual
>>> networks & passwordless ssh enabled.
>>> https://www.ovirt.org/blog/2017/04/up-and-running-with-ovirt
>>> -4.1-and-gluster-storage/
>>>
>>> Node install goes through without issue. Run into this error when i hit
>>> deploy.
>>>
>>> TASK [Run a shell script] **
>>> 
>>> fatal: [ovirt3]: FAILED! => {"failed": true, "msg": "The conditional
>>> check 'result.rc != 0' failed. The error was: error while evaluating
>>> conditional (result.rc != 0): 'dict object' has no attribute 'rc'"}
>>> fatal: [ovirt1]: FAILED! => {"failed": true, "msg": "The conditional
>>> check 'result.rc != 0' failed. The error was: error while evaluating
>>> conditional (result.rc != 0): 'dict object' has no attribute 'rc'"}
>>> fatal: [ovirt2]: FAILED! => {"failed": true, "msg": "The conditional
>>> check 'result.rc != 0' failed. The error was: error while evaluating
>>> conditional (result.rc != 0): 'dict object' has no attribute 'rc'"}
>>> to retry, use: --limit @/tmp/tmpbDBjAt/run-script.retry
>>>
>>>
>>> @Simone Marchioni were you able to find a solution ???.
>>>
>>> Thanks
>>> hk
>>>
>>>
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine

2017-11-13 Thread Kasturi Narra
Hello,

Can you please let me know which is the script it is failing  and
ansible and gdeploy versions?

Thanks
kasturi

On Mon, Nov 13, 2017 at 2:54 PM, Open tech  wrote:

> Hi All,
>I am new to Ovirt. I am hitting the exact same error while trying a new
> install in a nested virtualization setup on esxi 6.5.
> I am following this tutorial as well. Have three nodes on esxi with dual
> networks & passwordless ssh enabled.
> https://www.ovirt.org/blog/2017/04/up-and-running-with-
> ovirt-4.1-and-gluster-storage/
>
> Node install goes through without issue. Run into this error when i hit
> deploy.
>
> TASK [Run a shell script] **
> 
> fatal: [ovirt3]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt1]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> fatal: [ovirt2]: FAILED! => {"failed": true, "msg": "The conditional check
> 'result.rc != 0' failed. The error was: error while evaluating conditional
> (result.rc != 0): 'dict object' has no attribute 'rc'"}
> to retry, use: --limit @/tmp/tmpbDBjAt/run-script.retry
>
>
> @Simone Marchioni were you able to find a solution ???.
>
> Thanks
> hk
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine is down

2017-11-13 Thread Kasturi Narra
Hi Logan,

When i look at the hosted-engine --vm-status i see that vm is up but it
is health is bad. Can you try connecting to the vm using remote-viewer
using the command below ?

remote-viewer vnc://ovirttest1.wolfram.com:5900

Thanks
kasturi

On Fri, Nov 10, 2017 at 12:52 PM, Logan Kuhn  wrote:

> We lost the backend storage that hosts our self hosted engine tonight.
> We've recovered it and there was no data corruption on the volume
> containing the HE disk.  However, when we try to start the HE it doesn't
> give an error, but it also doesn't start.
>
> The VM isn't pingable and the liveliness check always fails.
>
>  [root@ovirttest1 ~]# hosted-engine --vm-status | grep -A20 ovirttest1
> Hostname   : ovirttest1.wolfram.com
> Host ID: 1
> Engine status  : {"reason": "failed liveliness check",
> "health": "bad", "vm": "up", "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 2c2f3ec9
> local_conf_timestamp   : 18980042
> Host timestamp : 18980039
> Extra metadata (valid at timestamp):
>metadata_parse_version=1
>metadata_feature_version=1
>timestamp=18980039 (Fri Nov 10 01:17:59 2017)
>host-id=1
>score=3400
>vm_conf_refresh_time=18980042 (Fri Nov 10 01:18:03 2017)
>conf_on_shared_storage=True
>maintenance=False
>state=GlobalMaintenance
>stopped=False
>
> The environment is in Global Maintenance so that we can isolate it to
> starting on a specific host to eliminate as many variables as possible.
> I've attached the agent and broker logs
>
> Regards,
> Logan Kuhn
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine installation + GlusterFS cluster

2017-11-10 Thread Kasturi Narra
Hello Artem,

May i know how did you deploy the Hosted Engine and glusterfs
volumes?  There is an easy way to do this using cockpit UI. You could log
into cockpit UI, click on Hosted Engine tab and there are two radio buttons
one for gluster deployment and another or HostedEngine deployment.

1) You could follow the gluster deployment screens which will create all
the volumes required for HC setup and opens all the ports
2) Then you can continue over to HE deployment and make sure you  provide
the answer as 'yes'  when the question 'Do you want to configure this host
and its cluster for gluster?' and answer no for iptables.
3) Once done with HE , you could go over to the UI, add the first master
domain (data) which imports hosted_storage and HE vm into the UI
automatically.
4) Then you can add the addtional hosts  & that is it, you are done.

Hope this helps !!!

Thanks
kasturi

On Thu, Nov 9, 2017 at 8:16 PM, Artem Tambovskiy  wrote:

> One more thing is  - firewall rules.
>
> For 3 gluster bricks I have configured following:
> firewall-cmd --zone=public --add-port=24007-24009/tcp
> --add-port=49152-49664/tcp --permanent
>
> and this seems not enough. have to stop the firewall in order to make the
> cluster working.
>
> I have noticed 490xx being used by gluster, any ideas on that documented
> range?
>
>  lsof -i | grep gluster | grep "490"
> glusterfs 32301root   10u  IPv4 148985  0t0  TCP
> ovirt1:49159->ovirt1:49099 (ESTABLISHED)
> glusterfs 32301root   17u  IPv4 153084  0t0  TCP
> ovirt1:49159->ovirt2:49096 (ESTABLISHED)
> glusterfs 46346root   17u  IPv4 156437  0t0  TCP
> ovirt1:49161->ovirt1:49093 (ESTABLISHED)
> glusterfs 46346root   18u  IPv4 149985  0t0  TCP
> ovirt1:49161->ovirt2:49090 (ESTABLISHED)
> glusterfs 46380root8u  IPv4 151389  0t0  TCP
> ovirt1:49090->ovirt3:49161 (ESTABLISHED)
> glusterfs 46380root   11u  IPv4 148986  0t0  TCP
> ovirt1:49091->ovirt2:49161 (ESTABLISHED)
> glusterfs 46380root   21u  IPv4 153074  0t0  TCP
> ovirt1:49099->ovirt1:49159 (ESTABLISHED)
> glusterfs 46380root   25u  IPv4 153075  0t0  TCP
> ovirt1:49097->ovirt2:49160 (ESTABLISHED)
> glusterfs 46380root   26u  IPv4 153076  0t0  TCP
> ovirt1:49095->ovirt3:49159 (ESTABLISHED)
> glusterfs 46380root   27u  IPv4 153077  0t0  TCP
> ovirt1:49093->ovirt1:49161 (ESTABLISHED)
>
> Regards,
> Artem
>
> On Thu, Nov 9, 2017 at 3:56 PM, Artem Tambovskiy <
> artem.tambovs...@gmail.com> wrote:
>
>> Hi,
>>
>> Just realized that I probably went in the wrong way. Reinstalled
>> everything from the scratch added 4 volumes (hosted_engine, data, export,
>> iso). All looks good so far.
>> But if go to the Cluster properties and tick the checkbox "Enable Cluster
>> Service" - the host will be marked as Non-Operational. Am I messing up the
>> things?
>> Or I'm just fine as long as I already have a Data (Master) Storage Domain
>> over GlusterFS?
>>
>> Regards,
>> Artem
>>
>> On Thu, Nov 9, 2017 at 2:46 PM, Fred Rolland  wrote:
>>
>>> Hi,
>>>
>>> The steps for this kind of setup are described in [1].
>>> However it seems you have already succeeded in installing, so maybe you
>>> need some additional steps [2]
>>> Did you add a storage domain that will act as Master Domain? It is
>>> needed, then the initial Storage Domain should be imported automatically.
>>>
>>>
>>> [1] https://www.ovirt.org/blog/2017/04/up-and-running-with-ovirt
>>> -4.1-and-gluster-storage/
>>> [2] https://www.ovirt.org/documentation/gluster-hyperconverged/c
>>> hap-Additional_Steps/
>>>
>>> On Thu, Nov 9, 2017 at 10:50 AM, Artem Tambovskiy <
>>> artem.tambovs...@gmail.com> wrote:
>>>
 Another yet attempt to get a help on hosted-engine deployment with
 glusterfs cluster.
 I already spend a day trying to get bring such a setup to work with no
 luck.

 The hosted engine being successfully deployed but I can't activate the
 host, the storage domain for the host is missing and I can't even add it.
 So either something went wrong during deployment or my glusterfs cluster
 doesn't configured properly.

 That are the prerequisites for this?

 - glusterfs cluster of 3 nodes with replica 3 volume
 - Any specific volume configs?
 - how many volumes should I prepare for hosted engine deployment?

 Any other thoughts?

 Regards,
 Artem

 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


>>>
>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine is down and won't start

2017-11-10 Thread Kasturi Narra
Hello Logan,

   One reason the liveliness check fails is host cannot ping your hosted
engine vm. you can try connecting to HE vm using remote-viewer
vnc://hypervisor-ip:5900 and from the hosted-engine --vm-status output
looks like the HE vm is up and running fine.


   - Please check internal dns setting like resolv.conf setting
   - Can not resolve virtual host name or ip address.

Thanks
kasturi


On Fri, Nov 10, 2017 at 12:56 PM, Logan Kuhn 
wrote:

> We lost the backend storage that hosts our self hosted engine tonight.
> We've recovered it and there was no data corruption on the volume
> containing the HE disk.  However, when we try to start the HE it doesn't
> give an error, but it also doesn't start.
>
> The VM isn't pingable and the liveliness check always fails.
>
>  [root@ovirttest1 ~]# hosted-engine --vm-status | grep -A20 ovirttest1
> Hostname   : ovirttest1.wolfram.com
> Host ID: 1
> Engine status  : {"reason": "failed liveliness check",
> "health": "bad", "vm": "up", "detail": "up"}
> Score  : 3400
> stopped: False
> Local maintenance  : False
> crc32  : 2c2f3ec9
> local_conf_timestamp   : 18980042
> Host timestamp : 18980039
> Extra metadata (valid at timestamp):
>metadata_parse_version=1
>metadata_feature_version=1
>timestamp=18980039 (Fri Nov 10 01:17:59 2017)
>host-id=1
>score=3400
>vm_conf_refresh_time=18980042 (Fri Nov 10 01:18:03 2017)
>conf_on_shared_storage=True
>maintenance=False
>state=GlobalMaintenance
>stopped=False
>
> The environment is in Global Maintenance so that we can isolate it to
> starting on a specific host to eliminate as many variables as possible.
> I've attached the agent and broker logs
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Cannot uplaod ot ISO domain

2017-10-23 Thread Kasturi Narra
Hi Alex,

  Can you check if you have the following on your setup ?

1) gluster volume which will be used as ISO storage domain should have
bricks connected using glusternw.
2) NFS time out is caused due to not having nfs.disable off on the volume.
Can you try to set this and try again.
3) self hosted engine requires an additional NIC on the gluster network for
the engine-iso-uploader to work sucessfully  if ovirtmgmt and glusternw are
on a different subnet.

Even after having all the above if things do not work, you can just scp the
file to /mnt//images/11--
-…/

Hope this helps !!!

Thanks
kasturi


On Thu, Oct 19, 2017 at 1:23 PM, Alex K  wrote:

> The only workaround that seems to work for my case is the following:
>
> Enable NFS on ISO gluster volume.
> Add ISO volume as NFS using hostname of server on the network that engine
> has access.
> This is not optimal as HA is not achieve -  if I loose that server then I
> will need to redefine the ISO domain.
>
> Alex
>
>
> On Thu, Oct 19, 2017 at 12:09 AM, Alex K  wrote:
>
>> If this is the case then I will try to set the gluster network to be at
>> the network that engine can reach prior to adding the ISO domain. I would
>> prefere not to add an additional interface to engine as then I will need to
>> make the storage network a bridge which might incur some overhead and
>> affect performance of storage network.
>>
>> Alex
>>
>> On Oct 18, 2017 11:36 PM, "Elad Ben Aharon"  wrote:
>>
>>> Hi,
>>>
>>> The storage where 'ISO' storage domain resides has to be reachable for
>>> the engine server. Please check network connectivity between the two.
>>>
>>> On Wed, Oct 18, 2017 at 7:41 PM, Alex K  wrote:
>>>
 Hi all,

 I have setup a self hosted setup with 3 servers with several volumes in
 replica 3.
 These gluster volumes are set on a separate network (on separate NICs
 on same servers), dedicated for gluster traffic. The engine does not have
 any NIC in this network.
 The ISO storage domain is also on top one of these gluster volumes.
 Hope this makes sense.

 When I try to upload an ISO file from engine, I get the following
 error:

 engine-iso-uploader -i ISO upload /usr/share/ovirt-guest-tools-i
 so/oVirt-toolsSetup_4.1-3.fc24.iso
 Please provide the REST API password for the admin@internal oVirt
 Engine user (CTRL+D to abort):
 Uploading, please wait...
 ERROR: mount.nfs: No route to host

 Seems that the engine is trying to mount the volume and gives a no
 route to host.

 What is the correct procedure to add an ISO domain? Can't ISO domain be
 on a separate dedicated network which is reachable from each host?

 Thanx,
 Alex


 ___
 Users mailing list
 Users@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/users


>>>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine storage on management IP

2017-10-18 Thread Kasturi Narra
yes, it will do that. Once you have the steps below done you should see
that hosted-storage get sync on gluster network.

1) create a new network with name 'blah'
2) Assign glusternw role to the newly created network
3) Now drag and drop gluster network to the interface with which gluster
trusted pool is connected.

Now when you go to volumes tab and look at hosted-storage volume brick
directory should have the path something like ':/

Hope this helps !!

Thanks
kasturi

On Wed, Oct 18, 2017 at 2:31 PM, TranceWorldLogic . <
tranceworldlo...@gmail.com> wrote:

> Hi,
>
> I saw that vdsm is updating gluster cluster with gluster network IP.
> (in file /var/lib/glusterd/peers/ I can see hostname2 entry)
>
> Will this also make sure hosted-storage get sync on gluster network ?
> Please help me.
>
> Thanks,
> ~Rohit
>
> On Wed, Oct 18, 2017 at 12:39 PM, TranceWorldLogic . <
> tranceworldlo...@gmail.com> wrote:
>
>> "That will take care of syncing the bricks in the UI of hosted-storage
>> gluster volume on gluster network."
>> Sorry, not understood, what do you mean UI here. Would you please explain
>> in detail ?
>>
>> On Wed, Oct 18, 2017 at 12:33 PM, Kasturi Narra <kna...@redhat.com>
>> wrote:
>>
>>> Hi,
>>>
>>> You can assign glusternw role to the newly created gluster network
>>>  and associate the interface with which you have configured gluster pool
>>> with gluster network. That will take care of syncing the bricks in the UI
>>> of hosted-storage gluster volume on gluster network.
>>>
>>> Thanks
>>> kasturi
>>>
>>> On Tue, Oct 17, 2017 at 9:59 PM, TranceWorldLogic . <
>>> tranceworldlo...@gmail.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> I was trying to setup separate gluster network in ovirt.
>>>> But I saw hosted-storage volume which I had created for hosted-engine
>>>> still running on managment network.
>>>>
>>>> Can I move that hosted-storage gluster volume on gluster network ?
>>>>
>>>> Thanks,
>>>> ~Rohit
>>>>
>>>> ___
>>>> Users mailing list
>>>> Users@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted Engine storage on management IP

2017-10-18 Thread Kasturi Narra
Hi,

You can assign glusternw role to the newly created gluster network  and
associate the interface with which you have configured gluster pool with
gluster network. That will take care of syncing the bricks in the UI of
hosted-storage gluster volume on gluster network.

Thanks
kasturi

On Tue, Oct 17, 2017 at 9:59 PM, TranceWorldLogic . <
tranceworldlo...@gmail.com> wrote:

> Hi,
>
> I was trying to setup separate gluster network in ovirt.
> But I saw hosted-storage volume which I had created for hosted-engine
> still running on managment network.
>
> Can I move that hosted-storage gluster volume on gluster network ?
>
> Thanks,
> ~Rohit
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fwd: Engine crash, storage won't activate, hosts won't shutdown, template locked, gpu passthrough failed

2017-10-09 Thread Kasturi Narra
On Sat, Sep 30, 2017 at 7:50 PM M R  wrote:

> Hello!
>
> I have been using Ovirt for last four weeks, testing and trying to get
> things working.
>
> I have collected here the problems I have found and this might be a bit
> long but help to any of these or maybe to all of them from several people
> would be wonderful.
>
> My version is ovirt node 4.1.5 and 4.1.6 downloaded from website latest
> stable release at the time. Also tested with CentOS minimal +ovirt repo. In
> this case, 3. is solved, but other problems persist.
>
>
> 1. Power off host
> First day after installing ovirt node, it was able to reboot and shutdown
> clean. No problems at all. After few days of using ovir, I have noticed
> that hosts are unable to shutdown. I have tested this in several different
> ways and come to the following conclusion. IF engine has not been started
> after boot, all hosts are able to shutdown clean. But if engine is started
> even once, none of the hosts are able to shutdown anymore. The only way to
> get power off is to unplug or press power button for a longer time as hard
> reset. I have failed to find a way to have the engine running and then
> shutdown host. This effects to all hosts in the cluster.
>
> 2. Glusterfs failed
> Every time I have booted hosts, glusterfs has failed. For some reason, it
> turns inactive state even if I have setup systemctl enable glusterd. Before
> this command it was just inactive. After this command, it will say "failed
> (inactive). There is still a way to get glusterfs working. I have to give
> command systemctl start glusterd manually and everything starts working.
> Why do I have to give manual commands to start glusterfs? I have used this
> for CentOS before and never had this problem before. Node installer is that
> much different from the CentOS core?
>

You need to fix this first, since this may cause the issues you report
later about
storage domains not activating.

You should ask about this in gluster mailing list.
Adding Sahina to advice on this.

There is a problem where glusterd fails to start  on nodes where glusterd
tries to come up even before network is up. You can find more info at bug
https://bugzilla.redhat.com/show_bug.cgi?id=1472267


>
> 3. Epel
> As I said that I have used CentOS before, I would like to able to install
> some packets from repo. But even if I install epel-release, it won't find
> packets such as nano or htop. I have read about how to add epel-release to
> ovirt node from here: https://www.ovirt.org/release/4.1.1/#epel
> I have tested even manually edit repolist, but it will fail to find normal
> epel packets. I have setup additional exclude=collectd* as guided in the
> link above. This doesn't make any difference. All being said I am able to
> install manually packets which are downloaded with other CentOS machine and
> transferred with scp to ovirt node. Still, this once again needs a lot of
> manual input and is just a workaround for the bug.
>
> 4.  Engine startup
> When I try to start the engine when glusterfs is up, it will say vm
> doesn't exist, starting up. Still, it won't startup automatically. I have
> to give several times command hosted-engine --vm-start. I wait for about
> 5minutes until I give it next time. This will take usually about 30minutes
> and then randomly. Completely randomly after one of the times, I give this
> command engine shoots up and is up in 1minute. This has happened every time
> I boot up. And the times that I have to give a command to start the engine,
> has been changing. At best it's been 3rd time at worst it has been 7th
> time. Calculating from there it might take from 15minutes to 35minutes to
> get the engine up.Nevertheless, it will eventually come up every time. If
> there is a way to get it up on the first try or even better, automatically
> up, it would be great.
>
> 5. Activate storage
> Once the engine is up, there has been a problem with storage. When I go to
> storage tab, it will show all sources red. Even if I wait for 15~20minutes,
> it won't get storage green itself. I have to go and press active button
> from main data storage. Then it will get main storage up in
> 2~3munutes.Sometimes it fails it once, but will definitely get main data
> storage up on the seconds try. And then magically at the same time all
> other storages instantly go green. Main storage is glusterfs and I have 3
> NFS storages as well. This is only a problem when starting up and once
> storages are on green they stay green. Still annoying that it cannot get it
> done by itself.
>

This may be related to the glusterfs issue (2).

When the nodes gets rebooted, gluster bricks goes down. If more than one
brick in the volume goes down quorum is lost and storage domain associated
goes to inactive state.

 Once two or more bricks are up and brick status gets synced to UI, storage
domains should automatically turn to active state and no issues should  be
seen there.

Does it happen after you fix 

Re: [ovirt-users] Maximum storage per VM?

2017-10-09 Thread Kasturi Narra
Hi Dmitri,

  If the vms are created on a Hyperconverged setup then the max disk
size  recommended is 2TB.

Thanks
kasturi.

On Sat, Oct 7, 2017 at 12:54 AM, Dmitri Chebotarov  wrote:

> Hello
>
> I'm trying to find any info on how much storage I can attach to a VM.
>
> Is there a recommended/maximum for number of disks and maximum disk size?
>
> I'm using GlusterFS as backend storage for the cluster.
>
> The VM uses LVM (/w striped volumes) to manage attached disks.
>
> Thank you,
> --
> Dmitri Chebotarov.
> George Mason University,
> 4400 University Drive,
> Fairfax, VA, 22030
> GPG Public key# 5E19F14D: [https://goo.gl/SlE8tj]
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Dead hosts

2017-10-09 Thread Kasturi Narra
Hi,

   You can run the command below which will remove these hosts from the
hosted-engine --vm-status output

‘hosted-engine --clean-metadata --host-id= --force-clean’

Thanks
kasturi



On Sun, Oct 8, 2017 at 12:01 AM, Maton, Brett 
wrote:

> Hi,
>
>   I've replaced some hardware, and didn't remote 'hosted engine deploy'
> before retireing the servers.
>   (actually I failed to remove them properly...)
>
>   How can I get rid of the 'old' hosts from the output of hosted-engine
> --vm-status ?
>
>
>   oVirt 4.2 - pre
>   centOS 7.4
>
> Regards,
> Brett
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed gdeploy

2017-09-25 Thread Kasturi Narra
Hi Sean,

 This error is expected and there is a bug to change this script so
that it can  properly disable multipath devices. To continue you can simply
add 'ignore_script_errors=yes'  under script3 which continues with this
failure. Please note that this script is used to disable multipath devices
and proceeding with this script simply means that you will have multipath
enabled for all your devices.

[script3]

file=/usr/share/gdeploy/scripts/disable-multipath.sh
ignore_script_errors=yes

Thanks
kasturi

On Fri, Sep 22, 2017 at 9:41 PM, Sean McMurray  wrote:

> After reading https://bugzilla.redhat.com/show_bug.cgi?id=1469469 and
> https://bugzilla.redhat.com/show_bug.cgi?id=1491548#c2 I changed
> gdeployConfig.conf so that [script3] has file=/usr/share/gdeploy/script
> s/disable-multipath.sh
>
> That gives me a different failture:
>
> TASK [Run a shell script] **
> 
> **
> task path: /tmp/tmp4kfKbY/run-script.yml:7
> failed: [192.168.1.3] (item=/usr/share/gdeploy/scripts/disable-multipath.sh)
> => {"changed": true, "failed": true, "failed_when_result": true, "item":
> "/usr/share/gdeploy/scripts/disable-multipath.sh", "rc": 1, "stderr":
> "Shared connection to 192.168.1.3 closed.\r\n", "stdout": "iscsiadm: No
> active sessions.\r\nThis script will prevent listing iscsi devices when
> multipath CLI is called\r\nwithout parameters, and so no LUNs will be
> discovered by applications like VDSM\r\n(oVirt, RHV) which shell-out to
> call `/usr/sbin/multipath` after target login\r\nSep 22 08:58:47 | DM
> multipath kernel driver not loaded\r\nSep 22 08:58:47 | DM multipath kernel
> driver not loaded\r\n", "stdout_lines": ["iscsiadm: No active sessions.",
> "This script will prevent listing iscsi devices when multipath CLI is
> called", "without parameters, and so no LUNs will be discovered by
> applications like VDSM", "(oVirt, RHV) which shell-out to call
> `/usr/sbin/multipath` after target login", "Sep 22 08:58:47 | DM multipath
> kernel driver not loaded", "Sep 22 08:58:47 | DM multipath kernel driver
> not loaded"]}
> failed: [192.168.1.2] (item=/usr/share/gdeploy/scripts/disable-multipath.sh)
> => {"changed": true, "failed": true, "failed_when_result": true, "item":
> "/usr/share/gdeploy/scripts/disable-multipath.sh", "rc": 1, "stderr":
> "Shared connection to 192.168.1.2 closed.\r\n", "stdout": "iscsiadm: No
> active sessions.\r\nThis script will prevent listing iscsi devices when
> multipath CLI is called\r\nwithout parameters, and so no LUNs will be
> discovered by applications like VDSM\r\n(oVirt, RHV) which shell-out to
> call `/usr/sbin/multipath` after target login\r\nSep 22 15:57:47 | DM
> multipath kernel driver not loaded\r\nSep 22 15:57:47 | DM multipath kernel
> driver not loaded\r\n", "stdout_lines": ["iscsiadm: No active sessions.",
> "This script will prevent listing iscsi devices when multipath CLI is
> called", "without parameters, and so no LUNs will be discovered by
> applications like VDSM", "(oVirt, RHV) which shell-out to call
> `/usr/sbin/multipath` after target login", "Sep 22 15:57:47 | DM multipath
> kernel driver not loaded", "Sep 22 15:57:47 | DM multipath kernel driver
> not loaded"]}
> failed: [192.168.1.1] (item=/usr/share/gdeploy/scripts/disable-multipath.sh)
> => {"changed": true, "failed": true, "failed_when_result": true, "item":
> "/usr/share/gdeploy/scripts/disable-multipath.sh", "rc": 1, "stderr":
> "Shared connection to 192.168.1.1 closed.\r\n", "stdout": "iscsiadm: No
> active sessions.\r\nThis script will prevent listing iscsi devices when
> multipath CLI is called\r\nwithout parameters, and so no LUNs will be
> discovered by applications like VDSM\r\n(oVirt, RHV) which shell-out to
> call `/usr/sbin/multipath` after target login\r\nSep 22 08:58:50 | DM
> multipath kernel driver not loaded\r\nSep 22 08:58:50 | DM multipath kernel
> driver not loaded\r\n", "stdout_lines": ["iscsiadm: No active sessions.",
> "This script will prevent listing iscsi devices when multipath CLI is
> called", "without parameters, and so no LUNs will be discovered by
> applications like VDSM", "(oVirt, RHV) which shell-out to call
> `/usr/sbin/multipath` after target login", "Sep 22 08:58:50 | DM multipath
> kernel driver not loaded", "Sep 22 08:58:50 | DM multipath kernel driver
> not loaded"]}
> to retry, use: --limit @/tmp/tmp4kfKbY/run-script.retry
>
> PLAY RECAP 
> ***
> 192.168.1.1: ok=0changed=0unreachable=0 failed=1
> 192.168.1.2: ok=0changed=0unreachable=0 failed=1
> 192.168.1.3: ok=0changed=0unreachable=0 failed=1
>
>
> On 09/22/2017 09:05 AM, Sean McMurray wrote:
>
>> My latest attempt to deploy went like this (/tmp/tmpaQJuTG/run-script.yml
>> 

Re: [ovirt-users] Odp: Re: How to upgrade self-hosted engine?

2017-09-20 Thread Kasturi Narra
Hi,

   upgrade HE (Hosted Engine ) by doing the steps below.

1) Move HE to global maintenance by running the command 'hosted-engine
--set-maintenance --mode=global'
2) Add the required repos which has higher package versions.
3) Run 'yum update ovirt\*setup\*'
4) engine-setup
5) Once the setup is done successfully please move HE out of global
maintenance mode by running the command 'hosted-engine --set-maintenance
--mode=none'

For upgrade node you can do the steps below:

1) If there is an upgrade link present in the 'General' tab of the UI, (if
ovirt node) click upgrade from UI which will move the host to maintenance
and run upgrade on the system, reboots and brings it back up to Normal
state.

2) If you do not see the upgrade link, then add the repos to which you
would want to update the node.

3) Run 'yum update'

4) Reboot the system.

5) make sure that you are on the latest image by running 'imgbase w' if it
is a ovirt node.

6)otherwise just make sure that you are on the latest kernel once your
upgrade finishes.

Hope this helps 

Thanks
kasturi

On Wed, Sep 20, 2017 at 6:34 PM, gabriel_skup...@o2.pl <
gabriel_skup...@o2.pl> wrote:

> To be honest, I don't see consistency regarding all that upgrade
> procedures. Have a look here:
> https://www.ovirt.org/documentation/upgrade-guide/chap-Updating_the_oVirt_
> Environment/
> It asks to update the system too :)
>
> I will appreciate if somebody clearly explains:
> -to upgrade the Node - please do this and this
> -to upgrade the Hosted-Engine - please do this and this.
>
> G.
>
> Dnia 20 września 2017 14:52 Johan Bernhardsson 
> napisał(a):
>
> If you follow that guide it will update everything on the virtual server
> running the engine. You should not run yum to upgrade the packages
> (correct me anyone if i am wrong about this)
>
> The engine-setup will download and install the packages for you.
>
> And on the nodes  upgrade first via the web interface and after that you
> can upgrade with yum on the nodes. Don't forget to have the node in
> maintence mode when you run yum update
>
> /Johan
> 2017-09-20 14:39 skrev gabriel_skup...@o2.pl:
>
> Thanks. What about the system itself?
>
> Is yum update enough?
>
> Dnia 20 września 2017 13:25 Johan Bernhardsson 
> napisał(a):
>
> Follow this guide if it is between minor releases
>
>
> https://www.ovirt.org/documentation/upgrade-guide/
> chap-Updates_between_Minor_Releases/
>
>
>
> Don't forget to send the hosted-engine to global maintenance
>
> /Johan
>
> On September 20, 2017 13:11:41 gabriel_skup...@o2.pl wrote:
>
> In oVirt Engine Web Administration portal I can see option to
> upgrade the nodes but can't see any option to upgrade
> hosted-engine itself?
>
> What is the recommended procedure for it?
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> --
> Security all the way ...
>
> Linux/CMS/Network/Print/Virtualisation/VoIP Consultant
>
> Kafit AB
> Orgnr:  556792-5945
> Mobile: +46705111751
> Sweden: +46101993005
> Seychelles: +2486478105
> Uk: +448701821792
> Email:  jo...@kafit.se
> Web:http://www.kafit.se
>
> Connect with me on LinkedIn: http://www.linkedin.com/in/smallone
>
> About me: http://about.me/smallone/bio
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Hosted-engine --deploy 4.1.6 - Web GUI Down

2017-09-20 Thread Kasturi Narra
Hi,

 can you check what 'hosted-engine --vm-status' reports and can you
check if you are able to ping the hostname of HE vm from your hosts and the
machine where you are trying to access the browser ?

Thanks
kasturi

On Tue, Sep 19, 2017 at 7:47 PM, Mat Gomes 
wrote:

> Hi Guys,
>
>
>
> I’m attempting to rebuild my environment for production testing, We’ve
> multiple locations NY and CH with a 2+1arbiter setup, this will run
> Geo-Replication.
>
>
>
> All goes smoothly, VM starts and I’m able to login onto it via SSH, I
> usually stop Firewalld as it usually causes the Web-GUI to be unreachable
>
> This time around the Web GUI can’t be reached, I’ve done more than 10
> deployments and tests but I’m not sure what’s causing this, 4.1.5 was
> deployed same exact way aswell. it all seems to be running fine, port 80 is
> listening, I’ve seem most of the logs but nothing stands out.
>
>
>
> Please help.
>
>
>
>
>
> Steps:
>
> #Puppet Installation/deployment.
>
> class ovirt {
>
>
>
> package { 'centos-release-gluster310':
>
> ensure => installed,
>
> }
>
> package { 'ovirt-release41-4.1.6-1.el7.centos.noarch':
>
> ensure  => installed,
>
> source => "http://resources.ovirt.org/
> pub/yum-repo/ovirt-release41.rpm",
>
> provider => 'rpm',
>
> install_options => ['--nosignature'],
>
> }
>
> $ovirt_packages = ['system-storage-manager','
> vdsm-gluster','ovirt-hosted-engine-setup','ovirt-engine-
> appliance','glusterfs-server',]
>
>
>
> package { $ovirt_packages:
>
> ensure => installed,
>
> require => [ Package['centos-release-gluster310','ovirt-release41-
> 4.1.6-1.el7.centos.noarch']],
>
> install_options => ['--disablerepo=epel'],
>
> }
>
> service { 'glusterd':
>
> ensure => running,
>
> enable => true,
>
> require => [ Package['vdsm-gluster'],
> File['/etc/glusterfs/glusterd.vol']],
>
> }
>
> file { '/etc/glusterfs/glusterd.vol':
>
> ensure  => file,
>
> source  => 'puppet:///modules/ovirt/glusterd.vol',
>
> owner   => root,
>
> group   => root,
>
> mode=> '0644',
>
> }  <- All works fine.
>
>
>
> Once everything is up/running deployed, peer-probed, ssh-keys are
> created/shared, volumes settings are set and started:
>
> gluster volume create engine replica 3 arbiter 1
> host1:/data/glusterfs/vol1/engine host2:/data/glusterfs/vol1/engine
> host3:/data/glusterfs/vol1/engine
>
> gluster volume set engine cluster.quorum-type auto
>
> gluster volume set engine network.ping-timeout 10
>
> gluster volume set engine auth.allow \*
>
> gluster volume set engine group virt
>
> gluster volume set engine storage.owner-uid 36
>
> gluster volume set engine storage.owner-gid 36
>
> gluster volume set engine server.allow-insecure on
>
> gluster volume start engine
>
>
>
> At last hosted-engine --deploy is ran,   --> https://pastebin.com/QXEwmSwT
> <--   answer file
>
>
>
> VM starts but no WEB GUI.
>
> Let me know if you need more info
>
>
>
> Best Regards,
>
> *Mat Gomes* | Assistant Vice President, IT
> *t*. 212-531-8594*  m*. 954-254-1294
> *e*. mgo...@clearpoolgroup.com
> *w*. clearpoolgroup.com 
>
>
>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Ovirt HC engine will not start

2017-09-18 Thread Kasturi Narra
Hi,

   In the agent.log file i see that it fails to connect to storage server.
Can you please check the following .

1) can you please check 'gluster peer status' and make sure that all the
nodes in the cluster are connected ?

2) can you gluster volume status  and make sure that at
least two of the bricks are online?

Thanks
kasturi

On Fri, Sep 15, 2017 at 12:29 PM, Jesper Andersson <
jesper.anders...@viconia.se> wrote:

> Hi.
>
> I have a HC environment (oVirt 4.1.4) where the engine does not start.
> Tryed to update it to the latest oVirt but something happend on the
> storage side.
> Seemed like there was no data on the second node so we did a full heal,
> restarted the environment and now the ha-agent, broker and the engine
> does not start.
> "The hosted engine configuration has not been retrieved from shared
> storage. Please ensure that ovirt-ha-agent is running and the storage
> server is reachable."
>
> Tried to start the ha-agent but receives this error:
> ovirt-ha-agent ovirt_hosted_engine_ha.agent.agent.Agent ERROR Traceback
> (most recent call last):
>   File
> "/usr/lib/python2.7/site-packages/ovirt_hosted_engine_ha/agent/agent.py",
> line 191, in _run_agent
>
> The broker wont start either:
> ovirt-ha-broker ovirt_hosted_engine_ha.broker.listener.ConnectionHandler
> ERROR Error handling request, data: 'set-storage-domain
> FilesystemBackend dom_type=glusterfs
> sd_uuid=2a79ea65-3df3-43cf-b1ba-304581f7918a'
> Traceback (most recent call last):
> File "/usr/lib/python2.7/site-packages/ovirt...
>
> I'm not that experienced with ovirt nor gluster storage but gluster
> seems OK as far as I can see.
> Does anyone have an idea of how to resolve this?
>
> Please see attached logs.
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] gluster replica volumes with different data usage

2017-09-18 Thread Kasturi Narra
Recommended would be creating a new storage domain with shard size as 64 MB
and migrating all the disks from 4MB storagedomain

On Mon, Sep 18, 2017 at 12:01 PM, Ravishankar N 
wrote:

> Possibly.  I don't think changing shard size on the fly is supported,
> especially when there are files on the volume that are sharded with a
> different size.
>
> -Ravi
>
>
> On 09/18/2017 11:40 AM, Alex K wrote:
>
> The heal status is showing that no pending files need healing (also shown
> at GUI).
> When checking the bricks on the file system I see that what is different
> between the server is the .shard folder of the volume. One server reports
> 835GB while the other 1.1 TB.
> I recall to have changed the shard size at some point from 4 MB to 64MB.
> Could this be the cause?
>
> Thanx,
> Alex
>
> On Mon, Sep 18, 2017 at 8:14 AM, Ravishankar N 
> wrote:
>
>>
>> On 09/18/2017 10:08 AM, Alex K wrote:
>>
>> Hi Ravishankar,
>>
>> I am not referring to the arbiter volume(which is showing 0% usage). I am
>> referring to the other 2 volumes which are replicas and should have the
>> exact same data. Checking the status of other bricks in ovirt (bricks used
>> from iso and export domain) I see that they all report same usage of data
>> on the data volumes, except the "vms" volume used for storing vms.
>>
>>
>> Ah, okay.  Some of the things that can cause a variation in disk usage:
>> - Pending self-heals in gluster (check if `gluster volume heal 
>> info` doesn't show any entries.  Also if there is anything under
>> `.glusterfs/landfill` folder of the bricks).
>> - XFS speculative preallocation
>> - Possibly some bug in self-healing of sparse files by gluster (although
>> we fixed known bugs a long time back in this area).
>>
>> Regards
>> Ravi
>>
>>
>> Thanx,
>> Alex
>>
>> On Sep 18, 2017 07:00, "Ravishankar N"  wrote:
>>
>>>
>>>
>>> On 09/17/2017 08:41 PM, Alex K wrote:
>>>
>>> Hi all,
>>>
>>> I have replica 3 with 1 arbiter.
>>> When checking the gluster volume bricks they are reported as using
>>> different space, as per attached. How come they use different space? One
>>> would expect to use exactly the same space since they are replica.
>>>
>>> The 3rd brick (arbiter ) only holds meta data, so it would not consume
>>> as much space as the other 2 data bricks. So what you are seeing is
>>> expected behaviour.
>>> Regards,
>>> Ravi
>>>
>>> Thanx,
>>> Alex
>>>
>>>
>>> ___
>>> Users mailing 
>>> listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>>
>>
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Recommended gluster setup

2017-09-15 Thread Kasturi Narra
Hi ,

   yes, you are right. Since arbiter brick  has only metadata and  data for
the vm has to be served from one of the other two replicas, read is slow.

Arbiter is a special subset of replica 3 volumes and is aimed at
preventing split-brains and providing same consistency as a normal replica
3 volume with out consuming 3x space. You could use replica 3 and no issues
with that.

Thanks
kasturi

On Fri, Sep 15, 2017 at 12:41 PM, Abi Askushi 
wrote:

> Hi all,
>
> I see in the ovirt guides that a gluster volume replica 3 with 1 arbiter
> is recommended.
> Why not simple replica 3? Is it due to the higher replication data that
> would cause performance issues?
>
> What I am observing is that a VM running on the server which has the
> arbiter brick has slower read performance then when the same VM runs on
> another server with a normal brick. Has anyone observed this? Is it because
> the arbiter does not have the real data on it?
>
> Thanx,
> Alex
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] First oVirt engine deploy: missing gateway on hosts

2017-09-04 Thread Kasturi Narra
Hi Mauro,

 yes, i do know this problem of gateway disappearing but i am not sure
why that happens because in my case it was happening on only on one server.
I used to add the gateway again on that node. I am just waiting for some
one to reply.

Thanks
kasturi

On Mon, Sep 4, 2017 at 1:02 PM, Mauro Tridici <mauro.trid...@cmcc.it> wrote:

> Hi Kasturi,
>
> thank you very much for your support and for the useful explanation.
> If possible, I would ask you if you know a way to correct the “missing
> gateway” problem that I described in my previous message.
>
> Thanks a lot,
> Mauro
>
>
> Il giorno 04 set 2017, alle ore 08:51, Kasturi Narra <kna...@redhat.com>
> ha scritto:
>
> Hi Mauro,
>
>Creating distributed dispersed volumes are not supported from ovirt
> UI yet but you should be able to sync them if cluster is imported into the
> UI. same holds true for add / remove bricks on disperse and distribute
> disperse volumes.
>
>you wont be able to see bricks created because ovirt excepts them
> to be mounted at /rhgs/. What you could simply do is uncheck
> the check box in the 'Add brick' dialog and type in the path or you could
> mount your bricks at above said location and that would show all the
> available bricks on the host.
>
>   Hope this helps.
>
> Thanks
> kasturi.
>
> On Sat, Sep 2, 2017 at 7:16 PM, Mauro Tridici <mauro.trid...@cmcc.it>
> wrote:
>
>> Hi all,
>>
>> I just started my first Ovirt Engine deploy using a dedicated (and
>> separated) virtual machine.
>> I’m trying to create and manage a test Gluster cluster using 3 “virtual”
>> hosts (hostnames are glu01, glu02, glu03)
>> 2 different networks have been defined on the hosts (192.168.213.0/24
>> for management network and 192.168.152.0/24 for gluster network).
>> Ovirt engine deploy completed without any problem, the hosts have been
>> added easily using ovirtmgmt network (bridgeless mgmt network) and
>> ovirtgluster (bridgeless gluster network).
>>
>> Everything seems to be ok for this first deploy, but I just noticed that
>> the gateway is missing on the target hosts:
>>
>> [root@glu01 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse
>> Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00
>> ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00
>> ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
>> ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
>> ens33
>>
>> [root@glu02 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse
>> Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00
>> ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00
>> ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
>> ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
>> ens33
>>
>> [root@glu03 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse
>> Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00
>> ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00
>> ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
>> ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
>> ens33
>>
>> Due to this problem I cannot reach internet from ens33 nic (management
>> network).
>> I just tried to add the gateway in ifcfg-ens33 configuration file but
>> gateway disappear after host reboot.
>>
>> [root@glu01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-ens33
>> # Generated by VDSM version 4.19.28-1.el7.centos
>> DEVICE=ens33
>> ONBOOT=yes
>> IPADDR=192.168.213.151
>> NETMASK=255.255.255.0
>> BOOTPROTO=none
>> MTU=1500
>> DEFROUTE=no
>> NM_CONTROLLED=no
>> IPV6INIT=yes
>> IPV6_AUTOCONF=yes
>>
>> The oVirt Engine network configuration is the following one:
>>
>> [host glu01]
>> ens33 -> ovirtmgmt (192.168.213.151, 255.255.255.0, 192.168.213.2)
>> ens34 -> ovirtgluster (192.168.152.151, 255.255.255.0)
>>
>> [host glu02]
>> ens33 -> ovirtmgmt (192.168.213.152, 255.255.255.0, 192.168.213.2)
>> ens34 -> ovirtgluster (192.168.1

Re: [ovirt-users] First oVirt engine deploy: missing gateway on hosts

2017-09-04 Thread Kasturi Narra
Hi Mauro,

   Creating distributed dispersed volumes are not supported from ovirt
UI yet but you should be able to sync them if cluster is imported into the
UI. same holds true for add / remove bricks on disperse and distribute
disperse volumes.

   you wont be able to see bricks created because ovirt excepts them to
be mounted at /rhgs/. What you could simply do is uncheck the
check box in the 'Add brick' dialog and type in the path or you could mount
your bricks at above said location and that would show all the available
bricks on the host.

  Hope this helps.

Thanks
kasturi.

On Sat, Sep 2, 2017 at 7:16 PM, Mauro Tridici  wrote:

> Hi all,
>
> I just started my first Ovirt Engine deploy using a dedicated (and
> separated) virtual machine.
> I’m trying to create and manage a test Gluster cluster using 3 “virtual”
> hosts (hostnames are glu01, glu02, glu03)
> 2 different networks have been defined on the hosts (192.168.213.0/24 for
> management network and 192.168.152.0/24 for gluster network).
> Ovirt engine deploy completed without any problem, the hosts have been
> added easily using ovirtmgmt network (bridgeless mgmt network) and
> ovirtgluster (bridgeless gluster network).
>
> Everything seems to be ok for this first deploy, but I just noticed that
> the gateway is missing on the target hosts:
>
> [root@glu01 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse
> Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00
> ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00
> ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
> ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
> ens33
>
> [root@glu02 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse
> Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00
> ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00
> ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
> ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
> ens33
>
> [root@glu03 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse
> Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00
> ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00
> ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
> ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
> ens33
>
> Due to this problem I cannot reach internet from ens33 nic (management
> network).
> I just tried to add the gateway in ifcfg-ens33 configuration file but
> gateway disappear after host reboot.
>
> [root@glu01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-ens33
> # Generated by VDSM version 4.19.28-1.el7.centos
> DEVICE=ens33
> ONBOOT=yes
> IPADDR=192.168.213.151
> NETMASK=255.255.255.0
> BOOTPROTO=none
> MTU=1500
> DEFROUTE=no
> NM_CONTROLLED=no
> IPV6INIT=yes
> IPV6_AUTOCONF=yes
>
> The oVirt Engine network configuration is the following one:
>
> [host glu01]
> ens33 -> ovirtmgmt (192.168.213.151, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.151, 255.255.255.0)
>
> [host glu02]
> ens33 -> ovirtmgmt (192.168.213.152, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.152, 255.255.255.0)
>
> [host glu03]
> ens33 -> ovirtmgmt (192.168.213.153, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.153, 255.255.255.0)
>
> Do you know the right way to set the gateway IP on all hosts?
>
> Just two last questions: I was able to import an existing gluster cluster
> using oVirt Engine, but I’m not able to create a new volume because:
>
> - I can’t select a distributed disperse volume configuration from oVirt
> Engine volume creation window
> - i can’t see the bricks to be used to create a new volume (but I can
> import an existing volume without problem).
>
> Is there something that I can do to resolve the issues and complete my
> first experience with oVirt?
>
> Thank you very much,
> Mauro T.
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hyperconverged question

2017-09-04 Thread Kasturi Narra
y dropping from 2 -> 1. You would need 4 node with 1 
>>>>>>>>>> being
>>>>>>>>>> arbiter to configure quorum which is in effect 3 writable nodes and 1
>>>>>>>>>> arbiter. If one gluster node drops, you still have 2 up. Although in 
>>>>>>>>>> this
>>>>>>>>>> case, you probably wouldnt need arbiter at all
>>>>>>>>>>
>>>>>>>>>> If you are configured, you can drop quorum settings and just let
>>>>>>>>>> arbiter run since you're not using arbiter node in your VM cluster 
>>>>>>>>>> part (I
>>>>>>>>>> believe), just storage cluster part. When using quorum, you need > 
>>>>>>>>>> 50% of
>>>>>>>>>> the cluster being up at one time. Since you have 3 nodes with 1 
>>>>>>>>>> arbiter,
>>>>>>>>>> you're actually losing 1/2 which == 50 which == degraded / hindered 
>>>>>>>>>> gluster
>>>>>>>>>>
>>>>>>>>>> Again, this is to the best of my knowledge based on other quorum
>>>>>>>>>> backed softwareand this is what I understand from testing with 
>>>>>>>>>> gluster
>>>>>>>>>> and ovirt thus far
>>>>>>>>>>
>>>>>>>>>> On Fri, Sep 1, 2017 at 11:53 AM, Jim Kusznir <j...@palousetech.com
>>>>>>>>>> > wrote:
>>>>>>>>>>
>>>>>>>>>>> Huh...Ok., how do I convert the arbitrar to full replica, then?
>>>>>>>>>>> I was misinformed when I created this setup.  I thought the 
>>>>>>>>>>> arbitrator held
>>>>>>>>>>> enough metadata that it could validate or refudiate  any one 
>>>>>>>>>>> replica (kinda
>>>>>>>>>>> like the parity drive for a RAID-4 array).  I was also under the 
>>>>>>>>>>> impression
>>>>>>>>>>> that one replica  + Arbitrator is enough to keep the array online 
>>>>>>>>>>> and
>>>>>>>>>>> functional.
>>>>>>>>>>>
>>>>>>>>>>> --Jim
>>>>>>>>>>>
>>>>>>>>>>> On Fri, Sep 1, 2017 at 5:22 AM, Charles Kozler <
>>>>>>>>>>> ckozler...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> @ Jim - you have only two data volumes and lost quorum.
>>>>>>>>>>>> Arbitrator only stores metadata, no actual files. So yes, you were 
>>>>>>>>>>>> running
>>>>>>>>>>>> in degraded mode so some operations were hindered.
>>>>>>>>>>>>
>>>>>>>>>>>> @ Sahina - Yes, this actually worked fine for me once I did
>>>>>>>>>>>> that. However, the issue I am still facing, is when I go to create 
>>>>>>>>>>>> a new
>>>>>>>>>>>> gluster storage domain (replica 3, hyperconverged) and I tell it 
>>>>>>>>>>>> "Host to
>>>>>>>>>>>> use" and I select that host. If I fail that host, all VMs halt. I 
>>>>>>>>>>>> do not
>>>>>>>>>>>> recall this in 3.6 or early 4.0. This to me makes it seem like 
>>>>>>>>>>>> this is
>>>>>>>>>>>> "pinning" a node to a volume and vice versa like you could, for 
>>>>>>>>>>>> instance,
>>>>>>>>>>>> for a singular hyperconverged to ex: export a local disk via NFS 
>>>>>>>>>>>> and then
>>>>>>>>>>>> mount it via ovirt domain. But of course, this has its caveats. To 
>>>>>>>>>>>> that
>>>>>>>>>>>> end, I am using gluster replica 3, when configuring it I say "host 
>>>>>>>>>>>> to use:
>>>>&

Re: [ovirt-users] hyperconverged question

2017-09-04 Thread Kasturi Narra
node with 1 being 
>>>>>>> arbiter
>>>>>>> to configure quorum which is in effect 3 writable nodes and 1 arbiter. 
>>>>>>> If
>>>>>>> one gluster node drops, you still have 2 up. Although in this case, you
>>>>>>> probably wouldnt need arbiter at all
>>>>>>>
>>>>>>> If you are configured, you can drop quorum settings and just let
>>>>>>> arbiter run since you're not using arbiter node in your VM cluster part 
>>>>>>> (I
>>>>>>> believe), just storage cluster part. When using quorum, you need > 50% 
>>>>>>> of
>>>>>>> the cluster being up at one time. Since you have 3 nodes with 1 arbiter,
>>>>>>> you're actually losing 1/2 which == 50 which == degraded / hindered 
>>>>>>> gluster
>>>>>>>
>>>>>>> Again, this is to the best of my knowledge based on other quorum
>>>>>>> backed softwareand this is what I understand from testing with 
>>>>>>> gluster
>>>>>>> and ovirt thus far
>>>>>>>
>>>>>>> On Fri, Sep 1, 2017 at 11:53 AM, Jim Kusznir <j...@palousetech.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Huh...Ok., how do I convert the arbitrar to full replica, then?  I
>>>>>>>> was misinformed when I created this setup.  I thought the arbitrator 
>>>>>>>> held
>>>>>>>> enough metadata that it could validate or refudiate  any one replica 
>>>>>>>> (kinda
>>>>>>>> like the parity drive for a RAID-4 array).  I was also under the 
>>>>>>>> impression
>>>>>>>> that one replica  + Arbitrator is enough to keep the array online and
>>>>>>>> functional.
>>>>>>>>
>>>>>>>> --Jim
>>>>>>>>
>>>>>>>> On Fri, Sep 1, 2017 at 5:22 AM, Charles Kozler <
>>>>>>>> ckozler...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> @ Jim - you have only two data volumes and lost quorum. Arbitrator
>>>>>>>>> only stores metadata, no actual files. So yes, you were running in 
>>>>>>>>> degraded
>>>>>>>>> mode so some operations were hindered.
>>>>>>>>>
>>>>>>>>> @ Sahina - Yes, this actually worked fine for me once I did that.
>>>>>>>>> However, the issue I am still facing, is when I go to create a new 
>>>>>>>>> gluster
>>>>>>>>> storage domain (replica 3, hyperconverged) and I tell it "Host to 
>>>>>>>>> use" and
>>>>>>>>> I select that host. If I fail that host, all VMs halt. I do not 
>>>>>>>>> recall this
>>>>>>>>> in 3.6 or early 4.0. This to me makes it seem like this is "pinning" 
>>>>>>>>> a node
>>>>>>>>> to a volume and vice versa like you could, for instance, for a 
>>>>>>>>> singular
>>>>>>>>> hyperconverged to ex: export a local disk via NFS and then mount it 
>>>>>>>>> via
>>>>>>>>> ovirt domain. But of course, this has its caveats. To that end, I am 
>>>>>>>>> using
>>>>>>>>> gluster replica 3, when configuring it I say "host to use: " node 1, 
>>>>>>>>> then
>>>>>>>>> in the connection details I give it node1:/data. I fail node1, all VMs
>>>>>>>>> halt. Did I miss something?
>>>>>>>>>
>>>>>>>>> On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose <sab...@redhat.com>
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> To the OP question, when you set up a gluster storage domain, you
>>>>>>>>>> need to specify backup-volfile-servers=: where
>>>>>>>>>> server2 and server3 also have bricks running. When server1 is down, 
>>>>>>>>>> and the
>>>>>>>>>> volume is mounted again - server2 or server3 are qu

Re: [ovirt-users] hyperconverged question

2017-09-01 Thread Kasturi Narra
yes, that is the same option i was asking about. Apologies that i had
mentioned a different name.

So, ovirt will automatically detect it if you select the option 'use
managed gluster volume'. While adding a storage domain after specifying the
host , you could just select the checkbox and that will list all the
volumes managed from ovirt UI + that will fill the mount options for you.



On Fri, Sep 1, 2017 at 6:40 PM, Charles Kozler <ckozler...@gmail.com> wrote:

> Are you referring to "Mount Options" - > http://i.imgur.com/bYfbyzz.png
>
> Then no, but that would explain why it wasnt working :-). I guess I had a
> silly assumption that oVirt would have detected it and automatically taken
> up the redundancy that was configured inside the replica set / brick
> detection.
>
> I will test and let you know
>
> Thanks!
>
> On Fri, Sep 1, 2017 at 8:52 AM, Kasturi Narra <kna...@redhat.com> wrote:
>
>> Hi Charles,
>>
>>   One question, while configuring a storage domain  you are
>> saying "host to use: " node1,  then in the connection details you say
>> node1:/data. What about the backup-volfile-servers option in the UI while
>> configuring storage domain? Are you specifying that too?
>>
>> Thanks
>> kasturi
>>
>>
>> On Fri, Sep 1, 2017 at 5:52 PM, Charles Kozler <ckozler...@gmail.com>
>> wrote:
>>
>>> @ Jim - you have only two data volumes and lost quorum. Arbitrator only
>>> stores metadata, no actual files. So yes, you were running in degraded mode
>>> so some operations were hindered.
>>>
>>> @ Sahina - Yes, this actually worked fine for me once I did that.
>>> However, the issue I am still facing, is when I go to create a new gluster
>>> storage domain (replica 3, hyperconverged) and I tell it "Host to use" and
>>> I select that host. If I fail that host, all VMs halt. I do not recall this
>>> in 3.6 or early 4.0. This to me makes it seem like this is "pinning" a node
>>> to a volume and vice versa like you could, for instance, for a singular
>>> hyperconverged to ex: export a local disk via NFS and then mount it via
>>> ovirt domain. But of course, this has its caveats. To that end, I am using
>>> gluster replica 3, when configuring it I say "host to use: " node 1, then
>>> in the connection details I give it node1:/data. I fail node1, all VMs
>>> halt. Did I miss something?
>>>
>>> On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose <sab...@redhat.com> wrote:
>>>
>>>> To the OP question, when you set up a gluster storage domain, you need
>>>> to specify backup-volfile-servers=: where server2
>>>> and server3 also have bricks running. When server1 is down, and the volume
>>>> is mounted again - server2 or server3 are queried to get the gluster
>>>> volfiles.
>>>>
>>>> @Jim, if this does not work, are you using 4.1.5 build with libgfapi
>>>> access? If not, please provide the vdsm and gluster mount logs to analyse
>>>>
>>>> If VMs go to paused state - this could mean the storage is not
>>>> available. You can check "gluster volume status " to see if
>>>> atleast 2 bricks are running.
>>>>
>>>> On Fri, Sep 1, 2017 at 11:31 AM, Johan Bernhardsson <jo...@kafit.se>
>>>> wrote:
>>>>
>>>>> If gluster drops in quorum so that it has less votes than it should it
>>>>> will stop file operations until quorum is back to normal.If i rember it
>>>>> right you need two bricks to write for quorum to be met and that the
>>>>> arbiter only is a vote to avoid split brain.
>>>>>
>>>>>
>>>>> Basically what you have is a raid5 solution without a spare. And when
>>>>> one disk dies it will run in degraded mode. And some raid systems will 
>>>>> stop
>>>>> the raid until you have removed the disk or forced it to run anyway.
>>>>>
>>>>> You can read up on it here: https://gluster.readthed
>>>>> ocs.io/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
>>>>>
>>>>> /Johan
>>>>>
>>>>> On Thu, 2017-08-31 at 22:33 -0700, Jim Kusznir wrote:
>>>>>
>>>>> Hi all:
>>>>>
>>>>> Sorry to hijack the thread, but I was about to start essentially the
>>>>> same thread.
>>>>>
>>>>> I have a 3 node cluster, all three are hosts

Re: [ovirt-users] hyperconverged question

2017-09-01 Thread Kasturi Narra
Hi Charles,

  One question, while configuring a storage domain  you are saying
"host to use: " node1,  then in the connection details you say node1:/data.
What about the backup-volfile-servers option in the UI while configuring
storage domain? Are you specifying that too?

Thanks
kasturi


On Fri, Sep 1, 2017 at 5:52 PM, Charles Kozler <ckozler...@gmail.com> wrote:

> @ Jim - you have only two data volumes and lost quorum. Arbitrator only
> stores metadata, no actual files. So yes, you were running in degraded mode
> so some operations were hindered.
>
> @ Sahina - Yes, this actually worked fine for me once I did that. However,
> the issue I am still facing, is when I go to create a new gluster storage
> domain (replica 3, hyperconverged) and I tell it "Host to use" and I select
> that host. If I fail that host, all VMs halt. I do not recall this in 3.6
> or early 4.0. This to me makes it seem like this is "pinning" a node to a
> volume and vice versa like you could, for instance, for a singular
> hyperconverged to ex: export a local disk via NFS and then mount it via
> ovirt domain. But of course, this has its caveats. To that end, I am using
> gluster replica 3, when configuring it I say "host to use: " node 1, then
> in the connection details I give it node1:/data. I fail node1, all VMs
> halt. Did I miss something?
>
> On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose <sab...@redhat.com> wrote:
>
>> To the OP question, when you set up a gluster storage domain, you need to
>> specify backup-volfile-servers=: where server2 and
>> server3 also have bricks running. When server1 is down, and the volume is
>> mounted again - server2 or server3 are queried to get the gluster volfiles.
>>
>> @Jim, if this does not work, are you using 4.1.5 build with libgfapi
>> access? If not, please provide the vdsm and gluster mount logs to analyse
>>
>> If VMs go to paused state - this could mean the storage is not available.
>> You can check "gluster volume status " to see if atleast 2 bricks
>> are running.
>>
>> On Fri, Sep 1, 2017 at 11:31 AM, Johan Bernhardsson <jo...@kafit.se>
>> wrote:
>>
>>> If gluster drops in quorum so that it has less votes than it should it
>>> will stop file operations until quorum is back to normal.If i rember it
>>> right you need two bricks to write for quorum to be met and that the
>>> arbiter only is a vote to avoid split brain.
>>>
>>>
>>> Basically what you have is a raid5 solution without a spare. And when
>>> one disk dies it will run in degraded mode. And some raid systems will stop
>>> the raid until you have removed the disk or forced it to run anyway.
>>>
>>> You can read up on it here: https://gluster.readthed
>>> ocs.io/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
>>>
>>> /Johan
>>>
>>> On Thu, 2017-08-31 at 22:33 -0700, Jim Kusznir wrote:
>>>
>>> Hi all:
>>>
>>> Sorry to hijack the thread, but I was about to start essentially the
>>> same thread.
>>>
>>> I have a 3 node cluster, all three are hosts and gluster nodes (replica
>>> 2 + arbitrar).  I DO have the mnt_options=backup-volfile-servers= set:
>>>
>>> storage=192.168.8.11:/engine
>>> mnt_options=backup-volfile-servers=192.168.8.12:192.168.8.13
>>>
>>> I had an issue today where 192.168.8.11 went down.  ALL VMs immediately
>>> paused, including the engine (all VMs were running on host2:192.168.8.12).
>>> I couldn't get any gluster stuff working until host1 (192.168.8.11) was
>>> restored.
>>>
>>> What's wrong / what did I miss?
>>>
>>> (this was set up "manually" through the article on setting up
>>> self-hosted gluster cluster back when 4.0 was new..I've upgraded it to 4.1
>>> since).
>>>
>>> Thanks!
>>> --Jim
>>>
>>>
>>> On Thu, Aug 31, 2017 at 12:31 PM, Charles Kozler <ckozler...@gmail.com>
>>> wrote:
>>>
>>> Typo..."Set it up and then failed that **HOST**"
>>>
>>> And upon that host going down, the storage domain went down. I only have
>>> hosted storage domain and this new one - is this why the DC went down and
>>> no SPM could be elected?
>>>
>>> I dont recall this working this way in early 4.0 or 3.6
>>>
>>> On Thu, Aug 31, 2017 at 3:30 PM, Charles Kozler <ckozler...@gmail.com>
>>> wrote:
>>>
>>> So I've tested this today and I failed a nod

Re: [ovirt-users] hyperconverged question

2017-08-31 Thread Kasturi Narra
yes, right.  What you can do is edit the hosted-engine.conf file and there
is a parameter as shown below [1] and replace h2 and h3 with your second
and third storage servers. Then you will need to restart ovirt-ha-agent and
ovirt-ha-broker services in all the nodes .

[1] 'mnt_options=backup-volfile-servers=:'

On Thu, Aug 31, 2017 at 5:54 PM, Charles Kozler <ckozler...@gmail.com>
wrote:

> Hi Kasturi -
>
> Thanks for feedback
>
> > If cockpit+gdeploy plugin would be have been used then that would have
> automatically detected glusterfs replica 3 volume created during Hosted
> Engine deployment and this question would not have been asked
>
> Actually, doing hosted-engine --deploy it too also auto detects
> glusterfs.  I know glusterfs fuse client has the ability to failover
> between all nodes in cluster, but I am still curious given the fact that I
> see in ovirt config node1:/engine (being node1 I set it to in hosted-engine
> --deploy). So my concern was to ensure and find out exactly how engine
> works when one node goes away and the fuse client moves over to the other
> node in the gluster cluster
>
> But you did somewhat answer my question, the answer seems to be no (as
> default) and I will have to use hosted-engine.conf and change the parameter
> as you list
>
> So I need to do something manual to create HA for engine on gluster? Yes?
>
> Thanks so much!
>
> On Thu, Aug 31, 2017 at 3:03 AM, Kasturi Narra <kna...@redhat.com> wrote:
>
>> Hi,
>>
>>During Hosted Engine setup question about glusterfs volume is being
>> asked because you have setup the volumes yourself. If cockpit+gdeploy
>> plugin would be have been used then that would have automatically detected
>> glusterfs replica 3 volume created during Hosted Engine deployment and this
>> question would not have been asked.
>>
>>During new storage domain creation when glusterfs is selected there is
>> a feature called 'use managed gluster volumes' and upon checking this all
>> glusterfs volumes managed will be listed and you could choose the volume of
>> your choice from the dropdown list.
>>
>> There is a conf file called /etc/hosted-engine/hosted-engine.conf
>> where there is a parameter called backup-volfile-servers="h1:h2" and if one
>> of the gluster node goes down engine uses this parameter to provide ha /
>> failover.
>>
>>  Hope this helps !!
>>
>> Thanks
>> kasturi
>>
>>
>>
>> On Wed, Aug 30, 2017 at 8:09 PM, Charles Kozler <ckozler...@gmail.com>
>> wrote:
>>
>>> Hello -
>>>
>>> I have successfully created a hyperconverged hosted engine setup
>>> consisting of 3 nodes - 2 for VM's and the third purely for storage. I
>>> manually configured it all, did not use ovirt node or anything. Built the
>>> gluster volumes myself
>>>
>>> However, I noticed that when setting up the hosted engine and even when
>>> adding a new storage domain with glusterfs type, it still asks for
>>> hostname:/volumename
>>>
>>> This leads me to believe that if that one node goes down (ex:
>>> node1:/data), then ovirt engine wont be able to communicate with that
>>> volume because its trying to reach it on node 1 and thus, go down
>>>
>>> I know glusterfs fuse client can connect to all nodes to provide
>>> failover/ha but how does the engine handle this?
>>>
>>> ___
>>> Users mailing list
>>> Users@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hyperconverged question

2017-08-31 Thread Kasturi Narra
Hi,

   During Hosted Engine setup question about glusterfs volume is being
asked because you have setup the volumes yourself. If cockpit+gdeploy
plugin would be have been used then that would have automatically detected
glusterfs replica 3 volume created during Hosted Engine deployment and this
question would not have been asked.

   During new storage domain creation when glusterfs is selected there is a
feature called 'use managed gluster volumes' and upon checking this all
glusterfs volumes managed will be listed and you could choose the volume of
your choice from the dropdown list.

There is a conf file called /etc/hosted-engine/hosted-engine.conf where
there is a parameter called backup-volfile-servers="h1:h2" and if one of
the gluster node goes down engine uses this parameter to provide ha /
failover.

 Hope this helps !!

Thanks
kasturi



On Wed, Aug 30, 2017 at 8:09 PM, Charles Kozler 
wrote:

> Hello -
>
> I have successfully created a hyperconverged hosted engine setup
> consisting of 3 nodes - 2 for VM's and the third purely for storage. I
> manually configured it all, did not use ovirt node or anything. Built the
> gluster volumes myself
>
> However, I noticed that when setting up the hosted engine and even when
> adding a new storage domain with glusterfs type, it still asks for
> hostname:/volumename
>
> This leads me to believe that if that one node goes down (ex:
> node1:/data), then ovirt engine wont be able to communicate with that
> volume because its trying to reach it on node 1 and thus, go down
>
> I know glusterfs fuse client can connect to all nodes to provide
> failover/ha but how does the engine handle this?
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Test Email

2017-08-29 Thread Kasturi Narra
Hi,

This is a test email and please ignore the mail

Thanks
kasturi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hosted engine setup with Gluster fail

2017-08-28 Thread Kasturi Narra
can you please check if you have any additional disk in the system? If you
have additional disk in the system other than the disk which is being used
for root partition then you could specify the disk in the cockpit UI (i
hope you are using cockpit UI to do the installation) with no partitions on
that. That will take care of the installation and make your life easier as
cockpit + gdeploy would take care of configuring gluster bricks and volumes
for you.

On Mon, Aug 28, 2017 at 2:55 PM, Anzar Esmail Sainudeen <
an...@it.thumbay.com> wrote:

> Dear Nara,
>
>
>
> All the partitions, pv and vg are created automatically during the initial
> setup time.
>
>
>
> [root@ovirtnode1 ~]# vgs
>
>   VG  #PV #LV #SN Attr   VSize   VFree
>
>   onn   1  12   0 wz--n- 555.73g 14.93g
>
>
>
> All space are mounted to the below location, all free space are mounted in
> /.
>
>
>
> Filesystem  Size  Used Avail
> Use% Mounted on
>
> /dev/mapper/onn-ovirt--node--ng--4.1.4--0.20170728.0+1  513G  4.2G
> 483G   1% /
>
> devtmpfs 44G 0
> 44G   0% /dev
>
> tmpfs44G  4.0K
> 44G   1% /dev/shm
>
> tmpfs44G   33M
> 44G   1% /run
>
> tmpfs44G 0
> 44G   0% /sys/fs/cgroup
>
> /dev/sda2   976M  135M  774M
> 15% /boot
>
> /dev/mapper/onn-home976M  2.6M
> 907M   1% /home
>
> /dev/mapper/onn-tmp 2.0G  6.3M
> 1.8G   1% /tmp
>
> /dev/sda1   200M  9.5M
> 191M   5% /boot/efi
>
> /dev/mapper/onn-var  15G  1.8G   13G
> 13% /var
>
> /dev/mapper/onn-var--log7.8G  224M
> 7.2G   3% /var/log
>
> /dev/mapper/onn-var--log--audit 2.0G   44M
> 1.8G   3% /var/log/audit
>
> tmpfs   8.7G 0
> 8.7G   0% /run/user/0
>
>
>
> If we need any space we want to reduce the vg size and create new
> one.(This is correct)
>
>
>
>
>
> If the above step is complicated, can you please suggest to setup
> glusterfs datastore in ovirt
>
>
>
> Anzar Esmail Sainudeen
>
> Group Datacenter Incharge| IT Infra Division | Thumbay Group
>
> P.O Box : 4184 | Ajman | United Arab Emirates.
>
> Mobile: 055-8633699|Tel: 06 7431333 |  Extn :1303
>
> Email: an...@it.thumbay.com | Website: www.thumbay.com
>
> [image: cid:image001.jpg@01D18D9D.15A17620]
>
>
>
> Disclaimer: This message contains confidential information and is intended
> only for the individual named. If you are not the named addressee, you are
> hereby notified that disclosing, copying, distributing or taking any action
> in reliance on the contents of this e-mail is strictly prohibited. Please
> notify the sender immediately by e-mail if you have received this e-mail by
> mistake, and delete this material. Thumbay Group accepts no liability for
> errors or omissions in the contents of this message, which arise as a
> result of e-mail transmission.
>
>
>
> *From:* Kasturi Narra [mailto:kna...@redhat.com]
> *Sent:* Monday, August 28, 2017 1:14 PM
>
> *To:* Anzar Esmail Sainudeen
> *Cc:* users
> *Subject:* Re: [ovirt-users] hosted engine setup with Gluster fail
>
>
>
> yes, you can create. I do not see any problems there.
>
>
>
> May i know how these vgs are created ? If they are not created using
> gdeploy then you will have to create bricks manually from the new vg you
> have created.
>
>
>
> On Mon, Aug 28, 2017 at 2:10 PM, Anzar Esmail Sainudeen <
> an...@it.thumbay.com> wrote:
>
> Dear Nara,
>
>
>
> Thank you for your great reply.
>
>
>
> 1) can you please check if the disks what would be used for brick creation
> does not have labels or any partitions on them ?
>
>
>
> Yes I agreed there is no labels partition available, my doubt is it
> possible to create required bricks partition from available 406.7G  Linux
> LVM. Following are the physical volume and volume group information.
>
>
>
>
>
> [root@ovirtnode1 ~]# pvdisplay
>
>   --- Physical volume ---
>
>   PV Name   /dev/sda3
>
>   VG Name   onn
>
>   PV Size   555.73 GiB / not usable 2.00 MiB
>
>   Allocatable   yes
>
>   PE Size   4.00 MiB
>
>   Total PE  142267

Re: [ovirt-users] hosted engine setup with Gluster fail

2017-08-28 Thread Kasturi Narra
yes, you can create. I do not see any problems there.

May i know how these vgs are created ? If they are not created using
gdeploy then you will have to create bricks manually from the new vg you
have created.

On Mon, Aug 28, 2017 at 2:10 PM, Anzar Esmail Sainudeen <
an...@it.thumbay.com> wrote:

> Dear Nara,
>
>
>
> Thank you for your great reply.
>
>
>
> 1) can you please check if the disks what would be used for brick creation
> does not have labels or any partitions on them ?
>
>
>
> Yes I agreed there is no labels partition available, my doubt is it
> possible to create required bricks partition from available 406.7G  Linux
> LVM. Following are the physical volume and volume group information.
>
>
>
>
>
> [root@ovirtnode1 ~]# pvdisplay
>
>   --- Physical volume ---
>
>   PV Name   /dev/sda3
>
>   VG Name   onn
>
>   PV Size   555.73 GiB / not usable 2.00 MiB
>
>   Allocatable   yes
>
>   PE Size   4.00 MiB
>
>   Total PE  142267
>
>   Free PE   3823
>
>   Allocated PE  138444
>
>   PV UUID   v1eGGf-r1he-3XZt-JUOM-8XiT-iGkf-0xClUe
>
>
>
> [root@ovirtnode1 ~]# vgdisplay
>
>   --- Volume group ---
>
>   VG Name   onn
>
>   System ID
>
>   Formatlvm2
>
>   Metadata Areas1
>
>   Metadata Sequence No  48
>
>   VG Access read/write
>
>   VG Status resizable
>
>   MAX LV0
>
>   Cur LV12
>
>   Open LV   7
>
>   Max PV0
>
>   Cur PV1
>
>   Act PV1
>
>   VG Size   555.73 GiB
>
>   PE Size   4.00 MiB
>
>   Total PE  142267
>
>   Alloc PE / Size   138444 / 540.80 GiB
>
>   Free  PE / Size   3823 / 14.93 GiB
>
>   VG UUID   nFfNXN-DcJt-bX1Q-UQ2U-07J5-ceT3-ULFtcy
>
>
>
>
>
> I am thinking, to reduce the vg size and create new vg for gluster. Is it
> a good thinking.
>
>
>
>
>
>
>
> Anzar Esmail Sainudeen
>
> Group Datacenter Incharge| IT Infra Division | Thumbay Group
>
> P.O Box : 4184 | Ajman | United Arab Emirates.
>
> Mobile: 055-8633699|Tel: 06 7431333 |  Extn :1303
>
> Email: an...@it.thumbay.com | Website: www.thumbay.com
>
> [image: cid:image001.jpg@01D18D9D.15A17620]
>
>
>
> Disclaimer: This message contains confidential information and is intended
> only for the individual named. If you are not the named addressee, you are
> hereby notified that disclosing, copying, distributing or taking any action
> in reliance on the contents of this e-mail is strictly prohibited. Please
> notify the sender immediately by e-mail if you have received this e-mail by
> mistake, and delete this material. Thumbay Group accepts no liability for
> errors or omissions in the contents of this message, which arise as a
> result of e-mail transmission.
>
>
>
> *From:* Kasturi Narra [mailto:kna...@redhat.com]
> *Sent:* Monday, August 28, 2017 9:48 AM
> *To:* Anzar Esmail Sainudeen
> *Cc:* users
> *Subject:* Re: [ovirt-users] hosted engine setup with Gluster fail
>
>
>
> Hi,
>
>
>
>If i understand right gdeploy script is failing at [1]. There could be
> two possible reasons why that would fail.
>
>
>
> 1) can you please check if the disks what would be used for brick creation
> does not have lables or any partitions on them ?
>
>
>
> 2) can you please check if the path [1] exists. If it does not can you
> please change the path of the script in gdeploy.conf file
> to /usr/share/gdeploy/scripts/grafton-sanity-check.sh
>
>
>
> [1] /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh
>
>
>
> Thanks
>
> kasturi
>
>
>
> On Sun, Aug 27, 2017 at 6:52 PM, Anzar Esmail Sainudeen <
> an...@it.thumbay.com> wrote:
>
> Dear Team Ovirt,
>
>
>
> I am trying to deploy hosted engine setup with Gluster. Hosted engine
> setup was failed. Total number of host is 3 server
>
>
>
>
>
> PLAY [gluster_servers] **
> ***
>
>
>
> TASK [Run a shell script] **
> 
>
> fatal: [ovirtnode4.thumbaytechlabs.int]: FAILED! => {"failed": true,
> "msg": "The conditional check 'result.rc != 0' failed. The error was: error
> while evaluating conditional (result.rc != 0): 'dict object' has no
> attribute 'rc'"}
>
> fatal: [ovirtnode3.thumbaytechlabs.i

Re: [ovirt-users] hosted engine setup with Gluster fail

2017-08-27 Thread Kasturi Narra
Hi,

   If i understand right gdeploy script is failing at [1]. There could be
two possible reasons why that would fail.

1) can you please check if the disks what would be used for brick creation
does not have lables or any partitions on them ?

2) can you please check if the path [1] exists. If it does not can you
please change the path of the script in gdeploy.conf file
to /usr/share/gdeploy/scripts/grafton-sanity-check.sh

[1] /usr/share/ansible/gdeploy/scripts/grafton-sanity-check.sh

Thanks
kasturi

On Sun, Aug 27, 2017 at 6:52 PM, Anzar Esmail Sainudeen <
an...@it.thumbay.com> wrote:

> Dear Team Ovirt,
>
>
>
> I am trying to deploy hosted engine setup with Gluster. Hosted engine
> setup was failed. Total number of host is 3 server
>
>
>
>
>
> PLAY [gluster_servers] **
> ***
>
>
>
> TASK [Run a shell script] **
> 
>
> fatal: [ovirtnode4.thumbaytechlabs.int]: FAILED! => {"failed": true,
> "msg": "The conditional check 'result.rc != 0' failed. The error was: error
> while evaluating conditional (result.rc != 0): 'dict object' has no
> attribute 'rc'"}
>
> fatal: [ovirtnode3.thumbaytechlabs.int]: FAILED! => {"failed": true,
> "msg": "The conditional check 'result.rc != 0' failed. The error was: error
> while evaluating conditional (result.rc != 0): 'dict object' has no
> attribute 'rc'"}
>
> fatal: [ovirtnode2.thumbaytechlabs.int]: FAILED! => {"failed": true,
> "msg": "The conditional check 'result.rc != 0' failed. The error was: error
> while evaluating conditional (result.rc != 0): 'dict object' has no
> attribute 'rc'"}
>
> to retry, use: --limit @/tmp/tmp59G7Vc/run-script.retry
>
>
>
> PLAY RECAP 
> *
>
> ovirtnode2.thumbaytechlabs.int : ok=0changed=0unreachable=0
> failed=1
>
> ovirtnode3.thumbaytechlabs.int : ok=0changed=0unreachable=0
> failed=1
>
> ovirtnode4.thumbaytechlabs.int : ok=0changed=0unreachable=0
> failed=1
>
>
>
>
>
> Please note my finding.
>
>
>
> 1.Still I am doubt with bricks setup ares . because during the ovirt
> node setup time automatically create partition and mount all space. Please
> find below #fdisk –l output
>
> 2.
>
> [root@ovirtnode4 ~]# fdisk –l
>
>
>
> WARNING: fdisk GPT support is currently new, and therefore in an
> experimental phase. Use at your own discretion.
>
>
>
> Disk /dev/sda: 438.0 GB, 437998583808 bytes, 855465984 sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 512 bytes / 512 bytes
>
> Disk label type: gpt
>
>
>
>
>
> # Start  EndSize  TypeName
>
> 1 2048   411647200M  EFI System  EFI System Partition
>
> 2   411648  2508799  1G  Microsoft basic
>
>  3  2508800855463935  406.7G  Linux LVM
>
>
>
> Disk /dev/mapper/onn-swap: 25.4 GB, 25367150592 bytes, 49545216 sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 512 bytes / 512 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-pool00_tmeta: 1073 MB, 1073741824 bytes, 2097152
> sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 512 bytes / 512 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-pool00_tdata: 394.2 GB, 394159718400 bytes, 769843200
> sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 512 bytes / 512 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-pool00-tpool: 394.2 GB, 394159718400 bytes, 769843200
> sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 131072 bytes / 262144 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-ovirt--node--ng--4.1.4--0.20170728.0+1: 378.1 GB,
> 378053591040 bytes, 738385920 sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 131072 bytes / 262144 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-pool00: 394.2 GB, 394159718400 bytes, 769843200
> sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 131072 bytes / 262144 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-var: 16.1 GB, 16106127360 bytes, 31457280 sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 131072 bytes / 262144 bytes
>
>
>
>
>
> Disk /dev/mapper/onn-root: 378.1 GB, 378053591040 bytes, 738385920 sectors
>
> Units = sectors of 1 * 512 = 512 bytes
>
> Sector size (logical/physical): 512 bytes / 512 bytes
>
> I/O size (minimum/optimal): 131072 bytes / 262144 bytes
>
>
>
>
>
> 

Re: [ovirt-users] glusterfs Error message constantly being reported

2017-08-18 Thread Kasturi Narra
Hi,

   Can you please check if you have vdsm-gluster package installed on the
system ?

Thanks
kasturi

On Wed, Aug 16, 2017 at 6:12 PM, Vadim  wrote:

> Hi, All
>
> ovirt 4.1.4 fresh install
> Constantly seeing this message in the logs, how to fix this:
>
>
> VDSM kvm04 command GlusterTasksListVDS failed: 'AutoProxy[instance]'
> object has no attribute 'glusterTasksList'
> VDSM kvm04 command GlusterTasksListVDS failed: 'AutoProxy[instance]'
> object has no attribute 'glusterTasksList'
> VDSM kvm04 command GlusterTasksListVDS failed: 'AutoProxy[instance]'
> object has no attribute 'glusterTasksList'
> VDSM kvm04 command GetGlusterVolumeAdvancedDetailsVDS failed:
> 'AutoProxy[instance]' object has no attribute 'glusterVolumeStatvfs'
> VDSM kvm04 command GetGlusterVolumeAdvancedDetailsVDS failed:
> 'AutoProxy[instance]' object has no attribute 'glusterVolumeStatvfs'
> VDSM kvm04 command GlusterTasksListVDS failed: 'AutoProxy[instance]'
> object has no attribute 'glusterTasksList'
> VDSM kvm04 command GlusterTasksListVDS failed: 'AutoProxy[instance]'
> object has no attribute 'glusterTasksList'
> VDSM kvm04 command GlusterTasksListVDS failed: 'AutoProxy[instance]'
> object has no attribute 'glusterTasksList'
>
> --
> Thanks,
> Vadim
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] How to shutdown an oVirt cluster with Gluster and hosted engine

2017-08-09 Thread Kasturi Narra
Hi,

  It would not do that as before shutting down the engine vm we are
setting maintenance mode to global. Main use is to allow the administrator
to start/stop/modify the engine VM with out any worry of interference from
the HA agents.

Thanks
kasturi

On Wed, Aug 9, 2017 at 2:16 PM, Moacir Ferreira <moacirferre...@hotmail.com>
wrote:

> That is great Kasturi, thanks!
>
>
> I will go over these steps just to make sure nothing is missing. But it
> looks the right way to do it. The only step that looks strange is the step
> 6 to shutdown. I always though that if I shutdown a HA protected machine
> oVirt would then launch it back again. So in my thoughts I would do the
> step 6 before step 4. This said, am I missing something?
>
> Moacir
>
> ------
> *From:* Kasturi Narra <kna...@redhat.com>
> *Sent:* Wednesday, August 9, 2017 5:51 AM
> *To:* Moacir Ferreira
> *Cc:* Erekle Magradze; users@ovirt.org
>
> *Subject:* Re: [ovirt-users] How to shutdown an oVirt cluster with
> Gluster and hosted engine
>
> Hi,
>You can follow the steps below to do that.
>
> 1) Stop all the virtual machines.
>
> 2) Move all the storage domains other than hosted_storage to maintenance
> which will unmount them from all the nodes.
>
> 3)  Move HE to global maintenance 'hosted-engine --set-maintenance --mode
> =global'
>
> 4) stop HE vm by running the command 'hosted-engine --vm-shutdown'
>
> 5) confirm that engine is down using the command 'hosted-engine
> --vm-status'
>
> 6) stop ha agent and broker services on all the nodes by running the
> command 'systemctl stop ovirt-ha-broker' ; 'systemctl stop ovirt-ha-agent'
>
> 7) umount hosted-engine from all the hypervisors 'hosted-engine
> --disconnect-storage'
>
> 8) stop all the volumes.
>
> 9) power off all the hypervisors.
>
>
> To bring it up back again below steps will help.
>
>
> 1) Power on all the hypervisors.
>
> 2) start all the volumes
>
> 3) start ha agent and broker services on all the nodes by running the
> command 'systemctl start ovirt-ha-broker' ; 'systemctl start ovirt-ha-agent'
>
> 4) Move hosted-engine out of global maintenance by running the command 
> hosted-engine
> --set-maintenance --mode =none
>
> 5) give some time for the HE to come up. check for hosted-engine
> --vm-status to see if HE vm is up.
>
> 6) Activate all storage domains from UI.
>
> 7) start all virtual machines.
>
> Hope this helps !!!
>
> Thanks
>
> kasturi.
>
> On Tue, Aug 8, 2017 at 2:27 AM, Moacir Ferreira <
> moacirferre...@hotmail.com> wrote:
>
>> Sorry Erekle, I am just a beginner...
>>
>>
>> From the hosted engine I can put the two other servers, that are not
>> hosting the hosted-engine, on maintenance, and that was what I did. When I
>> tried to put the last one on maintenance it did not allow me due to the
>> hosted-engine and I force it shunting down from the ssh CLI.
>>
>>
>> So, what should I do? My guess is that I should ssh to the hosted engine
>> and shut it down. As it would not have another node to re-launch itself,
>> then it would stay down. After this I should shutdown the oVirt node. Is it?
>>
>>
>> Anyway, I made a mistake and I forced it. When I try to bring the cluster
>> back, I noticed that glusterd was not enabled to start when the node
>> power-on. As I am a beginner, I "think" that there must be a reason to not
>> start glusterd when the node comes up. So I started glusterd on the
>> arbitrator (3rd server), then on the second node and finally on the host
>> that was hosting the hosted-engine. It worked, but when I try to change the
>> maintenance mode on the two nodes (2nd and 3rd) back to normal, the
>> hosted-engine went down and I had to start it manually.
>>
>>
>> All this said, I would like to know how to bring the cluster down and how
>> to bring it back on in the "right way" so I don't get problems. And yes, no
>> VM is running but the hosted-engine.
>>
>>
>> Thanks for sharing your knowledge.
>>
>> Moacir
>>
>>
>> --
>> *From:* Erekle Magradze <erekle.magra...@recogizer.de>
>> *Sent:* Monday, August 7, 2017 9:12 PM
>> *To:* Moacir Ferreira; users@ovirt.org
>> *Subject:* Re: [ovirt-users] How to shutdown an oVirt cluster with
>> Gluster and hosted engine
>>
>>
>> Hi Moacir,
>>
>> First switch off all VMs.
>>
>> Second you need to declare hosts maintenance mode, don't start with SRM
>> (of course if you are able use the ovirt-engine), it w

Re: [ovirt-users] How to shutdown an oVirt cluster with Gluster and hosted engine

2017-08-08 Thread Kasturi Narra
Hi,
   You can follow the steps below to do that.

1) Stop all the virtual machines.

2) Move all the storage domains other than hosted_storage to maintenance
which will unmount them from all the nodes.

3)  Move HE to global maintenance 'hosted-engine --set-maintenance --mode
=global'

4) stop HE vm by running the command 'hosted-engine --vm-shutdown'

5) confirm that engine is down using the command 'hosted-engine --vm-status'

6) stop ha agent and broker services on all the nodes by running the
command 'systemctl stop ovirt-ha-broker' ; 'systemctl stop ovirt-ha-agent'

7) umount hosted-engine from all the hypervisors 'hosted-engine
--disconnect-storage'

8) stop all the volumes.

9) power off all the hypervisors.


To bring it up back again below steps will help.


1) Power on all the hypervisors.

2) start all the volumes

3) start ha agent and broker services on all the nodes by running the
command 'systemctl start ovirt-ha-broker' ; 'systemctl start ovirt-ha-agent'

4) Move hosted-engine out of global maintenance by running the command
hosted-engine
--set-maintenance --mode =none

5) give some time for the HE to come up. check for hosted-engine
--vm-status to see if HE vm is up.

6) Activate all storage domains from UI.

7) start all virtual machines.

Hope this helps !!!

Thanks

kasturi.

On Tue, Aug 8, 2017 at 2:27 AM, Moacir Ferreira 
wrote:

> Sorry Erekle, I am just a beginner...
>
>
> From the hosted engine I can put the two other servers, that are not
> hosting the hosted-engine, on maintenance, and that was what I did. When I
> tried to put the last one on maintenance it did not allow me due to the
> hosted-engine and I force it shunting down from the ssh CLI.
>
>
> So, what should I do? My guess is that I should ssh to the hosted engine
> and shut it down. As it would not have another node to re-launch itself,
> then it would stay down. After this I should shutdown the oVirt node. Is it?
>
>
> Anyway, I made a mistake and I forced it. When I try to bring the cluster
> back, I noticed that glusterd was not enabled to start when the node
> power-on. As I am a beginner, I "think" that there must be a reason to not
> start glusterd when the node comes up. So I started glusterd on the
> arbitrator (3rd server), then on the second node and finally on the host
> that was hosting the hosted-engine. It worked, but when I try to change the
> maintenance mode on the two nodes (2nd and 3rd) back to normal, the
> hosted-engine went down and I had to start it manually.
>
>
> All this said, I would like to know how to bring the cluster down and how
> to bring it back on in the "right way" so I don't get problems. And yes, no
> VM is running but the hosted-engine.
>
>
> Thanks for sharing your knowledge.
>
> Moacir
>
>
> --
> *From:* Erekle Magradze 
> *Sent:* Monday, August 7, 2017 9:12 PM
> *To:* Moacir Ferreira; users@ovirt.org
> *Subject:* Re: [ovirt-users] How to shutdown an oVirt cluster with
> Gluster and hosted engine
>
>
> Hi Moacir,
>
> First switch off all VMs.
>
> Second you need to declare hosts maintenance mode, don't start with SRM
> (of course if you are able use the ovirt-engine), it will ask you to
> shutdown the glusterfs on a machine.
>
> Third if all machines are in maintenance mode, you can start shutting down
> them.
>
>
> If you have hosted engine setup follow this [1]
>
>
> Cheers
>
> Erekle
>
>
> [1] https://github.com/rharmonson/richtech/wiki/OSVDC-Series:-
> oVirt-3.6-Cluster-Shutdown-and-Startup
>
> 
> OSVDC Series: oVirt 3.6 Cluster Shutdown and Startup ...
> 
> github.com
> Contribute to richtech development by creating an account on GitHub.
>
> On 08/07/2017 08:58 PM, Moacir Ferreira wrote:
>
> I have installed a oVirt cluster in a KVM virtualized test environment.
> Now, how do I properly shutdown the oVirt cluster, with Gluster and the
> hosted engine?
>
> I.e.: I want to install a cluster of 3 servers and then send it to a
> remote office. How do I do it properly? I noticed that glusterd is not
> enabled to start automatically. And how do I deal with the hosted engine?
>
>
> Thanks,
>
> Moacir
>
>
> ___
> Users mailing listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>
>
> --
> Recogizer Group GmbH
>
> Dr.rer.nat. Erekle Magradze
> Lead Big Data Engineering & DevOps
> Rheinwerkallee 2, 53227 Bonn
> Tel: +49 228 29974555
>
> E-Mail erekle.magra...@recogizer.de
> Web: www.recogizer.com
>
> Recogizer auf LinkedIn https://www.linkedin.com/company-beta/10039182/
> Folgen Sie uns auf Twitter https://twitter.com/recogizer
>
> -
> Recogizer Group GmbH
> Geschäftsführer: Oliver 

Re: [ovirt-users] ovirt 4.1 hosted engine hyper converged on glusterfs 3.8.10 : "engine" storage domain alway complain about "unsynced" elements

2017-07-24 Thread Kasturi Narra
These errors are because not having glusternw assigned to the correct
interface. Once you attach that these errors should go away.  This has
nothing to do with the problem you are seeing.

sahina any idea about engine not showing the correct volume info ?

On Mon, Jul 24, 2017 at 7:30 PM, yayo (j) <jag...@gmail.com> wrote:

> Hi,
>
> UI refreshed but problem still remain ...
>
> No specific error, I've only these errors but I've read that there is no
> problem if I have this kind of errors:
>
>
> 2017-07-24 15:53:59,823+02 INFO  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler2)
> [b7590c4] START, GlusterServersListVDSCommand(HostName =
> node01.localdomain.local, VdsIdVDSCommandParametersBase:{runAsync='true',
> hostId='4c89baa5-e8f7-4132-a4b3-af332247570c'}), log id: 29a62417
> 2017-07-24 15:54:01,066+02 INFO  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler2)
> [b7590c4] FINISH, GlusterServersListVDSCommand, return: 
> [10.10.20.80/24:CONNECTED,
> node02.localdomain.local:CONNECTED, gdnode04:CONNECTED], log id: 29a62417
> 2017-07-24 15:54:01,076+02 INFO  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler2)
> [b7590c4] START, GlusterVolumesListVDSCommand(HostName =
> node01.localdomain.local, GlusterVolumesListVDSParameters:{runAsync='true',
> hostId='4c89baa5-e8f7-4132-a4b3-af332247570c'}), log id: 7fce25d3
> 2017-07-24 15:54:02,209+02 WARN  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler2) [b7590c4]
> Could not associate brick 'gdnode01:/gluster/engine/brick' of volume
> 'd19c19e3-910d-437b-8ba7-4f2a23d17515' with correct network as no gluster
> network found in cluster '0002-0002-0002-0002-017a'
> 2017-07-24 15:54:02,212+02 WARN  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler2) [b7590c4]
> Could not associate brick 'gdnode02:/gluster/engine/brick' of volume
> 'd19c19e3-910d-437b-8ba7-4f2a23d17515' with correct network as no gluster
> network found in cluster '0002-0002-0002-0002-017a'
> 2017-07-24 15:54:02,215+02 WARN  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler2) [b7590c4]
> Could not associate brick 'gdnode04:/gluster/engine/brick' of volume
> 'd19c19e3-910d-437b-8ba7-4f2a23d17515' with correct network as no gluster
> network found in cluster '0002-0002-0002-0002-017a'
> 2017-07-24 15:54:02,218+02 WARN  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler2) [b7590c4]
> Could not associate brick 'gdnode01:/gluster/data/brick' of volume
> 'c7a5dfc9-3e72-4ea1-843e-c8275d4a7c2d' with correct network as no gluster
> network found in cluster '0002-0002-0002-0002-017a'
> 2017-07-24 15:54:02,221+02 WARN  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler2) [b7590c4]
> Could not associate brick 'gdnode02:/gluster/data/brick' of volume
> 'c7a5dfc9-3e72-4ea1-843e-c8275d4a7c2d' with correct network as no gluster
> network found in cluster '0002-0002-0002-0002-017a'
> 2017-07-24 15:54:02,224+02 WARN  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler2) [b7590c4]
> Could not associate brick 'gdnode04:/gluster/data/brick' of volume
> 'c7a5dfc9-3e72-4ea1-843e-c8275d4a7c2d' with correct network as no gluster
> network found in cluster '0002-0002-0002-0002-017a'
> 2017-07-24 15:54:02,224+02 INFO  [org.ovirt.engine.core.vdsbro
> ker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler2)
> [b7590c4] FINISH, GlusterVolumesListVDSCommand, return: {d19c19e3-910d-437
> b-8ba7-4f2a23d17515=org.ovirt.engine.core.common.businessentities.gluste
> r.GlusterVolumeEntity@fdc91062, c7a5dfc9-3e72-4ea1-843e-c8275d
> 4a7c2d=org.ovirt.engine.core.common.businessentities.gluste
> r.GlusterVolumeEntity@999a6f23}, log id: 7fce25d3
>
>
> Thank you
>
>
> 2017-07-24 8:12 GMT+02:00 Kasturi Narra <kna...@redhat.com>:
>
>> Hi,
>>
>>Regarding the UI showing incorrect information about engine and data
>> volumes, can you please refresh the UI and see if the issue persists  plus
>> any errors in the engine.log files ?
>>
>> Thanks
>> kasturi
>>
>> On Sat, Jul 22, 2017 at 11:43 AM, Ravishankar N <ravishan...@redhat.com>
>> wrote:
>>
>>>
>>> On 07/21/2017 11:41 PM, yayo (j) wrote:
>>>
>>> Hi,
>>>
>>> Sorry for follow up again, but, checking the ovirt interface I've found
>>> that ovirt report the "e

Re: [ovirt-users] Ovirt 4.1 additional hosted-engine deploy setup on another host not working

2017-07-24 Thread Kasturi Narra
Hi,

This option appears in the host tab only when HostedEngine vm and
hosted_storage is present in the UI. Before adding another host make sure
that you add your first data domain to the UI which will automatically
import HostedEngine vm and hosted_storage. Once these two are imported you
will be able to see 'hosted-engine' sub tab in the 'Add host' / edit host
dialog box.

Thanks
kasturi

On Mon, Jul 24, 2017 at 4:05 PM, TranceWorldLogic . <
tranceworldlo...@gmail.com> wrote:

> Hi,
>
> I want to add another host to hosted-engine.
> Hence I tried to follow steps as shown in below link:
>
> http://www.ovirt.org/documentation/self-hosted/chap-Installing_Additional_
> Hosts_to_a_Self-Hosted_Environment/
> Topic :
>
> *Adding an Additional Self-Hosted Engine Host*
> But I not found any additional sub-tab call hosted-engine.
> Even adding host I tired to edit host but still not observe.
>
> Do I need to run some command hosted-engine --deploy to add another host ?
> Or is it handle by GUI automatically ?
>
> Thanks,
> ~Rohit
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt 4.1 hosted engine hyper converged on glusterfs 3.8.10 : "engine" storage domain alway complain about "unsynced" elements

2017-07-24 Thread Kasturi Narra
Hi,

   Regarding the UI showing incorrect information about engine and data
volumes, can you please refresh the UI and see if the issue persists  plus
any errors in the engine.log files ?

Thanks
kasturi

On Sat, Jul 22, 2017 at 11:43 AM, Ravishankar N 
wrote:

>
> On 07/21/2017 11:41 PM, yayo (j) wrote:
>
> Hi,
>
> Sorry for follow up again, but, checking the ovirt interface I've found
> that ovirt report the "engine" volume as an "arbiter" configuration and the
> "data" volume as full replicated volume. Check these screenshots:
>
>
> This is probably some refresh bug in the UI, Sahina might be able to tell
> you.
>
>
> https://drive.google.com/drive/folders/0ByUV7xQtP1gCTE8tUTFfVmR5aDQ?
> usp=sharing
>
> But the "gluster volume info" command report that all 2 volume are full
> replicated:
>
>
> *Volume Name: data*
> *Type: Replicate*
> *Volume ID: c7a5dfc9-3e72-4ea1-843e-c8275d4a7c2d*
> *Status: Started*
> *Snapshot Count: 0*
> *Number of Bricks: 1 x 3 = 3*
> *Transport-type: tcp*
> *Bricks:*
> *Brick1: gdnode01:/gluster/data/brick*
> *Brick2: gdnode02:/gluster/data/brick*
> *Brick3: gdnode04:/gluster/data/brick*
> *Options Reconfigured:*
> *nfs.disable: on*
> *performance.readdir-ahead: on*
> *transport.address-family: inet*
> *storage.owner-uid: 36*
> *performance.quick-read: off*
> *performance.read-ahead: off*
> *performance.io-cache: off*
> *performance.stat-prefetch: off*
> *performance.low-prio-threads: 32*
> *network.remote-dio: enable*
> *cluster.eager-lock: enable*
> *cluster.quorum-type: auto*
> *cluster.server-quorum-type: server*
> *cluster.data-self-heal-algorithm: full*
> *cluster.locking-scheme: granular*
> *cluster.shd-max-threads: 8*
> *cluster.shd-wait-qlength: 1*
> *features.shard: on*
> *user.cifs: off*
> *storage.owner-gid: 36*
> *features.shard-block-size: 512MB*
> *network.ping-timeout: 30*
> *performance.strict-o-direct: on*
> *cluster.granular-entry-heal: on*
> *auth.allow: **
> *server.allow-insecure: on*
>
>
>
>
>
> *Volume Name: engine*
> *Type: Replicate*
> *Volume ID: d19c19e3-910d-437b-8ba7-4f2a23d17515*
> *Status: Started*
> *Snapshot Count: 0*
> *Number of Bricks: 1 x 3 = 3*
> *Transport-type: tcp*
> *Bricks:*
> *Brick1: gdnode01:/gluster/engine/brick*
> *Brick2: gdnode02:/gluster/engine/brick*
> *Brick3: gdnode04:/gluster/engine/brick*
> *Options Reconfigured:*
> *nfs.disable: on*
> *performance.readdir-ahead: on*
> *transport.address-family: inet*
> *storage.owner-uid: 36*
> *performance.quick-read: off*
> *performance.read-ahead: off*
> *performance.io-cache: off*
> *performance.stat-prefetch: off*
> *performance.low-prio-threads: 32*
> *network.remote-dio: off*
> *cluster.eager-lock: enable*
> *cluster.quorum-type: auto*
> *cluster.server-quorum-type: server*
> *cluster.data-self-heal-algorithm: full*
> *cluster.locking-scheme: granular*
> *cluster.shd-max-threads: 8*
> *cluster.shd-wait-qlength: 1*
> *features.shard: on*
> *user.cifs: off*
> *storage.owner-gid: 36*
> *features.shard-block-size: 512MB*
> *network.ping-timeout: 30*
> *performance.strict-o-direct: on*
> *cluster.granular-entry-heal: on*
> *auth.allow: **
>
>   server.allow-insecure: on
>
>
> 2017-07-21 19:13 GMT+02:00 yayo (j) :
>
>> 2017-07-20 14:48 GMT+02:00 Ravishankar N :
>>
>>>
>>> But it does  say something. All these gfids of completed heals in the
>>> log below are the for the ones that you have given the getfattr output of.
>>> So what is likely happening is there is an intermittent connection problem
>>> between your mount and the brick process, leading to pending heals again
>>> after the heal gets completed, which is why the numbers are varying each
>>> time. You would need to check why that is the case.
>>> Hope this helps,
>>> Ravi
>>>
>>>
>>>
>>> *[2017-07-20 09:58:46.573079] I [MSGID: 108026]
>>> [afr-self-heal-common.c:1254:afr_log_selfheal] 0-engine-replicate-0:
>>> Completed data selfheal on e6dfd556-340b-4b76-b47b-7b6f5bd74327.
>>> sources=[0] 1  sinks=2*
>>> *[2017-07-20 09:59:22.995003] I [MSGID: 108026]
>>> [afr-self-heal-metadata.c:51:__afr_selfheal_metadata_do]
>>> 0-engine-replicate-0: performing metadata selfheal on
>>> f05b9742-2771-484a-85fc-5b6974bcef81*
>>> *[2017-07-20 09:59:22.999372] I [MSGID: 108026]
>>> [afr-self-heal-common.c:1254:afr_log_selfheal] 0-engine-replicate-0:
>>> Completed metadata selfheal on f05b9742-2771-484a-85fc-5b6974bcef81.
>>> sources=[0] 1  sinks=2*
>>>
>>>
>>
>> Hi,
>>
>> following your suggestion, I've checked the "peer" status and I found
>> that there is too many name for the hosts, I don't know if this can be the
>> problem or part of it:
>>
>> *gluster peer status on NODE01:*
>> *Number of Peers: 2*
>>
>> *Hostname: dnode02.localdomain.local*
>> *Uuid: 7c0ebfa3-5676-4d3f-9bfa-7fff6afea0dd*
>> *State: Peer in Cluster (Connected)*
>> *Other names:*
>> *192.168.10.52*
>> *dnode02.localdomain.local*
>> *10.10.20.90*
>> *10.10.10.20*
>>
>>
>>
>>
>> *gluster peer 

Re: [ovirt-users] Engine HA-Issues

2017-07-17 Thread Kasturi Narra
Hi ,

  Can you please check the following. Following could be one of the reason
why HE vm restarts every minute.

Check the error or engine health state. If it’s to do with Liveliness
check, then this is mostly an issue connecting to engine.

- Check if engine FQDN is reachable from all hosts

-  curl -v http:///ovirt-engine/services/health - does this
return ok?

- Access the HE console and check if ovirt-engine is running.

- Check /var/log/ovirt-engine/server.log or
/var/log/ovirt-engine/engine.log if there are errors starting ovirt-engine


Thanks

kasturi


On Fri, Jul 14, 2017 at 10:28 PM, Sven Achtelik 
wrote:

> Hi All,
>
>
>
> after running solid for several month my ovirt-engine started rebooting on
> several hosts. I’ve looked into the hostend-engine –vm-status and it sees
> that the engine is up on one host but not reachable. At the same time I can
> access the gui and everything is working fine. After some time the engine
> is shutting down and all hosts are trying to start the engine until one is
> the winner, at least it looks like this. Any clues where to look at and
> find the issue with the liveliness check ?
>
>
>
> 
> 
>
>
>
> --== Host 1 status ==--
>
>
>
> conf_on_shared_storage : True
>
> Status up-to-date  : True
>
> Hostname   : ovirt-node01
>
> Host ID: 1
>
> Engine status  : {"reason": "vm not running on this
> host", "health": "bad", "vm": "down", "detail": "unknown"}
>
> Score  : 3400
>
> stopped: False
>
> Local maintenance  : False
>
> crc32  : 3eb33843
>
> local_conf_timestamp   : 17128
>
> Host timestamp : 17113
>
> Extra metadata (valid at timestamp):
>
> metadata_parse_version=1
>
> metadata_feature_version=1
>
> timestamp=17113 (Fri Jul 14 11:50:23 2017)
>
> host-id=1
>
> score=3400
>
> vm_conf_refresh_time=17128 (Fri Jul 14 11:50:38 2017)
>
> conf_on_shared_storage=True
>
> maintenance=False
>
> state=EngineDown
>
> stopped=False
>
>
>
>
>
> --== Host 2 status ==--
>
>
>
> conf_on_shared_storage : True
>
> Status up-to-date  : True
>
> Hostname   : ovirt-node02.mgmt.lan
>
> Host ID: 2
>
> Engine status  : {"reason": "failed liveliness check",
> "health": "bad", "vm": "up", "detail": "up"}
>
> Score  : 3400
>
> stopped: False
>
> Local maintenance  : False
>
> crc32  : 2a8c86cc
>
> local_conf_timestamp   : 523182
>
> Host timestamp : 523167
>
> Extra metadata (valid at timestamp):
>
> metadata_parse_version=1
>
> metadata_feature_version=1
>
> timestamp=523167 (Fri Jul 14 11:50:25 2017)
>
> host-id=2
>
> score=3400
>
> vm_conf_refresh_time=523182 (Fri Jul 14 11:50:40 2017)
>
> conf_on_shared_storage=True
>
> maintenance=False
>
> state=EngineStarting
>
> stopped=False
>
>
>
>
>
> --== Host 3 status ==--
>
>
>
> conf_on_shared_storage : True
>
> Status up-to-date  : True
>
> Hostname   : ovirt-node03.mgmt.lan
>
> Host ID: 3
>
> Engine status  : {"reason": "vm not running on this
> host", "health": "bad", "vm": "down", "detail": "unknown"}
>
> Score  : 3400
>
> stopped: False
>
> Local maintenance  : False
>
> crc32  : f8490d79
>
> local_conf_timestamp   : 527698
>
> Host timestamp : 527683
>
> Extra metadata (valid at timestamp):
>
> metadata_parse_version=1
>
> metadata_feature_version=1
>
> timestamp=527683 (Fri Jul 14 11:50:33 2017)
>
> host-id=3
>
> score=3400
>
> vm_conf_refresh_time=527698 (Fri Jul 14 11:50:47 2017)
>
> conf_on_shared_storage=True
>
> maintenance=False
>
> state=EngineDown
>
> stopped=False
>
>
>
> 
> --
>
> Thank you,
>
> Sven
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users