Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Simone Tiraboschi
On Mon, Sep 4, 2017 at 7:32 PM, Arsène Gschwind 
wrote:

>
>
> On 09/04/2017 06:32 PM, Simone Tiraboschi wrote:
>
>
>
> On Mon, Sep 4, 2017 at 6:24 PM, Arsène Gschwind  > wrote:
>
>>
>>
>> On 09/04/2017 02:51 PM, Simone Tiraboschi wrote:
>>
>>
>>
>> On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind <
>> arsene.gschw...@unibas.ch> wrote:
>>
>>>
>>>
>>> On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:
>>>
>>>
>>>
>>> On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind <
>>> arsene.gschw...@unibas.ch> wrote:
>>>


 On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:



 On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind <
 arsene.gschw...@unibas.ch> wrote:

> Hi Simone,
>
> On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:
>
>
>
> On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind <
> arsene.gschw...@unibas.ch> wrote:
>
>> Hi Didi,
>>
>> On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:
>>
>> On Mon, Sep 4, 2017 at 10:16 AM, Arsène 
>> Gschwind  wrote:
>>
>> Hi all,
>>
>> A while ago I had some problem with hosted-engine network which wasn't 
>> set
>> correctly at deploy time, so I finally decided to redeploy the hosted 
>> engine
>> in the hope the network will be set correctly this time. I've followed 
>> this
>> procedure:
>>
>> Stop all VMs
>> Full backup of HE DB and export to safe place
>> Cleanup HE storage following https://access.redhat.com/solutions/2121581
>> Reboot Hosts
>> Re-deploy HE until DB recovery
>> Recover DB adding the following param:
>>  --he-remove-storage-vm Removes the hosted-engine storage
>> domain, all its entities and the hosted-engine VM during restore.
>>  --he-remove-hosts  Removes all the hosted-engine hosts
>> during restore.
>>
>> Finalize HE deployment.
>>
>> Everything did run without errors and I'm able to access Web UI.
>>
>> But now I don't see my HE VM and its respective Storage Domain, the logs
>> says it isn't able to import it. I see all other SD and I'm able to 
>> manage
>> my VMs as before.
>>
>> Please find attached engine.log
>>
>> I think this is your problem:
>>
>> 2017-09-04 03:26:14,272+02 INFO
>> [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
>> (org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
>> in the system which are part of VG id
>> 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
>>
>> I don't see a VG with this ID, here the IDs I see on the hosts:
>>
>>   VG   #PV #LV #SN Attr   VSize
>> VFree
>>   6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g
>> 14.50g
>>   b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t
>> 2.31t
>>   b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n-  5.27t
>> 2.50t
>>
>
>
> Could you please repeat the command on host adm-kvmh70 ?
>
> 2017-09-04 09:04:18,163+02 INFO  [org.ovirt.engine.core.bll.st
> orage.domain.ImportHostedEngineStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
> ImportHostedEngineStorageDomainCommand internal: true.
> 2017-09-04 09:04:18,189+02 INFO  [org.ovirt.engine.core.vdsbro
> ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
> [7d2e6cb2] START, GetVGInfoVDSCommand(HostName = adm-kvmh70,
> GetVGInfoVDSCommandParameters:{runAsync='true',
> hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
> VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
> 2017-09-04 09:04:18,232+02 INFO  [org.ovirt.engine.core.vdsbro
> ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
> [7d2e6cb2] FINISH, GetVGInfoVDSCommand, return:
> [LUNs:{id='repl_HostedEngine', 
> physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
> volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
> serial='SHITACHI_OPEN-V_5048', lunMapping='4',
> vendorId='HITACHI', productId='OPEN-V', lunConnections='[]',
> deviceSize='100', pvSize='0', peCount='797', peAllocatedCount='681',
> vendorName='HITACHI', pathsDictionary='[sdf=true, sdu=true, sdk=true,
> sdp=true]', pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]',
> lunType='FCP', status='null', diskId='null', diskAlias='null',
> storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246',
> storageDomainName='null', discardMaxSize='268435456',
> discardZeroesData='true'}], log id: 6693b98a
> 2017-09-04 09:04:18,245+02 INFO  [org.ovirt.engine.core.bll.st
> orage.domain.AddExistingBlockStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-34) 

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Arsène Gschwind



On 09/04/2017 06:32 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 6:24 PM, Arsène Gschwind 
> wrote:




On 09/04/2017 02:51 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind
> wrote:



On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind
> wrote:



On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind
> wrote:

Hi Simone,


On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind
> wrote:

Hi Didi,


On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:

On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind

  wrote:

Hi all,

A while ago I had some problem with hosted-engine network 
which wasn't set
correctly at deploy time, so I finally decided to redeploy 
the hosted engine
in the hope the network will be set correctly this time. 
I've followed this
procedure:

Stop all VMs
Full backup of HE DB and export to safe place
Cleanup HE storage 
followinghttps://access.redhat.com/solutions/2121581

Reboot Hosts
Re-deploy HE until DB recovery
Recover DB adding the following param:
  --he-remove-storage-vm Removes the 
hosted-engine storage
domain, all its entities and the hosted-engine VM during 
restore.
  --he-remove-hosts  Removes all the 
hosted-engine hosts
during restore.

Finalize HE deployment.

Everything did run without errors and I'm able to access 
Web UI.

But now I don't see my HE VM and its respective Storage 
Domain, the logs
says it isn't able to import it. I see all other SD and I'm 
able to manage
my VMs as before.

Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO

[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are 
existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

I don't see a VG with this ID, here the IDs I
see on the hosts:

VG #PV #LV #SN Attr   VSize VFree
6b62cc06-fc44-4c38-af6d-bfd9cbe73246 1  10   0
wz--n- 99.62g 14.50g
b0414c06-d984-4001-a998-fd9a2e79fb83 2  70   0
wz--n- 10.00t 2.31t
b2e30961-7cff-4cca-83d6-bee3a4f890ee 2  47   0
wz--n-  5.27t 2.50t



Could you please repeat the command on host
adm-kvmh70 ?

2017-09-04 09:04:18,163+02 INFO
 [org.ovirt.engine.core.bll.st

orage.domain.ImportHostedEngineStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [247a3718]
Running command:
ImportHostedEngineStorageDomainCommand internal: true.
2017-09-04 09:04:18,189+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2]
START, GetVGInfoVDSCommand(HostName = adm-kvmh70,
GetVGInfoVDSCommandParameters:{runAsync='true',
hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}),
log id: 6693b98a
2017-09-04 09:04:18,232+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2]
FINISH, GetVGInfoVDSCommand, return:
[LUNs:{id='repl_HostedEngine',

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Simone Tiraboschi
On Mon, Sep 4, 2017 at 6:24 PM, Arsène Gschwind 
wrote:

>
>
> On 09/04/2017 02:51 PM, Simone Tiraboschi wrote:
>
>
>
> On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind  > wrote:
>
>>
>>
>> On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:
>>
>>
>>
>> On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind <
>> arsene.gschw...@unibas.ch> wrote:
>>
>>>
>>>
>>> On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:
>>>
>>>
>>>
>>> On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind <
>>> arsene.gschw...@unibas.ch> wrote:
>>>
 Hi Simone,

 On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:



 On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind <
 arsene.gschw...@unibas.ch> wrote:

> Hi Didi,
>
> On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:
>
> On Mon, Sep 4, 2017 at 10:16 AM, Arsène 
> Gschwind  wrote:
>
> Hi all,
>
> A while ago I had some problem with hosted-engine network which wasn't set
> correctly at deploy time, so I finally decided to redeploy the hosted 
> engine
> in the hope the network will be set correctly this time. I've followed 
> this
> procedure:
>
> Stop all VMs
> Full backup of HE DB and export to safe place
> Cleanup HE storage following https://access.redhat.com/solutions/2121581
> Reboot Hosts
> Re-deploy HE until DB recovery
> Recover DB adding the following param:
>  --he-remove-storage-vm Removes the hosted-engine storage
> domain, all its entities and the hosted-engine VM during restore.
>  --he-remove-hosts  Removes all the hosted-engine hosts
> during restore.
>
> Finalize HE deployment.
>
> Everything did run without errors and I'm able to access Web UI.
>
> But now I don't see my HE VM and its respective Storage Domain, the logs
> says it isn't able to import it. I see all other SD and I'm able to manage
> my VMs as before.
>
> Please find attached engine.log
>
> I think this is your problem:
>
> 2017-09-04 03:26:14,272+02 INFO
> [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
> in the system which are part of VG id
> 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
>
> I don't see a VG with this ID, here the IDs I see on the hosts:
>
>   VG   #PV #LV #SN Attr   VSize
> VFree
>   6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g
> 14.50g
>   b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t
> 2.31t
>   b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n-  5.27t
> 2.50t
>


 Could you please repeat the command on host adm-kvmh70 ?

 2017-09-04 09:04:18,163+02 INFO  [org.ovirt.engine.core.bll.st
 orage.domain.ImportHostedEngineStorageDomainCommand]
 (org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
 ImportHostedEngineStorageDomainCommand internal: true.
 2017-09-04 09:04:18,189+02 INFO  [org.ovirt.engine.core.vdsbro
 ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
 [7d2e6cb2] START, GetVGInfoVDSCommand(HostName = adm-kvmh70,
 GetVGInfoVDSCommandParameters:{runAsync='true',
 hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
 VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
 2017-09-04 09:04:18,232+02 INFO  [org.ovirt.engine.core.vdsbro
 ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
 [7d2e6cb2] FINISH, GetVGInfoVDSCommand, return:
 [LUNs:{id='repl_HostedEngine', 
 physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
 volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
 serial='SHITACHI_OPEN-V_5048', lunMapping='4', vendorId='HITACHI',
 productId='OPEN-V', lunConnections='[]', deviceSize='100', pvSize='0',
 peCount='797', peAllocatedCount='681', vendorName='HITACHI',
 pathsDictionary='[sdf=true, sdu=true, sdk=true, sdp=true]',
 pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]', lunType='FCP',
 status='null', diskId='null', diskAlias='null',
 storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246',
 storageDomainName='null', discardMaxSize='268435456',
 discardZeroesData='true'}], log id: 6693b98a
 2017-09-04 09:04:18,245+02 INFO  [org.ovirt.engine.core.bll.st
 orage.domain.AddExistingBlockStorageDomainCommand]
 (org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] There are existing luns
 in the system which are part of VG id 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL
 2z-wbTU8g'
 2017-09-04 09:04:18,245+02 WARN  [org.ovirt.engine.core.bll.st
 orage.domain.AddExistingBlockStorageDomainCommand]
 (org.ovirt.thread.pool-6-thread-34) 

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Arsène Gschwind



On 09/04/2017 02:51 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind 
> wrote:




On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind
> wrote:



On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind
> wrote:

Hi Simone,


On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind
> wrote:

Hi Didi,


On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:

On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind

  wrote:

Hi all,

A while ago I had some problem with hosted-engine network which 
wasn't set
correctly at deploy time, so I finally decided to redeploy the 
hosted engine
in the hope the network will be set correctly this time. I've 
followed this
procedure:

Stop all VMs
Full backup of HE DB and export to safe place
Cleanup HE storage 
followinghttps://access.redhat.com/solutions/2121581

Reboot Hosts
Re-deploy HE until DB recovery
Recover DB adding the following param:
  --he-remove-storage-vm Removes the hosted-engine 
storage
domain, all its entities and the hosted-engine VM during 
restore.
  --he-remove-hosts  Removes all the 
hosted-engine hosts
during restore.

Finalize HE deployment.

Everything did run without errors and I'm able to access Web UI.

But now I don't see my HE VM and its respective Storage Domain, 
the logs
says it isn't able to import it. I see all other SD and I'm 
able to manage
my VMs as before.

Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO

[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are 
existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

I don't see a VG with this ID, here the IDs I see
on the hosts:

VG #PV #LV #SN Attr   VSize VFree
6b62cc06-fc44-4c38-af6d-bfd9cbe73246 1  10   0
wz--n- 99.62g 14.50g
b0414c06-d984-4001-a998-fd9a2e79fb83 2  70   0
wz--n- 10.00t 2.31t
b2e30961-7cff-4cca-83d6-bee3a4f890ee 2  47   0
wz--n-  5.27t 2.50t



Could you please repeat the command on host adm-kvmh70 ?

2017-09-04 09:04:18,163+02 INFO
 [org.ovirt.engine.core.bll.st

orage.domain.ImportHostedEngineStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [247a3718] Running
command: ImportHostedEngineStorageDomainCommand
internal: true.
2017-09-04 09:04:18,189+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] START,
GetVGInfoVDSCommand(HostName = adm-kvmh70,
GetVGInfoVDSCommandParameters:{runAsync='true',
hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log
id: 6693b98a
2017-09-04 09:04:18,232+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] FINISH,
GetVGInfoVDSCommand, return:
[LUNs:{id='repl_HostedEngine',
physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
serial='SHITACHI_OPEN-V_5048', lunMapping='4',
vendorId='HITACHI', productId='OPEN-V',
lunConnections='[]', deviceSize='100', pvSize='0',
peCount='797', peAllocatedCount='681',
vendorName='HITACHI', pathsDictionary='[sdf=true,
sdu=true, sdk=true, sdp=true]',
pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]',

Re: [ovirt-users] hyperconverged question

2017-09-04 Thread FERNANDO FREDIANI
I had the very same impression. It doesn't look like that it works then. 
So for a fully redundant where you can loose a complete host you must 
have at least 3 nodes then ?


Fernando


On 01/09/2017 12:53, Jim Kusznir wrote:
Huh...Ok., how do I convert the arbitrar to full replica, then?  I was 
misinformed when I created this setup.  I thought the arbitrator held 
enough metadata that it could validate or refudiate  any one replica 
(kinda like the parity drive for a RAID-4 array).  I was also under 
the impression that one replica  + Arbitrator is enough to keep the 
array online and functional.


--Jim

On Fri, Sep 1, 2017 at 5:22 AM, Charles Kozler > wrote:


@ Jim - you have only two data volumes and lost quorum. Arbitrator
only stores metadata, no actual files. So yes, you were running in
degraded mode so some operations were hindered.

@ Sahina - Yes, this actually worked fine for me once I did that.
However, the issue I am still facing, is when I go to create a new
gluster storage domain (replica 3, hyperconverged) and I tell it
"Host to use" and I select that host. If I fail that host, all VMs
halt. I do not recall this in 3.6 or early 4.0. This to me makes
it seem like this is "pinning" a node to a volume and vice versa
like you could, for instance, for a singular hyperconverged to ex:
export a local disk via NFS and then mount it via ovirt domain.
But of course, this has its caveats. To that end, I am using
gluster replica 3, when configuring it I say "host to use: " node
1, then in the connection details I give it node1:/data. I fail
node1, all VMs halt. Did I miss something?

On Fri, Sep 1, 2017 at 2:13 AM, Sahina Bose > wrote:

To the OP question, when you set up a gluster storage domain,
you need to specify backup-volfile-servers=:
where server2 and server3 also have bricks running. When
server1 is down, and the volume is mounted again - server2 or
server3 are queried to get the gluster volfiles.

@Jim, if this does not work, are you using 4.1.5 build with
libgfapi access? If not, please provide the vdsm and gluster
mount logs to analyse

If VMs go to paused state - this could mean the storage is not
available. You can check "gluster volume status " to
see if atleast 2 bricks are running.

On Fri, Sep 1, 2017 at 11:31 AM, Johan Bernhardsson
> wrote:

If gluster drops in quorum so that it has less votes than
it should it will stop file operations until quorum is
back to normal.If i rember it right you need two bricks to
write for quorum to be met and that the arbiter only is a
vote to avoid split brain.


Basically what you have is a raid5 solution without a
spare. And when one disk dies it will run in degraded
mode. And some raid systems will stop the raid until you
have removed the disk or forced it to run anyway.

You can read up on it here:

https://gluster.readthedocs.io/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/



/Johan

On Thu, 2017-08-31 at 22:33 -0700, Jim Kusznir wrote:

Hi all:

Sorry to hijack the thread, but I was about to start
essentially the same thread.

I have a 3 node cluster, all three are hosts and gluster
nodes (replica 2 + arbitrar).  I DO have the
mnt_options=backup-volfile-servers= set:

storage=192.168.8.11:/engine
mnt_options=backup-volfile-servers=192.168.8.12:192.168.8.13

I had an issue today where 192.168.8.11 went down.  ALL
VMs immediately paused, including the engine (all VMs
were running on host2:192.168.8.12).  I couldn't get any
gluster stuff working until host1 (192.168.8.11) was
restored.

What's wrong / what did I miss?

(this was set up "manually" through the article on
setting up self-hosted gluster cluster back when 4.0 was
new..I've upgraded it to 4.1 since).

Thanks!
--Jim


On Thu, Aug 31, 2017 at 12:31 PM, Charles Kozler
> wrote:

Typo..."Set it up and then failed that **HOST**"

And upon that host going down, the storage domain went
down. I only have hosted storage domain and this new one
- is this why the DC went down and no SPM could be elected?

I dont recall this working this way in early 4.0 or 3.6

   

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Simone Tiraboschi
On Mon, Sep 4, 2017 at 2:21 PM, Arsène Gschwind 
wrote:

>
>
> On 09/04/2017 02:01 PM, Simone Tiraboschi wrote:
>
>
>
> On Mon, Sep 4, 2017 at 1:55 PM, Arsène Gschwind  > wrote:
>
>>
>>
>> On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:
>>
>>
>>
>> On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind <
>> arsene.gschw...@unibas.ch> wrote:
>>
>>> Hi Simone,
>>>
>>> On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:
>>>
>>>
>>>
>>> On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind <
>>> arsene.gschw...@unibas.ch> wrote:
>>>
 Hi Didi,

 On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:

 On Mon, Sep 4, 2017 at 10:16 AM, Arsène 
 Gschwind  wrote:

 Hi all,

 A while ago I had some problem with hosted-engine network which wasn't set
 correctly at deploy time, so I finally decided to redeploy the hosted 
 engine
 in the hope the network will be set correctly this time. I've followed this
 procedure:

 Stop all VMs
 Full backup of HE DB and export to safe place
 Cleanup HE storage following https://access.redhat.com/solutions/2121581
 Reboot Hosts
 Re-deploy HE until DB recovery
 Recover DB adding the following param:
  --he-remove-storage-vm Removes the hosted-engine storage
 domain, all its entities and the hosted-engine VM during restore.
  --he-remove-hosts  Removes all the hosted-engine hosts
 during restore.

 Finalize HE deployment.

 Everything did run without errors and I'm able to access Web UI.

 But now I don't see my HE VM and its respective Storage Domain, the logs
 says it isn't able to import it. I see all other SD and I'm able to manage
 my VMs as before.

 Please find attached engine.log

 I think this is your problem:

 2017-09-04 03:26:14,272+02 INFO
 [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
 (org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
 in the system which are part of VG id
 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

 I don't see a VG with this ID, here the IDs I see on the hosts:

   VG   #PV #LV #SN Attr   VSize  VFree
   6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g 14.50g
   b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t  2.31t
   b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n-  5.27t  2.50t

>>>
>>>
>>> Could you please repeat the command on host adm-kvmh70 ?
>>>
>>> 2017-09-04 09:04:18,163+02 INFO  [org.ovirt.engine.core.bll.st
>>> orage.domain.ImportHostedEngineStorageDomainCommand]
>>> (org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
>>> ImportHostedEngineStorageDomainCommand internal: true.
>>> 2017-09-04 09:04:18,189+02 INFO  [org.ovirt.engine.core.vdsbro
>>> ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
>>> [7d2e6cb2] START, GetVGInfoVDSCommand(HostName = adm-kvmh70,
>>> GetVGInfoVDSCommandParameters:{runAsync='true',
>>> hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
>>> VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
>>> 2017-09-04 09:04:18,232+02 INFO  [org.ovirt.engine.core.vdsbro
>>> ker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
>>> [7d2e6cb2] FINISH, GetVGInfoVDSCommand, return:
>>> [LUNs:{id='repl_HostedEngine', 
>>> physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
>>> volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
>>> serial='SHITACHI_OPEN-V_5048', lunMapping='4', vendorId='HITACHI',
>>> productId='OPEN-V', lunConnections='[]', deviceSize='100', pvSize='0',
>>> peCount='797', peAllocatedCount='681', vendorName='HITACHI',
>>> pathsDictionary='[sdf=true, sdu=true, sdk=true, sdp=true]',
>>> pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]', lunType='FCP',
>>> status='null', diskId='null', diskAlias='null',
>>> storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246',
>>> storageDomainName='null', discardMaxSize='268435456',
>>> discardZeroesData='true'}], log id: 6693b98a
>>> 2017-09-04 09:04:18,245+02 INFO  [org.ovirt.engine.core.bll.st
>>> orage.domain.AddExistingBlockStorageDomainCommand]
>>> (org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] There are existing luns
>>> in the system which are part of VG id 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL
>>> 2z-wbTU8g'
>>> 2017-09-04 09:04:18,245+02 WARN  [org.ovirt.engine.core.bll.st
>>> orage.domain.AddExistingBlockStorageDomainCommand]
>>> (org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] Validation of action
>>> 'AddExistingBlockStorageDomain' failed for user SYSTEM. Reasons:
>>> VAR__TYPE__STORAGE__DOMAIN,VAR__ACTION__ADD,ACTION_TYPE_FAIL
>>> ED_IMPORT_STORAGE_DOMAIN_EXTERNAL_LUN_DISK_EXIST
>>>
>>> I don't know which command you are talking about, I didn't run any
>>> 

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Arsène Gschwind



On 09/04/2017 01:52 PM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind 
> wrote:


Hi Simone,


On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind
> wrote:

Hi Didi,


On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:

On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind
   wrote:

Hi all,

A while ago I had some problem with hosted-engine network which wasn't 
set
correctly at deploy time, so I finally decided to redeploy the hosted 
engine
in the hope the network will be set correctly this time. I've followed 
this
procedure:

Stop all VMs
Full backup of HE DB and export to safe place
Cleanup HE storage followinghttps://access.redhat.com/solutions/2121581

Reboot Hosts
Re-deploy HE until DB recovery
Recover DB adding the following param:
  --he-remove-storage-vm Removes the hosted-engine storage
domain, all its entities and the hosted-engine VM during restore.
  --he-remove-hosts  Removes all the hosted-engine hosts
during restore.

Finalize HE deployment.

Everything did run without errors and I'm able to access Web UI.

But now I don't see my HE VM and its respective Storage Domain, the logs
says it isn't able to import it. I see all other SD and I'm able to 
manage
my VMs as before.

Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO

[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

I don't see a VG with this ID, here the IDs I see on the hosts:

  VG #PV #LV #SN Attr   VSize  VFree
6b62cc06-fc44-4c38-af6d-bfd9cbe73246 1  10   0 wz--n- 99.62g
14.50g
b0414c06-d984-4001-a998-fd9a2e79fb83 2  70   0 wz--n- 10.00t 
2.31t
b2e30961-7cff-4cca-83d6-bee3a4f890ee 2  47   0 wz--n-  5.27t 
2.50t



Could you please repeat the command on host adm-kvmh70 ?

2017-09-04 09:04:18,163+02 INFO
 
[org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
ImportHostedEngineStorageDomainCommand internal: true.
2017-09-04 09:04:18,189+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] START,
GetVGInfoVDSCommand(HostName = adm-kvmh70,
GetVGInfoVDSCommandParameters:{runAsync='true',
hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
2017-09-04 09:04:18,232+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] FINISH,
GetVGInfoVDSCommand, return: [LUNs:{id='repl_HostedEngine',
physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
serial='SHITACHI_OPEN-V_5048', lunMapping='4',
vendorId='HITACHI', productId='OPEN-V', lunConnections='[]',
deviceSize='100', pvSize='0', peCount='797',
peAllocatedCount='681', vendorName='HITACHI',
pathsDictionary='[sdf=true, sdu=true, sdk=true, sdp=true]',
pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]',
lunType='FCP', status='null', diskId='null', diskAlias='null',
storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246',
storageDomainName='null', discardMaxSize='268435456',
discardZeroesData='true'}], log id: 6693b98a
2017-09-04 09:04:18,245+02 INFO
 
[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] There are existing
luns in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
2017-09-04 09:04:18,245+02 WARN
 
[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] Validation of
action 'AddExistingBlockStorageDomain' failed for user SYSTEM.
Reasons:

VAR__TYPE__STORAGE__DOMAIN,VAR__ACTION__ADD,ACTION_TYPE_FAILED_IMPORT_STORAGE_DOMAIN_EXTERNAL_LUN_DISK_EXIST

I don't know which command you are talking about, I didn't run any
command since it tries to import the SD automatically.


Sorry, can you please run vgdisplay on adm-kvmh70 ?

No 

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Simone Tiraboschi
On Mon, Sep 4, 2017 at 12:23 PM, Arsène Gschwind 
wrote:

> Hi Simone,
>
> On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:
>
>
>
> On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind <
> arsene.gschw...@unibas.ch> wrote:
>
>> Hi Didi,
>>
>> On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:
>>
>> On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind 
>>  wrote:
>>
>> Hi all,
>>
>> A while ago I had some problem with hosted-engine network which wasn't set
>> correctly at deploy time, so I finally decided to redeploy the hosted engine
>> in the hope the network will be set correctly this time. I've followed this
>> procedure:
>>
>> Stop all VMs
>> Full backup of HE DB and export to safe place
>> Cleanup HE storage following https://access.redhat.com/solutions/2121581
>> Reboot Hosts
>> Re-deploy HE until DB recovery
>> Recover DB adding the following param:
>>  --he-remove-storage-vm Removes the hosted-engine storage
>> domain, all its entities and the hosted-engine VM during restore.
>>  --he-remove-hosts  Removes all the hosted-engine hosts
>> during restore.
>>
>> Finalize HE deployment.
>>
>> Everything did run without errors and I'm able to access Web UI.
>>
>> But now I don't see my HE VM and its respective Storage Domain, the logs
>> says it isn't able to import it. I see all other SD and I'm able to manage
>> my VMs as before.
>>
>> Please find attached engine.log
>>
>> I think this is your problem:
>>
>> 2017-09-04 03:26:14,272+02 INFO
>> [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
>> (org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
>> in the system which are part of VG id
>> 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
>>
>> I don't see a VG with this ID, here the IDs I see on the hosts:
>>
>>   VG   #PV #LV #SN Attr   VSize  VFree
>>   6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g 14.50g
>>   b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t  2.31t
>>   b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n-  5.27t  2.50t
>>
>
>
> Could you please repeat the command on host adm-kvmh70 ?
>
> 2017-09-04 09:04:18,163+02 INFO  [org.ovirt.engine.core.bll.
> storage.domain.ImportHostedEngineStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
> ImportHostedEngineStorageDomainCommand internal: true.
> 2017-09-04 09:04:18,189+02 INFO  [org.ovirt.engine.core.
> vdsbroker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
> [7d2e6cb2] START, GetVGInfoVDSCommand(HostName = adm-kvmh70,
> GetVGInfoVDSCommandParameters:{runAsync='true',
> hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
> VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
> 2017-09-04 09:04:18,232+02 INFO  [org.ovirt.engine.core.
> vdsbroker.vdsbroker.GetVGInfoVDSCommand] (org.ovirt.thread.pool-6-thread-34)
> [7d2e6cb2] FINISH, GetVGInfoVDSCommand, return:
> [LUNs:{id='repl_HostedEngine', 
> physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
> volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
> serial='SHITACHI_OPEN-V_5048', lunMapping='4', vendorId='HITACHI',
> productId='OPEN-V', lunConnections='[]', deviceSize='100', pvSize='0',
> peCount='797', peAllocatedCount='681', vendorName='HITACHI',
> pathsDictionary='[sdf=true, sdu=true, sdk=true, sdp=true]',
> pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]', lunType='FCP',
> status='null', diskId='null', diskAlias='null', 
> storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246',
> storageDomainName='null', discardMaxSize='268435456',
> discardZeroesData='true'}], log id: 6693b98a
> 2017-09-04 09:04:18,245+02 INFO  [org.ovirt.engine.core.bll.
> storage.domain.AddExistingBlockStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] There are existing luns in
> the system which are part of VG id 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-
> SL2z-wbTU8g'
> 2017-09-04 09:04:18,245+02 WARN  [org.ovirt.engine.core.bll.
> storage.domain.AddExistingBlockStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] Validation of action 
> 'AddExistingBlockStorageDomain'
> failed for user SYSTEM. Reasons: VAR__TYPE__STORAGE__DOMAIN,
> VAR__ACTION__ADD,ACTION_TYPE_FAILED_IMPORT_STORAGE_DOMAIN_
> EXTERNAL_LUN_DISK_EXIST
>
> I don't know which command you are talking about, I didn't run any command
> since it tries to import the SD automatically.
>

Sorry, can you please run vgdisplay on adm-kvmh70 ?


>
> Thanks,
> Arsene
>
>
>
>>
>>
>> Thanks for any help to resolve that issue.
>>
>> I guess you can try to remove this disk/lun from the engine and let it retry.
>>
>> Could let me know how to remove that lun from the engine?
>>
>> If the only disk is of the hosted-engine, I guess it should have been
>> removed by '--he-remove-storage-vm' - if so, please open a bug
>> describing your flow in detail. 

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Arsène Gschwind

Hi Simone,


On 09/04/2017 11:14 AM, Simone Tiraboschi wrote:



On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind 
> wrote:


Hi Didi,


On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:

On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind
   wrote:

Hi all,

A while ago I had some problem with hosted-engine network which wasn't set
correctly at deploy time, so I finally decided to redeploy the hosted engine
in the hope the network will be set correctly this time. I've followed this
procedure:

Stop all VMs
Full backup of HE DB and export to safe place
Cleanup HE storage followinghttps://access.redhat.com/solutions/2121581

Reboot Hosts
Re-deploy HE until DB recovery
Recover DB adding the following param:
  --he-remove-storage-vm Removes the hosted-engine storage
domain, all its entities and the hosted-engine VM during restore.
  --he-remove-hosts  Removes all the hosted-engine hosts
during restore.

Finalize HE deployment.

Everything did run without errors and I'm able to access Web UI.

But now I don't see my HE VM and its respective Storage Domain, the logs
says it isn't able to import it. I see all other SD and I'm able to manage
my VMs as before.

Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO

[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

I don't see a VG with this ID, here the IDs I see on the hosts:

  VG    #PV #LV #SN Attr   VSize  VFree
  6b62cc06-fc44-4c38-af6d-bfd9cbe73246 1  10   0 wz--n- 99.62g 14.50g
  b0414c06-d984-4001-a998-fd9a2e79fb83 2  70   0 wz--n- 10.00t  2.31t
  b2e30961-7cff-4cca-83d6-bee3a4f890ee 2  47   0 wz--n-  5.27t  2.50t



Could you please repeat the command on host adm-kvmh70 ?

2017-09-04 09:04:18,163+02 INFO 
 [org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand] 
(org.ovirt.thread.pool-6-thread-34) [247a3718] Running command: 
ImportHostedEngineStorageDomainCommand internal: true.
2017-09-04 09:04:18,189+02 INFO 
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand] 
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] START, 
GetVGInfoVDSCommand(HostName = adm-kvmh70, 
GetVGInfoVDSCommandParameters:{runAsync='true', 
hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b', 
VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
2017-09-04 09:04:18,232+02 INFO 
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand] 
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] FINISH, 
GetVGInfoVDSCommand, return: [LUNs:{id='repl_HostedEngine', 
physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO', 
volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g', 
serial='SHITACHI_OPEN-V_5048', lunMapping='4', vendorId='HITACHI', 
productId='OPEN-V', lunConnections='[]', deviceSize='100', pvSize='0', 
peCount='797', peAllocatedCount='681', vendorName='HITACHI', 
pathsDictionary='[sdf=true, sdu=true, sdk=true, sdp=true]', 
pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]', lunType='FCP', 
status='null', diskId='null', diskAlias='null', 
storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246', 
storageDomainName='null', discardMaxSize='268435456', 
discardZeroesData='true'}], log id: 6693b98a
2017-09-04 09:04:18,245+02 INFO 
 [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand] 
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] There are existing luns 
in the system which are part of VG id 
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
2017-09-04 09:04:18,245+02 WARN 
 [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand] 
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] Validation of action 
'AddExistingBlockStorageDomain' failed for user SYSTEM. Reasons: 
VAR__TYPE__STORAGE__DOMAIN,VAR__ACTION__ADD,ACTION_TYPE_FAILED_IMPORT_STORAGE_DOMAIN_EXTERNAL_LUN_DISK_EXIST
I don't know which command you are talking about, I didn't run any 
command since it tries to import the SD automatically.


Thanks,
Arsene





Thanks for any help to resolve that issue.

I guess you can try to remove this disk/lun from the engine and let it 
retry.

Could let me know how to remove that lun from the engine?

If the only disk is of the hosted-engine, I guess it should have been
removed by '--he-remove-storage-vm' - if so, please open a bug
describing your flow in detail. Thanks.

It seems that this option didn't remove the he storage information
and that this VG is still the old one.

Many thanks for your help

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Simone Tiraboschi
On Mon, Sep 4, 2017 at 10:56 AM, Arsène Gschwind 
wrote:

> Hi Didi,
>
> On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:
>
> On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind 
>  wrote:
>
> Hi all,
>
> A while ago I had some problem with hosted-engine network which wasn't set
> correctly at deploy time, so I finally decided to redeploy the hosted engine
> in the hope the network will be set correctly this time. I've followed this
> procedure:
>
> Stop all VMs
> Full backup of HE DB and export to safe place
> Cleanup HE storage following https://access.redhat.com/solutions/2121581
> Reboot Hosts
> Re-deploy HE until DB recovery
> Recover DB adding the following param:
>  --he-remove-storage-vm Removes the hosted-engine storage
> domain, all its entities and the hosted-engine VM during restore.
>  --he-remove-hosts  Removes all the hosted-engine hosts
> during restore.
>
> Finalize HE deployment.
>
> Everything did run without errors and I'm able to access Web UI.
>
> But now I don't see my HE VM and its respective Storage Domain, the logs
> says it isn't able to import it. I see all other SD and I'm able to manage
> my VMs as before.
>
> Please find attached engine.log
>
> I think this is your problem:
>
> 2017-09-04 03:26:14,272+02 INFO
> [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
> (org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
> in the system which are part of VG id
> 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
>
> I don't see a VG with this ID, here the IDs I see on the hosts:
>
>   VG   #PV #LV #SN Attr   VSize  VFree
>   6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g 14.50g
>   b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t  2.31t
>   b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n-  5.27t  2.50t
>


Could you please repeat the command on host adm-kvmh70 ?

2017-09-04 09:04:18,163+02 INFO
 
[org.ovirt.engine.core.bll.storage.domain.ImportHostedEngineStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [247a3718] Running command:
ImportHostedEngineStorageDomainCommand internal: true.
2017-09-04 09:04:18,189+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] START,
GetVGInfoVDSCommand(HostName = adm-kvmh70,
GetVGInfoVDSCommandParameters:{runAsync='true',
hostId='acbacabb-6c4a-43fd-a1e2-2d7ff2f6f98b',
VGID='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'}), log id: 6693b98a
2017-09-04 09:04:18,232+02 INFO
 [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVGInfoVDSCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] FINISH, GetVGInfoVDSCommand,
return: [LUNs:{id='repl_HostedEngine',
physicalVolumeId='kYN8Jj-FBDw-MhxI-XcoZ-w1zH-eQL8-IRIgzO',
volumeGroupId='vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g',
serial='SHITACHI_OPEN-V_5048', lunMapping='4', vendorId='HITACHI',
productId='OPEN-V', lunConnections='[]', deviceSize='100', pvSize='0',
peCount='797', peAllocatedCount='681', vendorName='HITACHI',
pathsDictionary='[sdf=true, sdu=true, sdk=true, sdp=true]',
pathsCapacity='[sdf=100, sdu=100, sdk=100, sdp=100]', lunType='FCP',
status='null', diskId='null', diskAlias='null',
storageDomainId='6b62cc06-fc44-4c38-af6d-bfd9cbe73246',
storageDomainName='null', discardMaxSize='268435456',
discardZeroesData='true'}], log id: 6693b98a
2017-09-04 09:04:18,245+02 INFO
 [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] There are existing luns in
the system which are part of VG id 'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'
2017-09-04 09:04:18,245+02 WARN
 [org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-34) [7d2e6cb2] Validation of action
'AddExistingBlockStorageDomain' failed for user SYSTEM. Reasons:
VAR__TYPE__STORAGE__DOMAIN,VAR__ACTION__ADD,ACTION_TYPE_FAILED_IMPORT_STORAGE_DOMAIN_EXTERNAL_LUN_DISK_EXIST


>
>
> Thanks for any help to resolve that issue.
>
> I guess you can try to remove this disk/lun from the engine and let it retry.
>
> Could let me know how to remove that lun from the engine?
>
> If the only disk is of the hosted-engine, I guess it should have been
> removed by '--he-remove-storage-vm' - if so, please open a bug
> describing your flow in detail. Thanks.
>
> It seems that this option didn't remove the he storage information and
> that this VG is still the old one.
>
> Many thanks for your help
> Rgds,
> Arsene
>
> Best,
>
>
> Arsène
>
> --
>
> Arsène Gschwind
> Fa. Sapify AG im Auftrag der Universität Basel
> IT Services
> Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
> Tel. +41 79 449 25 63 <+41%2079%20449%2025%2063>  |  http://its.unibas.ch
> ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11 
> <+41%2061%20267%2014%2011>
>
>
> 

Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Arsène Gschwind

Hi Didi,


On 09/04/2017 10:15 AM, Yedidyah Bar David wrote:

On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind
 wrote:

Hi all,

A while ago I had some problem with hosted-engine network which wasn't set
correctly at deploy time, so I finally decided to redeploy the hosted engine
in the hope the network will be set correctly this time. I've followed this
procedure:

Stop all VMs
Full backup of HE DB and export to safe place
Cleanup HE storage following https://access.redhat.com/solutions/2121581
Reboot Hosts
Re-deploy HE until DB recovery
Recover DB adding the following param:
  --he-remove-storage-vm Removes the hosted-engine storage
domain, all its entities and the hosted-engine VM during restore.
  --he-remove-hosts  Removes all the hosted-engine hosts
during restore.

Finalize HE deployment.

Everything did run without errors and I'm able to access Web UI.

But now I don't see my HE VM and its respective Storage Domain, the logs
says it isn't able to import it. I see all other SD and I'm able to manage
my VMs as before.

Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO
[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

I don't see a VG with this ID, here the IDs I see on the hosts:

  VG   #PV #LV #SN Attr   VSize VFree
  6b62cc06-fc44-4c38-af6d-bfd9cbe73246   1  10   0 wz--n- 99.62g 14.50g
  b0414c06-d984-4001-a998-fd9a2e79fb83   2  70   0 wz--n- 10.00t  2.31t
  b2e30961-7cff-4cca-83d6-bee3a4f890ee   2  47   0 wz--n- 5.27t  2.50t


Thanks for any help to resolve that issue.


I guess you can try to remove this disk/lun from the engine and let it retry.

Could let me know how to remove that lun from the engine?


If the only disk is of the hosted-engine, I guess it should have been
removed by '--he-remove-storage-vm' - if so, please open a bug
describing your flow in detail. Thanks.
It seems that this option didn't remove the he storage information and 
that this VG is still the old one.


Many thanks for your help
Rgds,
Arsene


Best,


Arsène

--

Arsène Gschwind
Fa. Sapify AG im Auftrag der Universität Basel
IT Services
Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
Tel. +41 79 449 25 63  |  http://its.unibas.ch
ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users






--

*Arsène Gschwind*
Fa. Sapify AG im Auftrag der Universität Basel
IT Services
Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
Tel. +41 79 449 25 63  | http://its.unibas.ch 
ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] First oVirt engine deploy: missing gateway on hosts

2017-09-04 Thread Mauro Tridici

Hi Kasturi,

I just tried to set, using the oVirt Engine UI, a fake gateway (different from 
the real one) for the ovirtmgmt network attached to the hosts.
After that, I set again the old and real gateway: it seems that, at the end, 
the new configuration has been stored correctly in the oVirt Engine environment.

[root@glu01 ~]# route
Kernel IP routing table
Destination Gateway Genmask Flags Metric RefUse Iface
default gateway 0.0.0.0 UG0  00 ens33
link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33

It seems that, finally, the ovirtmgmt network configuration has been updated 
with the right gateway value.

Thanks for your feedback.
Mauro

> Il giorno 04 set 2017, alle ore 09:48, Kasturi Narra  ha 
> scritto:
> 
> Hi Mauro,
> 
>  yes, i do know this problem of gateway disappearing but i am not sure 
> why that happens because in my case it was happening on only on one server. I 
> used to add the gateway again on that node. I am just waiting for some one to 
> reply.
> 
> Thanks
> kasturi
> 
> On Mon, Sep 4, 2017 at 1:02 PM, Mauro Tridici  > wrote:
> Hi Kasturi,
> 
> thank you very much for your support and for the useful explanation.
> If possible, I would ask you if you know a way to correct the “missing 
> gateway” problem that I described in my previous message.
> 
> Thanks a lot,
> Mauro
> 
> 
>> Il giorno 04 set 2017, alle ore 08:51, Kasturi Narra > > ha scritto:
>> 
>> Hi Mauro,
>> 
>>Creating distributed dispersed volumes are not supported from ovirt 
>> UI yet but you should be able to sync them if cluster is imported into the 
>> UI. same holds true for add / remove bricks on disperse and distribute 
>> disperse volumes.
>> 
>>you wont be able to see bricks created because ovirt excepts them to 
>> be mounted at /rhgs/. What you could simply do is uncheck the 
>> check box in the 'Add brick' dialog and type in the path or you could mount 
>> your bricks at above said location and that would show all the available 
>> bricks on the host.
>> 
>>   Hope this helps.
>> 
>> Thanks
>> kasturi.
>> 
>> On Sat, Sep 2, 2017 at 7:16 PM, Mauro Tridici > > wrote:
>> Hi all,
>> 
>> I just started my first Ovirt Engine deploy using a dedicated (and 
>> separated) virtual machine.
>> I’m trying to create and manage a test Gluster cluster using 3 “virtual” 
>> hosts (hostnames are glu01, glu02, glu03)
>> 2 different networks have been defined on the hosts (192.168.213.0/24 
>>  for management network and 192.168.152.0/24 
>>  for gluster network).
>> Ovirt engine deploy completed without any problem, the hosts have been added 
>> easily using ovirtmgmt network (bridgeless mgmt network) and ovirtgluster 
>> (bridgeless gluster network).
>> 
>> Everything seems to be ok for this first deploy, but I just noticed that the 
>> gateway is missing on the target hosts:
>> 
>> [root@glu01 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33
>> 
>> [root@glu02 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33
>> 
>> [root@glu03 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33
>> 
>> Due to this problem I cannot reach internet from ens33 nic (management 
>> network).
>> I just tried to add the gateway in ifcfg-ens33 configuration file but 
>> gateway disappear after host reboot. 
>> 
>> [root@glu01 ~]# cat 

Re: [ovirt-users] failed upgrade oVirt node 4.1.3 -> 4.1.5

2017-09-04 Thread Matthias Leopold

thanks, so i'll wait for 4.1.6 before upgrading my other nodes

Regards
matthias

Am 2017-09-03 um 15:57 schrieb Yuval Turgeman:

Hi,

Seems to be a bug that was resolved here https://gerrit.ovirt.org/c/80716/

Thanks,
Yuval.


On Fri, Sep 1, 2017 at 3:55 PM, Matthias Leopold 
> wrote:


hi,

i'm sorry to write to this list again, but i failed to upgrade a
freshly installed oVirt Node from version 4.1.3 to 4.1.5. it seems
to be a SELinux related problem. i'm attaching imgbased.log +
relevant lines from engine.log.

is the skipped version (4.1.4) the problem?
can i force upgrade to version 4.1.4?

thx
matthias


___
Users mailing list
Users@ovirt.org 
http://lists.ovirt.org/mailman/listinfo/users





--
Matthias Leopold
IT Systems & Communications
Medizinische Universität Wien
Spitalgasse 23 / BT 88 /Ebene 00
A-1090 Wien
Tel: +43 1 40160-21241
Fax: +43 1 40160-921200
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Failed to import the Hosted Engine Storage Domain

2017-09-04 Thread Yedidyah Bar David
On Mon, Sep 4, 2017 at 10:16 AM, Arsène Gschwind
 wrote:
> Hi all,
>
> A while ago I had some problem with hosted-engine network which wasn't set
> correctly at deploy time, so I finally decided to redeploy the hosted engine
> in the hope the network will be set correctly this time. I've followed this
> procedure:
>
> Stop all VMs
> Full backup of HE DB and export to safe place
> Cleanup HE storage following https://access.redhat.com/solutions/2121581
> Reboot Hosts
> Re-deploy HE until DB recovery
> Recover DB adding the following param:
>  --he-remove-storage-vm Removes the hosted-engine storage
> domain, all its entities and the hosted-engine VM during restore.
>  --he-remove-hosts  Removes all the hosted-engine hosts
> during restore.
>
> Finalize HE deployment.
>
> Everything did run without errors and I'm able to access Web UI.
>
> But now I don't see my HE VM and its respective Storage Domain, the logs
> says it isn't able to import it. I see all other SD and I'm able to manage
> my VMs as before.
>
> Please find attached engine.log

I think this is your problem:

2017-09-04 03:26:14,272+02 INFO
[org.ovirt.engine.core.bll.storage.domain.AddExistingBlockStorageDomainCommand]
(org.ovirt.thread.pool-6-thread-24) [2383eaa0] There are existing luns
in the system which are part of VG id
'vvIoS2-fZTZ-99Ox-Ltzq-pr8U-SL2z-wbTU8g'

>
> Thanks for any help to resolve that issue.


I guess you can try to remove this disk/lun from the engine and let it retry.

If the only disk is of the hosted-engine, I guess it should have been
removed by '--he-remove-storage-vm' - if so, please open a bug
describing your flow in detail. Thanks.

Best,

>
> Arsène
>
> --
>
> Arsène Gschwind
> Fa. Sapify AG im Auftrag der Universität Basel
> IT Services
> Klingelbergstr. 70 |  CH-4056 Basel  |  Switzerland
> Tel. +41 79 449 25 63  |  http://its.unibas.ch
> ITS-ServiceDesk: support-...@unibas.ch | +41 61 267 14 11
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>



-- 
Didi
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] First oVirt engine deploy: missing gateway on hosts

2017-09-04 Thread Kasturi Narra
Hi Mauro,

 yes, i do know this problem of gateway disappearing but i am not sure
why that happens because in my case it was happening on only on one server.
I used to add the gateway again on that node. I am just waiting for some
one to reply.

Thanks
kasturi

On Mon, Sep 4, 2017 at 1:02 PM, Mauro Tridici  wrote:

> Hi Kasturi,
>
> thank you very much for your support and for the useful explanation.
> If possible, I would ask you if you know a way to correct the “missing
> gateway” problem that I described in my previous message.
>
> Thanks a lot,
> Mauro
>
>
> Il giorno 04 set 2017, alle ore 08:51, Kasturi Narra 
> ha scritto:
>
> Hi Mauro,
>
>Creating distributed dispersed volumes are not supported from ovirt
> UI yet but you should be able to sync them if cluster is imported into the
> UI. same holds true for add / remove bricks on disperse and distribute
> disperse volumes.
>
>you wont be able to see bricks created because ovirt excepts them
> to be mounted at /rhgs/. What you could simply do is uncheck
> the check box in the 'Add brick' dialog and type in the path or you could
> mount your bricks at above said location and that would show all the
> available bricks on the host.
>
>   Hope this helps.
>
> Thanks
> kasturi.
>
> On Sat, Sep 2, 2017 at 7:16 PM, Mauro Tridici 
> wrote:
>
>> Hi all,
>>
>> I just started my first Ovirt Engine deploy using a dedicated (and
>> separated) virtual machine.
>> I’m trying to create and manage a test Gluster cluster using 3 “virtual”
>> hosts (hostnames are glu01, glu02, glu03)
>> 2 different networks have been defined on the hosts (192.168.213.0/24
>> for management network and 192.168.152.0/24 for gluster network).
>> Ovirt engine deploy completed without any problem, the hosts have been
>> added easily using ovirtmgmt network (bridgeless mgmt network) and
>> ovirtgluster (bridgeless gluster network).
>>
>> Everything seems to be ok for this first deploy, but I just noticed that
>> the gateway is missing on the target hosts:
>>
>> [root@glu01 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse
>> Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00
>> ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00
>> ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
>> ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
>> ens33
>>
>> [root@glu02 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse
>> Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00
>> ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00
>> ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
>> ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
>> ens33
>>
>> [root@glu03 ~]# route
>> Kernel IP routing table
>> Destination Gateway Genmask Flags Metric RefUse
>> Iface
>> link-local  0.0.0.0 255.255.0.0 U 1002   00
>> ens33
>> link-local  0.0.0.0 255.255.0.0 U 1003   00
>> ens34
>> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
>> ens34
>> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
>> ens33
>>
>> Due to this problem I cannot reach internet from ens33 nic (management
>> network).
>> I just tried to add the gateway in ifcfg-ens33 configuration file but
>> gateway disappear after host reboot.
>>
>> [root@glu01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-ens33
>> # Generated by VDSM version 4.19.28-1.el7.centos
>> DEVICE=ens33
>> ONBOOT=yes
>> IPADDR=192.168.213.151
>> NETMASK=255.255.255.0
>> BOOTPROTO=none
>> MTU=1500
>> DEFROUTE=no
>> NM_CONTROLLED=no
>> IPV6INIT=yes
>> IPV6_AUTOCONF=yes
>>
>> The oVirt Engine network configuration is the following one:
>>
>> [host glu01]
>> ens33 -> ovirtmgmt (192.168.213.151, 255.255.255.0, 192.168.213.2)
>> ens34 -> ovirtgluster (192.168.152.151, 255.255.255.0)
>>
>> [host glu02]
>> ens33 -> ovirtmgmt (192.168.213.152, 255.255.255.0, 192.168.213.2)
>> ens34 -> ovirtgluster (192.168.152.152, 255.255.255.0)
>>
>> [host glu03]
>> ens33 -> ovirtmgmt (192.168.213.153, 255.255.255.0, 192.168.213.2)
>> ens34 -> ovirtgluster (192.168.152.153, 255.255.255.0)
>>
>> Do you know the right way to set the gateway IP on all hosts?
>>
>> Just two last questions: I was able to import an existing gluster cluster
>> using oVirt Engine, but I’m not able to create a new volume because:
>>
>> - I can’t select a distributed disperse volume configuration from oVirt
>> Engine volume creation window
>> - i can’t see the bricks to be used to create a new volume (but I can
>> import an existing volume without problem).
>>
>> Is there 

[ovirt-users] Limit number of vms

2017-09-04 Thread qinglong.d...@horebdata.cn
Hi all,
I want to limit number of running vms in ovirt. For example, a vm 
should fail to start when there are enough running vms. Is there an easy way to 
do this such as modifing an vdsm script? Wish someone can help. Thanks!
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] First oVirt engine deploy: missing gateway on hosts

2017-09-04 Thread Mauro Tridici
Hi Kasturi,

thank you very much for your support and for the useful explanation.
If possible, I would ask you if you know a way to correct the “missing gateway” 
problem that I described in my previous message.

Thanks a lot,
Mauro


> Il giorno 04 set 2017, alle ore 08:51, Kasturi Narra  ha 
> scritto:
> 
> Hi Mauro,
> 
>Creating distributed dispersed volumes are not supported from ovirt UI 
> yet but you should be able to sync them if cluster is imported into the UI. 
> same holds true for add / remove bricks on disperse and distribute disperse 
> volumes.
> 
>you wont be able to see bricks created because ovirt excepts them to 
> be mounted at /rhgs/. What you could simply do is uncheck the 
> check box in the 'Add brick' dialog and type in the path or you could mount 
> your bricks at above said location and that would show all the available 
> bricks on the host.
> 
>   Hope this helps.
> 
> Thanks
> kasturi.
> 
> On Sat, Sep 2, 2017 at 7:16 PM, Mauro Tridici  > wrote:
> Hi all,
> 
> I just started my first Ovirt Engine deploy using a dedicated (and separated) 
> virtual machine.
> I’m trying to create and manage a test Gluster cluster using 3 “virtual” 
> hosts (hostnames are glu01, glu02, glu03)
> 2 different networks have been defined on the hosts (192.168.213.0/24 
>  for management network and 192.168.152.0/24 
>  for gluster network).
> Ovirt engine deploy completed without any problem, the hosts have been added 
> easily using ovirtmgmt network (bridgeless mgmt network) and ovirtgluster 
> (bridgeless gluster network).
> 
> Everything seems to be ok for this first deploy, but I just noticed that the 
> gateway is missing on the target hosts:
> 
> [root@glu01 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33
> 
> [root@glu02 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33
> 
> [root@glu03 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00 ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00 ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00 ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00 ens33
> 
> Due to this problem I cannot reach internet from ens33 nic (management 
> network).
> I just tried to add the gateway in ifcfg-ens33 configuration file but gateway 
> disappear after host reboot. 
> 
> [root@glu01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-ens33
> # Generated by VDSM version 4.19.28-1.el7.centos
> DEVICE=ens33
> ONBOOT=yes
> IPADDR=192.168.213.151
> NETMASK=255.255.255.0
> BOOTPROTO=none
> MTU=1500
> DEFROUTE=no
> NM_CONTROLLED=no
> IPV6INIT=yes
> IPV6_AUTOCONF=yes
> 
> The oVirt Engine network configuration is the following one:
> 
> [host glu01]
> ens33 -> ovirtmgmt (192.168.213.151, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.151, 255.255.255.0)
> 
> [host glu02]
> ens33 -> ovirtmgmt (192.168.213.152, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.152, 255.255.255.0)
> 
> [host glu03]
> ens33 -> ovirtmgmt (192.168.213.153, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.153, 255.255.255.0)
> 
> Do you know the right way to set the gateway IP on all hosts?
> 
> Just two last questions: I was able to import an existing gluster cluster 
> using oVirt Engine, but I’m not able to create a new volume because:
> 
> - I can’t select a distributed disperse volume configuration from oVirt 
> Engine volume creation window
> - i can’t see the bricks to be used to create a new volume (but I can import 
> an existing volume without problem).
> 
> Is there something that I can do to resolve the issues and complete my first 
> experience with oVirt?
> 
> Thank you very much,
> Mauro T.
> 
> 
> ___
> Users mailing list
> Users@ovirt.org 
> http://lists.ovirt.org/mailman/listinfo/users 
> 
> 
> 



Re: [ovirt-users] First oVirt engine deploy: missing gateway on hosts

2017-09-04 Thread Kasturi Narra
Hi Mauro,

   Creating distributed dispersed volumes are not supported from ovirt
UI yet but you should be able to sync them if cluster is imported into the
UI. same holds true for add / remove bricks on disperse and distribute
disperse volumes.

   you wont be able to see bricks created because ovirt excepts them to
be mounted at /rhgs/. What you could simply do is uncheck the
check box in the 'Add brick' dialog and type in the path or you could mount
your bricks at above said location and that would show all the available
bricks on the host.

  Hope this helps.

Thanks
kasturi.

On Sat, Sep 2, 2017 at 7:16 PM, Mauro Tridici  wrote:

> Hi all,
>
> I just started my first Ovirt Engine deploy using a dedicated (and
> separated) virtual machine.
> I’m trying to create and manage a test Gluster cluster using 3 “virtual”
> hosts (hostnames are glu01, glu02, glu03)
> 2 different networks have been defined on the hosts (192.168.213.0/24 for
> management network and 192.168.152.0/24 for gluster network).
> Ovirt engine deploy completed without any problem, the hosts have been
> added easily using ovirtmgmt network (bridgeless mgmt network) and
> ovirtgluster (bridgeless gluster network).
>
> Everything seems to be ok for this first deploy, but I just noticed that
> the gateway is missing on the target hosts:
>
> [root@glu01 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse
> Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00
> ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00
> ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
> ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
> ens33
>
> [root@glu02 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse
> Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00
> ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00
> ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
> ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
> ens33
>
> [root@glu03 ~]# route
> Kernel IP routing table
> Destination Gateway Genmask Flags Metric RefUse
> Iface
> link-local  0.0.0.0 255.255.0.0 U 1002   00
> ens33
> link-local  0.0.0.0 255.255.0.0 U 1003   00
> ens34
> 192.168.152.0   0.0.0.0 255.255.255.0   U 0  00
> ens34
> 192.168.213.0   0.0.0.0 255.255.255.0   U 0  00
> ens33
>
> Due to this problem I cannot reach internet from ens33 nic (management
> network).
> I just tried to add the gateway in ifcfg-ens33 configuration file but
> gateway disappear after host reboot.
>
> [root@glu01 ~]# cat /etc/sysconfig/network-scripts/ifcfg-ens33
> # Generated by VDSM version 4.19.28-1.el7.centos
> DEVICE=ens33
> ONBOOT=yes
> IPADDR=192.168.213.151
> NETMASK=255.255.255.0
> BOOTPROTO=none
> MTU=1500
> DEFROUTE=no
> NM_CONTROLLED=no
> IPV6INIT=yes
> IPV6_AUTOCONF=yes
>
> The oVirt Engine network configuration is the following one:
>
> [host glu01]
> ens33 -> ovirtmgmt (192.168.213.151, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.151, 255.255.255.0)
>
> [host glu02]
> ens33 -> ovirtmgmt (192.168.213.152, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.152, 255.255.255.0)
>
> [host glu03]
> ens33 -> ovirtmgmt (192.168.213.153, 255.255.255.0, 192.168.213.2)
> ens34 -> ovirtgluster (192.168.152.153, 255.255.255.0)
>
> Do you know the right way to set the gateway IP on all hosts?
>
> Just two last questions: I was able to import an existing gluster cluster
> using oVirt Engine, but I’m not able to create a new volume because:
>
> - I can’t select a distributed disperse volume configuration from oVirt
> Engine volume creation window
> - i can’t see the bricks to be used to create a new volume (but I can
> import an existing volume without problem).
>
> Is there something that I can do to resolve the issues and complete my
> first experience with oVirt?
>
> Thank you very much,
> Mauro T.
>
>
> ___
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] hyperconverged question

2017-09-04 Thread Kasturi Narra
Hi charles,

 The right option is backup-volfile-servers and not
'backupvolfile-server'.
So can you please use the first one and test ?

Thanks
kasturi

On Sat, Sep 2, 2017 at 5:23 AM, Charles Kozler  wrote:

> Jim -
>
> result of this test...engine crashed but all VM's on the gluster domain
> (backed by the same physical nodes/hardware/gluster process/etc) stayed up
> fine
>
> I guess there is some functional difference between 'backupvolfile-server'
> and 'backup-volfile-servers'?
>
> Perhaps try latter and see what happens. My next test is going to be to
> configure hosted-engine.conf with backupvolfile-server=node2:node3 and
> see if engine VM still shuts down. Seems odd engine VM would shut itself
> down (or vdsm would shut it down) but not other VMs. Perhaps built in HA
> functionality of sorts
>
> On Fri, Sep 1, 2017 at 7:38 PM, Charles Kozler 
> wrote:
>
>> Jim -
>>
>> One thing I noticed is that, by accident, I used
>> 'backupvolfile-server=node2:node3' which is apparently a supported
>> setting. It would appear, by reading the man page of mount.glusterfs, the
>> syntax is slightly different. not sure if my setting being different has
>> different impacts
>>
>> hosted-engine.conf:
>>
>> # cat /etc/ovirt-hosted-engine/hosted-engine.conf | grep -i option
>> mnt_options=backup-volfile-servers=node2:node3
>>
>> And for my datatest gluster domain I have:
>>
>> backupvolfile-server=node2:node3
>>
>> I am now curious what happens when I move everything to node1 and drop
>> node2
>>
>> To that end, will follow up with that test
>>
>>
>>
>>
>> On Fri, Sep 1, 2017 at 7:20 PM, Charles Kozler 
>> wrote:
>>
>>> Jim -
>>>
>>> here is my test:
>>>
>>> - All VM's on node2: hosted engine and 1 test VM
>>> - Test VM on gluster storage domain (with mount options set)
>>> - hosted engine is on gluster as well, with settings persisted to
>>> hosted-engine.conf for backupvol
>>>
>>> All VM's stayed up. Nothing in dmesg of the test vm indicating a pause
>>> or an issue or anything
>>>
>>> However, what I did notice during this, is my /datatest volume doesnt
>>> have quorum set. So I will set that now and report back what happens
>>>
>>> # gluster volume info datatest
>>>
>>> Volume Name: datatest
>>> Type: Replicate
>>> Volume ID: 229c25f9-405e-4fe7-b008-1d3aea065069
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x 3 = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: node1:/gluster/data/datatest/brick1
>>> Brick2: node2:/gluster/data/datatest/brick1
>>> Brick3: node3:/gluster/data/datatest/brick1
>>> Options Reconfigured:
>>> transport.address-family: inet
>>> nfs.disable: on
>>>
>>> Perhaps quorum may be more trouble than its worth when you have 3 nodes
>>> and/or 2 nodes + arbiter?
>>>
>>> Since I am keeping my 3rd node out of ovirt, I am more content on
>>> keeping it as a warm spare if I **had** to swap it in to ovirt cluster, but
>>> keeps my storage 100% quorum
>>>
>>> On Fri, Sep 1, 2017 at 5:18 PM, Jim Kusznir  wrote:
>>>
 I can confirm that I did set it up manually, and I did specify
 backupvol, and in the "manage domain" storage settings, I do have under
 mount options, backup-volfile-servers=192.168.8.12:192.168.8.13  (and
 this was done at initial install time).

 The "used managed gluster" checkbox is NOT checked, and if I check it
 and save settings, next time I go in it is not checked.

 --Jim

 On Fri, Sep 1, 2017 at 2:08 PM, Charles Kozler 
 wrote:

> @ Jim - here is my setup which I will test in a few (brand new
> cluster) and report back what I found in my tests
>
> - 3x servers direct connected via 10Gb
> - 2 of those 3 setup in ovirt as hosts
> - Hosted engine
> - Gluster replica 3 (no arbiter) for all volumes
> - 1x engine volume gluster replica 3 manually configured (not using
> ovirt managed gluster)
> - 1x datatest volume (20gb) replica 3 manually configured (not using
> ovirt managed gluster)
> - 1x nfstest domain served from some other server in my infrastructure
> which, at the time of my original testing, was master domain
>
> I tested this earlier and all VMs stayed online. However, ovirt
> cluster reported DC/cluster down, all VM's stayed up
>
> As I am now typing this, can you confirm you setup your gluster
> storage domain with backupvol? Also, confirm you updated 
> hosted-engine.conf
> with backupvol mount option as well?
>
> On Fri, Sep 1, 2017 at 4:22 PM, Jim Kusznir 
> wrote:
>
>> So, after reading the first document twice and the 2nd link
>> thoroughly once, I believe that the arbitrator volume should be 
>> sufficient
>> and count for replica / split brain.  EG, if any one full replica is 
>> down,
>> and the arbitrator and the other replica is up, 

Re: [ovirt-users] hyperconverged question

2017-09-04 Thread Kasturi Narra
Hi Jim,

  I looked at the gluster volume info and that looks to be fine for me.
Recommended config is arbiter for data and vmstore and for engine it should
be replica 3 since we would want HE to be available always.

 If i understand right the problem you are facing is when you shut down one
of the node all the HE vms and app vms goes to paused state right ?

 For debugging further and to ensure that volume has been mounted
using  backup-volfile-servers option, you can move the storage domain to
maintenance which will umount the volume , activate back which will mount
it again. During this time you can check the mount command passed in vdsm
logs and that should have the backup-volfile-servers option.

Can you please confirm if you have ovirt-guest-agent installed on the
app vms and power management enabled ? ovirt-guest-agent is required on the
app vms to ensure HA functionality

Thanks
kasturi

On Sat, Sep 2, 2017 at 2:48 AM, Jim Kusznir  wrote:

> I can confirm that I did set it up manually, and I did specify backupvol,
> and in the "manage domain" storage settings, I do have under mount
> options, backup-volfile-servers=192.168.8.12:192.168.8.13  (and this was
> done at initial install time).
>
> The "used managed gluster" checkbox is NOT checked, and if I check it and
> save settings, next time I go in it is not checked.
>
> --Jim
>
> On Fri, Sep 1, 2017 at 2:08 PM, Charles Kozler 
> wrote:
>
>> @ Jim - here is my setup which I will test in a few (brand new cluster)
>> and report back what I found in my tests
>>
>> - 3x servers direct connected via 10Gb
>> - 2 of those 3 setup in ovirt as hosts
>> - Hosted engine
>> - Gluster replica 3 (no arbiter) for all volumes
>> - 1x engine volume gluster replica 3 manually configured (not using ovirt
>> managed gluster)
>> - 1x datatest volume (20gb) replica 3 manually configured (not using
>> ovirt managed gluster)
>> - 1x nfstest domain served from some other server in my infrastructure
>> which, at the time of my original testing, was master domain
>>
>> I tested this earlier and all VMs stayed online. However, ovirt cluster
>> reported DC/cluster down, all VM's stayed up
>>
>> As I am now typing this, can you confirm you setup your gluster storage
>> domain with backupvol? Also, confirm you updated hosted-engine.conf with
>> backupvol mount option as well?
>>
>> On Fri, Sep 1, 2017 at 4:22 PM, Jim Kusznir  wrote:
>>
>>> So, after reading the first document twice and the 2nd link thoroughly
>>> once, I believe that the arbitrator volume should be sufficient and count
>>> for replica / split brain.  EG, if any one full replica is down, and the
>>> arbitrator and the other replica is up, then it should have quorum and all
>>> should be good.
>>>
>>> I think my underlying problem has to do more with config than the
>>> replica state.  That said, I did size the drive on my 3rd node planning to
>>> have an identical copy of all data on it, so I'm still not opposed to
>>> making it a full replica.
>>>
>>> Did I miss something here?
>>>
>>> Thanks!
>>>
>>> On Fri, Sep 1, 2017 at 11:59 AM, Charles Kozler 
>>> wrote:
>>>
 These can get a little confusing but this explains it best:
 https://gluster.readthedocs.io/en/latest/Administrator
 %20Guide/arbiter-volumes-and-quorum/#replica-2-and-replica-3-volumes

 Basically in the first paragraph they are explaining why you cant have
 HA with quorum for 2 nodes. Here is another overview doc that explains some
 more

 http://openmymind.net/Does-My-Replica-Set-Need-An-Arbiter/

 From my understanding arbiter is good for resolving split brains.
 Quorum and arbiter are two different things though quorum is a mechanism to
 help you **avoid** split brain and the arbiter is to help gluster resolve
 split brain by voting and other internal mechanics (as outlined in link 1).
 How did you create the volume exactly - what command? It looks to me like
 you created it with 'gluster volume create replica 2 arbiter 1 {}' per
 your earlier mention of "replica 2 arbiter 1". That being said, if you did
 that and then setup quorum in the volume configuration, this would cause
 your gluster to halt up since quorum was lost (as you saw until you
 recovered node 1)

 As you can see from the docs, there is still a corner case for getting
 in to split brain with replica 3, which again, is where arbiter would help
 gluster resolve it

 I need to amend my previous statement: I was told that arbiter volume
 does not store data, only metadata. I cannot find anything in the docs
 backing this up however it would make sense for it to be. That being said,
 in my setup, I would not include my arbiter or my third node in my ovirt VM
 cluster component. I would keep it completely separate


 On Fri, Sep 1, 2017 at 2:46 PM,