[ovirt-users] Re: ovirt 4.2.7-1 - adding virtual host ( nested virt. )

2019-01-10 Thread p...@email.cz

Hello,
problem was fixed now. Missing nested virtualization =  enabled option. 
( set on VM config > CPU -> hardware virtualization )

Thx for response.
Pa.


On 1/10/19 4:39 PM, Shani Leviim wrote:

Hi Paul,

Can you please attach vdsm & engine logs?

*Regards,
*
*Shani Leviim
*


On Mon, Jan 7, 2019 at 8:44 PM p...@email.cz <mailto:p...@email.cz> 
mailto:p...@email.cz>> wrote:


Hello guys,
I've got problem with adding new host (ESX-virtual) to ovirt
4.2.7-1  ( gluster included)

Is this feature supported ???

2019-01-07 19:38:30,168+01 ERROR
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
(DefaultQuartzScheduler1) [15a4029b] Error while refreshing server
data for cluster 'MID' from database: null

regs.
Paul
___
Users mailing list -- users@ovirt.org <mailto:users@ovirt.org>
To unsubscribe send an email to users-le...@ovirt.org
<mailto:users-le...@ovirt.org>
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:

https://lists.ovirt.org/archives/list/users@ovirt.org/message/TREULSOFHYWLGLK7H557NT4WEYPBUXPU/



___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RFH43FBXQOFRA6WTSLOMTQTS2N6OY5VA/


[ovirt-users] ovirt 4.2.7-1 - adding virtual host ( nested virt. )

2019-01-07 Thread p...@email.cz

Hello guys,
I've got problem with adding new host (ESX-virtual) to ovirt 4.2.7-1  ( 
gluster included)


Is this feature supported ???

2019-01-07 19:38:30,168+01 ERROR 
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
(DefaultQuartzScheduler1) [15a4029b] Error while refreshing server data 
for cluster 'MID' from database: null


regs.
Paul
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/TREULSOFHYWLGLK7H557NT4WEYPBUXPU/


Re: [ovirt-users] VDSM hang

2017-03-13 Thread p...@email.cz
 0,00 0,000,000,00 0,00 0,00 
0,00 0,000,000,000,00   0,00 0,00
dm-17 0,00 0,000,000,00 0,00 0,00 
0,00 0,000,000,000,00   0,00 0,00
dm-18 0,00 0,000,000,00 0,00 0,00 
0,00 0,000,000,000,00   0,00 0,00


look w_await ( write queue )

concurent tasks:
- The data import is about 1,5Gb over then 10Gb LAN
- VM clone on the same volume ( 300G )
- VM snapshot deleting on the same volume ( cca 1TB )
- VM with Oracle DB  start/stop
- VM seq.write ( 500GB )

Yes - it could  happen  in one time


regards
Paf1

On 03/13/2017 06:01 AM, Krutika Dhananjay wrote:

Hi,

Could you please share your volume info output?

-Krutika

On Fri, Mar 10, 2017 at 6:41 PM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


freez / freezing
IO operations are paused from any reasons
available posibilities are
1) net - any tcp framework collapse
2) gluster interconnect due gluster daemon - process hang ??
3) VSD - pause managed services
4) XFS - RW issues
5) swap overfulled - any processes are killed - but why swap is
full if max 30% of mem (196 GB )  is used by VMs ? ( unmanaged
process forking )

regs


On 03/10/2017 01:56 PM, Nir Soffer wrote:

    On Fri, Mar 10, 2017 at 1:07 PM,p...@email.cz <mailto:p...@email.cz>  
<p...@email.cz> <mailto:p...@email.cz>  wrote:

Hello everybody,

for production usage i'm testing  ovirt with gluster.
All components seems to be running fine but whenever I'm testing huge
workload, then node freez. Not the main OS, but VDSM mgmt and attached
services, VMs eg.

What do you mean by freez?


mgmt
oVirt - 4.1.0.4
centos 7.3-1611


nodes ( installed from ovirt image
"ovirt-node-ng-installer-ovirt-4.1-2017030804.iso"  )

OS Version: == RHEL - 7 - 3.1611.el7.centos
OS Description:== oVirt Node 4.1.0
Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64
KVM Version:== 2.6.0 - 28.el7_3.3.1
LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5
VDSM Version:== vdsm-4.19.4-1.el7.centos
SPICE Version:== 0.12.4 - 20.el7_3
GlusterFS Version:== glusterfs-3.8.9-1.el7  ( LVM thinprovisioning in
replica 2 - created from ovirt GUI )

concurently running
- huge import from export domain( net workload )
- sequential write to VMs local disk ( gluster replica sequential workload )
- VMs database huge select  (  random IOps )
- huge old snapshot delete  ( random IOps )

In this configuration / workload  is  runnig one hour eg, with no exceptions
, with 70-80% disk load, but in some point VDSM freez  all jobs for a
timeout and VM's are in "uknown" status .
The whole system revitalize then automaticaly in cca 20min time frame (
except the import and snapshot deleting(rollback) )

engine.log  - focus 10:39:07 time  ( Failed in 'HSMGetAllTasksStatusesVDS'
method )


n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:01,727+01 INFO

[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler2) [759c8e1f] Command 'RemoveSnapshotSingleDiskLive'
(id: 'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id:
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete
2017-03-10 10:39:03,929+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3
2017-03-10 10:39:04,343+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED 
<http://172.16.5.163/24:CONNECTED>,
16.0.0.164:CONNECTED], log id: 446d0cd3
2017-03-10 10:39:04,353+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda
2017-03-10 10:39:05,128+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterVolumesListVDSCommand, return:

{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.glus

Re: [ovirt-users] VDSM hang

2017-03-10 Thread p...@email.cz

freez / freezing
IO operations are paused from any reasons
available posibilities are
1) net - any tcp framework collapse
2) gluster interconnect due gluster daemon - process hang ??
3) VSD - pause managed services
4) XFS - RW issues
5) swap overfulled - any processes are killed - but why swap is full if 
max 30% of mem (196 GB )  is used by VMs ? ( unmanaged process forking )


regs


On 03/10/2017 01:56 PM, Nir Soffer wrote:

On Fri, Mar 10, 2017 at 1:07 PM, p...@email.cz <p...@email.cz> wrote:

Hello everybody,

for production usage i'm testing  ovirt with gluster.
All components seems to be running fine but whenever I'm testing huge
workload, then node freez. Not the main OS, but VDSM mgmt and attached
services, VMs eg.

What do you mean by freez?


mgmt
oVirt - 4.1.0.4
centos 7.3-1611


nodes ( installed from ovirt image
"ovirt-node-ng-installer-ovirt-4.1-2017030804.iso"  )

OS Version: == RHEL - 7 - 3.1611.el7.centos
OS Description:== oVirt Node 4.1.0
Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64
KVM Version:== 2.6.0 - 28.el7_3.3.1
LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5
VDSM Version:== vdsm-4.19.4-1.el7.centos
SPICE Version:== 0.12.4 - 20.el7_3
GlusterFS Version:== glusterfs-3.8.9-1.el7  ( LVM thinprovisioning in
replica 2 - created from ovirt GUI )

concurently running
- huge import from export domain( net workload )
- sequential write to VMs local disk ( gluster replica sequential workload )
- VMs database huge select  (  random IOps )
- huge old snapshot delete  ( random IOps )

In this configuration / workload  is  runnig one hour eg, with no exceptions
, with 70-80% disk load, but in some point VDSM freez  all jobs for a
timeout and VM's are in "uknown" status .
The whole system revitalize then automaticaly in cca 20min time frame (
except the import and snapshot deleting(rollback) )

engine.log  - focus 10:39:07 time  ( Failed in 'HSMGetAllTasksStatusesVDS'
method )


n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508'
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:01,727+01 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(DefaultQuartzScheduler2) [759c8e1f] Command 'RemoveSnapshotSingleDiskLive'
(id: 'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id:
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete
2017-03-10 10:39:03,929+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterServersListVDSCommand(HostName = 2kvm1,
VdsIdVDSCommandParametersBase:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3
2017-03-10 10:39:04,343+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED,
16.0.0.164:CONNECTED], log id: 446d0cd3
2017-03-10 10:39:04,353+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START,
GlusterVolumesListVDSCommand(HostName = 2kvm1,
GlusterVolumesListVDSParameters:{runAsync='true',
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda
2017-03-10 10:39:05,128+01 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH,
GlusterVolumesListVDSCommand, return:
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad},
log id: 69ea1fda
2017-03-10 10:39:07,163+01 ERROR
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand]
(DefaultQuartzScheduler2) [759c8e1f] Failed in 'HSMGetAllTasksStatusesVDS'
method
2017-03-10 10:39:07,178+01 ERROR
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(DefaultQuartzScheduler2) [759c8e1f] EVENT_ID:
VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: null, Call Stack: null,
Custom Event ID: -1, Message: VDSM 2kvm2 command HSMGetAllTasksStatusesVDS
failed: Connection timed out
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2)
[759c8e1f] BaseAsyncTask::onTaskEndSuccess: Task
'f594bf69-619b-4d1b-8f6d-a9826997e478' (Parent Command 'ImportVm',
Parameters Type
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') ended
successfully.
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.core.bll.CommandMultiAsyncTasks] (DefaultQuartzScheduler2)
[759c8e1f] Task with DB Task ID 'a05c7c07-9b98-4ab2-ac7b-9e70a75ba7b7' and
VDSM Task ID '7c60369f-70a3-4a6a-80c9-4753ac9ed372' is in state Polling. End
action for command 8deb3fe3-4a83-4605-816c-ffdc63fd9ac1 will proceed when
all the entity's tasks are completed.
2017-03-10 10:39:07,182+01 INFO
[org.ovirt.engine.c

[ovirt-users] VDSM hang

2017-03-10 Thread p...@email.cz

Hello everybody,

for production usage i'm testing  ovirt with gluster.
All components seems to be running fine but whenever I'm testing huge 
workload, then node freez. Not the main OS, but VDSM mgmt and attached 
services, VMs eg.


*mgmt *
oVirt - 4.1.0.4
centos 7.3-1611


*nodes* ( installed from ovirt image 
/"ovirt-node-ng-installer-ovirt-4.1-2017030804.iso" )/


OS Version:== RHEL - 7 - 3.1611.el7.centos
OS Description:== oVirt Node 4.1.0
Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64
KVM Version:== 2.6.0 - 28.el7_3.3.1
LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5
VDSM Version:== vdsm-4.19.4-1.el7.centos
SPICE Version:== 0.12.4 - 20.el7_3
GlusterFS Version:== glusterfs-3.8.9-1.el7  ( LVM thinprovisioning in 
replica 2 - created from ovirt GUI )


concurently running
- huge import from export domain( net workload )
- sequential write to VMs local disk ( gluster replica sequential workload )
- VMs database huge select  (  random IOps )
- huge old snapshot delete  ( random IOps )

In this configuration / workload  is  runnig one hour eg, with no 
exceptions , with 70-80% disk load, but in some point VDSM freez all 
jobs for a timeout and VM's are in "uknown" status .
The whole system revitalize then automaticaly in cca 20min time frame ( 
except the import and snapshot deleting(rollback) )


engine.log  - focus 10:39:07 time  ( Failed in 
'HSMGetAllTasksStatusesVDS' method )



n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508' 
type:'RemoveSnapshotSingleDiskLive' to complete
2017-03-10 10:39:01,727+01 INFO 
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback] 
(DefaultQuartzScheduler2) [759c8e1f] Command 
'RemoveSnapshotSingleDiskLive' (id: 
'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id: 
'33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete
2017-03-10 10:39:03,929+01 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] 
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, 
GlusterServersListVDSCommand(HostName = 2kvm1, 
VdsIdVDSCommandParametersBase:{runAsync='true', 
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3
2017-03-10 10:39:04,343+01 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] 
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, 
GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED, 
16.0.0.164:CONNECTED], log id: 446d0cd3
2017-03-10 10:39:04,353+01 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, 
GlusterVolumesListVDSCommand(HostName = 2kvm1, 
GlusterVolumesListVDSParameters:{runAsync='true', 
hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda
2017-03-10 10:39:05,128+01 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, 
GlusterVolumesListVDSCommand, return: 
{8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad}, 
log id: 69ea1fda
2017-03-10 10:39:07,163+01 ERROR 
[org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand] 
(DefaultQuartzScheduler2) [759c8e1f] Failed in 
'HSMGetAllTasksStatusesVDS' method
2017-03-10 10:39:07,178+01 ERROR 
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(DefaultQuartzScheduler2) [759c8e1f] EVENT_ID: 
VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: null, Call Stack: 
null, Custom Event ID: -1, Message: VDSM 2kvm2 command 
HSMGetAllTasksStatusesVDS failed: Connection timed out
2017-03-10 10:39:07,182+01 INFO 
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) 
[759c8e1f] BaseAsyncTask::onTaskEndSuccess: Task 
'f594bf69-619b-4d1b-8f6d-a9826997e478' (Parent Command 'ImportVm', 
Parameters Type 
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') ended 
successfully.
2017-03-10 10:39:07,182+01 INFO 
[org.ovirt.engine.core.bll.CommandMultiAsyncTasks] 
(DefaultQuartzScheduler2) [759c8e1f] Task with DB Task ID 
'a05c7c07-9b98-4ab2-ac7b-9e70a75ba7b7' and VDSM Task ID 
'7c60369f-70a3-4a6a-80c9-4753ac9ed372' is in state Polling. End action 
for command 8deb3fe3-4a83-4605-816c-ffdc63fd9ac1 will proceed when all 
the entity's tasks are completed.
2017-03-10 10:39:07,182+01 INFO 
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) 
[759c8e1f] SPMAsyncTask::PollTask: Polling task 
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent Command 'DestroyImage', 
Parameters Type 
'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') returned 
status 'finished', result 'cleanSuccess'.
2017-03-10 10:39:07,182+01 ERROR 
[org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) 
[759c8e1f] BaseAsyncTask::logEndTaskFailure: Task 
'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent 

[ovirt-users] GUI node detail long delay

2017-02-06 Thread p...@email.cz

Hello everybody,

We are using oVirt Engine Version: 4.0.6.3-1.el7.centos on centos 7.3 
with gluster replica 3 arbiter  = (1+1)+1


I'm confused with GUI delaying -  if node details are wanted ( cluster 
-> nodes -> node detail = click on node raw ) then request generate over 
10 min delay to display details. This unexpected mistake  didn't occure 
initially, but later - not specified when .


The followed partial list of "engine.log" shows requests to "arbiter 
node" ( 16.0.0.159)  connectivity.
This requested 3rd node of gluster(arbiter)  is NOT included in oVirt 
environment and will NOT.

Maybe this is that problem, but I'm not shure, especially how to fix this.

2017-02-06 13:20:03,924 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] 
(DefaultQuartzScheduler3) [49cebf0] START, 
GlusterServersListVDSCommand(HostName = 1kvm2, 
VdsIdVDSCommandParametersBase:{runAsync='true', 
hostId='258decac-46f4-4c15-b855-ad97b570ee60'}), log id: 6873151
2017-02-06 13:20:04,796 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] 
(DefaultQuartzScheduler3) [49cebf0] FINISH, 
GlusterServersListVDSCommand, return: [172.16.5.162/24:CONNECTED, 
172.16.5.161:CONNECTED, 16.0.0.159:CONNECTED], log id: 6873151
2017-02-06 13:20:04,814 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler3) [49cebf0] START, 
GlusterVolumesListVDSCommand(HostName = 1kvm2, 
GlusterVolumesListVDSParameters:{runAsync='true', 
hostId='258decac-46f4-4c15-b855-ad97b570ee60'}), log id: 381ae630
2017-02-06 13:20:05,970 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] 
(DefaultQuartzScheduler3) [49cebf0] Could not add brick 
'16.0.0.159:/GLUSTER/1KVM12-sda2/GFS' to volume 
'19c27787-f1c9-4dee-8415-c6d1c81e3aa2' - server uuid 
'f7670ea9-2204-4310-96a6-243c2c6a00de' not found in cluster 
'587fa2d8-017d-03b3-0003-030d'
2017-02-06 13:20:05,987 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] 
(DefaultQuartzScheduler3) [49cebf0] Could not add brick 
'16.0.0.159:/GLUSTER/1KVM12-sda1/GFS' to volume 
'96adac2a-0dc4-4bd8-ad79-23dd3448f73b' - server uuid 
'f7670ea9-2204-4310-96a6-243c2c6a00de' not found in cluster 
'587fa2d8-017d-03b3-0003-030d'
2017-02-06 13:20:05,987 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler3) [49cebf0] FINISH, 
GlusterVolumesListVDSCommand, return: 
{19c27787-f1c9-4dee-8415-c6d1c81e3aa2=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@b9f51962, 
96adac2a-0dc4-4bd8-ad79-23dd3448f73b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@86597dda}, 
log id: 381ae630


repeatelly occured several times per minute, so huge filling logs


OS Version:RHEL - 7 - 3.1611.el7.centos
OS Description:CentOS Linux 7 (Core)
Kernel Version:3.10.0 - 514.6.1.el7.x86_64
KVM Version:2.6.0 - 28.el7_3.3.1
LIBVIRT Version:libvirt-2.0.0-10.el7_3.4
VDSM Version:vdsm-4.18.21-1.el7.centos
SPICE Version:0.12.4 - 19.el7
GlusterFS Version:glusterfs-3.8.8-1.el7
CEPH Version:librbd1-0.94.5-1.el7


regards
Paf1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] fast import to ovirt

2017-01-19 Thread p...@email.cz

Hello,
how can I import Vm from different ovirt envir..? There is no common 
mgmt ovirt. ( ovirt 3.5 -> 4.0 )

Gluster FS used.
Will ovirt accept "rsync" file migrations , meaning will update oVirt DB 
automaticaly  ?

I'd prefer more quickly method then export-umount oV1-mount oV2-import .

regards
paf1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt with glusterfs - big test - unwanted results

2016-03-31 Thread p...@email.cz

Hello Yaniv,

we tried another small test - reboot two nodes from replica3 a1 ( 
1HP12-R3A1P1 ) which keep master domain.
All domains went down  = master down, but master domain didn't move to 
another available domain ( eg. 2HP12-R3A1P1 ).


It looks that "master domain" management isn't correct ( has a bug ?? )

regs.
Pavel


On 31.3.2016 14:30, Yaniv Kaul wrote:

Hi Pavel,

Thanks for the report. Can you begin with a more accurate description 
of your environment?
Begin with host, oVirt and Gluster versions. Then continue with the 
exact setup (what are 'A', 'B', 'C' - domains? Volumes? What is the 
mapping between domains and volumes?).


Are there any logs you can share with us?

I'm sure with more information, we'd be happy to look at the issue.
Y.


On Thu, Mar 31, 2016 at 3:09 PM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


Hello,
we tried the  following test - with unwanted results

input:
5 node gluster
A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 )
B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 )
C = distributed replica 3 arbiter 1  ( node1+node2, node3+node4,
each arbiter on node 5)
node 5 has only arbiter replica ( 4x )

TEST:
1)  directly reboot one node - OK ( is not important which ( data
node or arbiter node ))
2)  directly reboot two nodes - OK ( if  nodes are not from the
same replica )
3)  directly reboot three nodes - yes, this is the main problem
and a questions 
- rebooted all three nodes from replica "B"  ( not so
possible, but who knows ... )
- all VMs with data on this replica was paused ( no data
access ) - OK
- all VMs running on replica "B" nodes lost (  started
manually, later )( datas on other replicas ) - acceptable
BUT
- !!! all oVIrt domains went down !! - master domain is on
replica "A" which lost only one member from three !!!
so we are not expecting that all domain will go down,
especially master with 2 live members.

Results:
- the whole cluster unreachable until at all domains up -
depent of all nodes up !!!
- all paused VMs started back - OK
- rest of all VMs rebooted and runnig - OK

Questions:
1) why all domains down if master domain ( on replica "A" )
has two runnig members ( 2 of 3 )  ??
2) how to fix that colaps without waiting to all nodes up ? (
in worste case if node has HW error eg. ) ??
3) which oVirt  cluster  policy  can prevent that situation ??
( if any )

regs.
Pavel



___
Users mailing list
Users@ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt with glusterfs - big test - unwanted results

2016-03-31 Thread p...@email.cz

Hi,
rest of logs:
www.uschovna.cz/en/zasilka/HYGXR57CNHM3TP39-L3W 
<http://www.uschovna.cz/en/zasilka/HYGXR57CNHM3TP39-L3W>


The TEST is the last big event in logs 
TEST TIME : about 14:00-14:30  CET

regs.Pavel

On 31.3.2016 14:30, Yaniv Kaul wrote:

Hi Pavel,

Thanks for the report. Can you begin with a more accurate description 
of your environment?
Begin with host, oVirt and Gluster versions. Then continue with the 
exact setup (what are 'A', 'B', 'C' - domains? Volumes? What is the 
mapping between domains and volumes?).


Are there any logs you can share with us?

I'm sure with more information, we'd be happy to look at the issue.
Y.


On Thu, Mar 31, 2016 at 3:09 PM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


Hello,
we tried the  following test - with unwanted results

input:
5 node gluster
A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 )
B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 )
C = distributed replica 3 arbiter 1  ( node1+node2, node3+node4,
each arbiter on node 5)
node 5 has only arbiter replica ( 4x )

TEST:
1)  directly reboot one node - OK ( is not important which ( data
node or arbiter node ))
2)  directly reboot two nodes - OK ( if  nodes are not from the
same replica )
3)  directly reboot three nodes - yes, this is the main problem
and a questions 
- rebooted all three nodes from replica "B"  ( not so
possible, but who knows ... )
- all VMs with data on this replica was paused ( no data
access ) - OK
- all VMs running on replica "B" nodes lost (  started
manually, later )( datas on other replicas ) - acceptable
BUT
- !!! all oVIrt domains went down !! - master domain is on
replica "A" which lost only one member from three !!!
so we are not expecting that all domain will go down,
especially master with 2 live members.

Results:
- the whole cluster unreachable until at all domains up -
depent of all nodes up !!!
- all paused VMs started back - OK
- rest of all VMs rebooted and runnig - OK

Questions:
1) why all domains down if master domain ( on replica "A" )
has two runnig members ( 2 of 3 )  ??
2) how to fix that colaps without waiting to all nodes up ? (
in worste case if node has HW error eg. ) ??
3) which oVirt  cluster  policy  can prevent that situation ??
( if any )

regs.
Pavel



___
Users mailing list
Users@ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt with glusterfs - big test - unwanted results

2016-03-31 Thread p...@email.cz

Hello,
some envir. answers :
*
OS = RHEL - 7 - 2.151
kernel = 3.10.0 - 327.10.1.el7.x86_64
KVM = 2.3.0 - 31.el7_2.7.1
libvirt = libvirt-1.2.17-13.el7_2.3
vdsm = vdsm-4.17.23.2-0.el7
glusterfs = glusterfs-3.7.9-1.el7
ovirt = 3.5.6.2-1
*
# gluster peer status
Number of Peers: 4

Hostname: 1hp2
Uuid: 8e87cf18-8958-41b7-8d24-7ee420a1ef9f
State: Peer in Cluster (Connected)

Hostname: 2hp2
Uuid: b1d987d8-0b42-4ce4-b85f-83b4072e0990
State: Peer in Cluster (Connected)

Hostname: 2hp1
Uuid: a1cbe1a8-88ad-4e89-8a0e-d2bb2b6786d8
State: Peer in Cluster (Connected)

Hostname: kvmarbiter
Uuid: bb1d63f1-7757-4c07-b70d-aa2f68449e21
State: Peer in Cluster (Connected)
*
== "C" ==
Volume Name: 12HP12-D2R3A1P2
Type: Distributed-Replicate
Volume ID: 3c22d3dc-7c6e-4e37-9e0b-78410873ed6d
Status: Started
Number of Bricks: 2 x (2 + 1) = 6
Transport-type: tcp
Bricks:
Brick1: 1hp1:/STORAGES/P2/GFS
Brick2: 1hp2:/STORAGES/P2/GFS
Brick3: kvmarbiter:/STORAGES/P2-1/GFS (arbiter)
Brick4: 2hp1:/STORAGES/P2/GFS
Brick5: 2hp2:/STORAGES/P2/GFS
Brick6: kvmarbiter:/STORAGES/P2-2/GFS (arbiter)
Options Reconfigured:
performance.readdir-ahead: on
*
== "A" ==
Volume Name: 1HP12-R3A1P1
Type: Replicate
Volume ID: e4121610-6128-4ecc-86d3-1429ab3b8356
Status: Started
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: 1hp1:/STORAGES/P1/GFS
Brick2: 1hp2:/STORAGES/P1/GFS
Brick3: kvmarbiter:/STORAGES/P1-1/GFS (arbiter)
Options Reconfigured:
performance.readdir-ahead: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
storage.owner-uid: 36
storage.owner-gid: 36
features.shard: on
features.shard-block-size: 512MB
cluster.data-self-heal-algorithm: full
performance.write-behind: on
performance.low-prio-threads: 32
performance.write-behind-window-size: 128MB
network.ping-timeout: 10
*
== "B" ==
Volume Name: 2HP12-R3A1P1
Type: Replicate
Volume ID: d3d260cd-455f-42d6-9580-d88ae6df0519
Status: Started
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: 2hp1:/STORAGES/P1/GFS
Brick2: 2hp2:/STORAGES/P1/GFS
Brick3: kvmarbiter:/STORAGES/P1-2/GFS (arbiter)
Options Reconfigured:
performance.readdir-ahead: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
storage.owner-uid: 36
storage.owner-gid: 36
features.shard: on
features.shard-block-size: 512MB
cluster.data-self-heal-algorithm: full
performance.write-behind: on
performance.low-prio-threads: 32
performance.write-behind-window-size: 128MB
network.ping-timeout: 10


The oVirt volumes(storages) have the same name as gluster volumes ( eg: 
"B" = 2HP12-R3A1P1( ovirt storage ) = 2HP12-R3A1P1( gluster volume name ) )

In the test the master volume was  "A" = 1HP12-R3A1P1

regs. Pavel
PS: logs will follow as webstore pointer ... this takes some time


On 31.3.2016 14:30, Yaniv Kaul wrote:

Hi Pavel,

Thanks for the report. Can you begin with a more accurate description 
of your environment?
Begin with host, oVirt and Gluster versions. Then continue with the 
exact setup (what are 'A', 'B', 'C' - domains? Volumes? What is the 
mapping between domains and volumes?).


Are there any logs you can share with us?

I'm sure with more information, we'd be happy to look at the issue.
Y.


On Thu, Mar 31, 2016 at 3:09 PM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


Hello,
we tried the  following test - with unwanted results

input:
5 node gluster
A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 )
B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 )
C = distributed replica 3 arbiter 1  ( node1+node2, node3+node4,
each arbiter on node 5)
node 5 has only arbiter replica ( 4x )

TEST:
1)  directly reboot one node - OK ( is not important which ( data
node or arbiter node ))
2)  directly reboot two nodes - OK ( if  nodes are not from the
same replica )
3)  directly reboot three nodes - yes, this is the main problem
and a questions 
- rebooted all three nodes from replica "B"  ( not so
possible, but who knows ... )
- all VMs with data on this replica was paused ( no data
access ) - OK
- all VMs running on replica "B" nodes lost (  started
manually, later )( datas on other replicas ) - acceptable
BUT
- !!! all oVIrt domains 

[ovirt-users] ovirt with glusterfs - big test - unwanted results

2016-03-31 Thread p...@email.cz

Hello,
we tried the  following test - with unwanted results

input:
5 node gluster
A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 )
B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 )
C = distributed replica 3 arbiter 1  ( node1+node2, node3+node4, each 
arbiter on node 5)

node 5 has only arbiter replica ( 4x )

TEST:
1)  directly reboot one node - OK ( is not important which ( data node 
or arbiter node ))
2)  directly reboot two nodes - OK ( if  nodes are not from the same 
replica )
3)  directly reboot three nodes - yes, this is the main problem and a 
questions 
- rebooted all three nodes from replica "B"  ( not so possible, but 
who knows ... )

- all VMs with data on this replica was paused ( no data access ) - OK
- all VMs running on replica "B" nodes lost (  started manually, 
later )( datas on other replicas ) - acceptable

BUT
- !!! all oVIrt domains went down !! - master domain is on replica 
"A" which lost only one member from three !!!
so we are not expecting that all domain will go down, especially 
master with 2 live members.


Results:
- the whole cluster unreachable until at all domains up - depent of 
all nodes up !!!

- all paused VMs started back - OK
- rest of all VMs rebooted and runnig - OK

Questions:
1) why all domains down if master domain ( on replica "A" ) has two 
runnig members ( 2 of 3 )  ??
2) how to fix that colaps without waiting to all nodes up ? ( in 
worste case if node has HW error eg. ) ??
3) which oVirt  cluster  policy  can prevent that situation ?? ( if 
any )


regs.
Pavel


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt - unsupported GFS config ??

2016-03-28 Thread p...@email.cz
- Yes, database will run in VM , but with RMAN online backups. So VM 
takes  over 1,4TB . This VM wil  run 5-6x   :o(
- I have no special issues with  512MB shard size, looks OK but not 
tested for such databases yet ( Waiting for 3.7.10 gluster (dupl.healing 
fix))


regs.
Pavel

On 28.3.2016 08:30, Sahina Bose wrote:




On 03/27/2016 10:24 PM, p...@email.cz wrote:
What's the recommended shard size for databases ( espacially Oracle ) 
, I'm afraid that 512M is too large.
I found that someones using  about 16MB , but it generate a lot of 
files for healing if volumes  splitted. (eg. for 500GB DB in worse case )


Would the database be running within the guest VM?

Did you run into any specific issue with 512M shard size?
What we have noticed is that with smaller shard sizes , like 4MB, the 
number of entries in the .shard directory is too high, and hence 
affects the performance when it comes to healing entries in directory. 
The 256M/512M shard size is a good balance between the number of 
entries created and data size to heal.




Pa.

On 27.3.2016 16:57, Sahina Bose wrote:

Stripe is not supported.

What you need to do instead is turn on sharding for the volume.

So:

gluster volume create 12HP12-S2R3A1P2 replica 3 arbiter 1 
1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS 
kvmarbiter:/STORAGES/P2-1/GFS  force


gluster volume set 12HP12-S2R3A1P2 features.shard on
gluster volume set 12HP12-S2R3A1P2 features.shard-block-size 512MB

If you want to utilize the additional nodes as well, you can change 
this to a distributed replicate volume - instead of the volume 
creation in step above , use below


gluster volume create 12HP12-S2R3A1P2 replica 3 arbiter 1 
1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS 
kvmarbiter:/STORAGES/P2-1/GFS 2hp1:/STORAGES/P2/GFS 
2hp2:/STORAGES/P2/GFS  kvmarbiter:/STORAGES/P2-2/GFS  force



On 03/24/2016 07:49 PM, p...@email.cz wrote:

Hello,
I tried create stripe 2 replica 3 arbiter1 gluster volume for testing.
So , glusterFS such type from commandline was successfull, but 
domain creation looks to be unsupported. with oVirt message "Error 
while executing action AddGlusterFsStorageDomain: Storage Domain 
target is unsupported".

Cam U tell me if is it  error or really unsuported ??

exam:
gluster volume create 12HP12-S2R3A1P2 stripe 2 replica 3 arbiter 1 
1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS 
kvmarbiter:/STORAGES/P2-1/GFS 2hp1:/STORAGES/P2/GFS 
2hp2:/STORAGES/P2/GFS  kvmarbiter:/STORAGES/P2-2/GFS  force



RHEL 7-2.1511
vdsm - vdsm-4.17.23-1.el7
gluster - glusterfs-3.7.9-1.el7
ovirt - 3.5.6.2-1

regs.Pavel


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users








___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] ovirt - unsupported GFS config ??

2016-03-24 Thread p...@email.cz

Hello,
I tried create stripe 2 replica 3 arbiter1 gluster volume for testing.
So , glusterFS such type from commandline was successfull, but domain 
creation looks to be unsupported. with oVirt message "Error while 
executing action AddGlusterFsStorageDomain: Storage Domain target is 
unsupported".

Cam U tell me if is it  error or really unsuported ??

exam:
gluster volume create 12HP12-S2R3A1P2 stripe 2 replica 3 arbiter 1 
1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS 
kvmarbiter:/STORAGES/P2-1/GFS 2hp1:/STORAGES/P2/GFS 
2hp2:/STORAGES/P2/GFS  kvmarbiter:/STORAGES/P2-2/GFS  force



RHEL 7-2.1511
vdsm - vdsm-4.17.23-1.el7
gluster - glusterfs-3.7.9-1.el7
ovirt - 3.5.6.2-1

regs.Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] engine.log messages in cycle

2016-03-24 Thread p...@email.cz

Hello,
can anybody explain why is engine.log filled with following messages, 
even thought this is fresh installation ??
Especially START / FINISH rows in cycle, ... in cycle,  in cycle  
 - it takes a lot of space and is really needed ??


RHEL 7-2.1511
vdsm - vdsm-4.17.23-1.el7
gluster - glusterfs-3.7.9-1.el7
ovirt - 3.5.6.2-1

regs.Pavel


2016-03-24 13:39:10,758 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] 
(DefaultQuartzScheduler_Worker-89) START, 
GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 2hp1, HostId = 
45f76a0f-9616-420a-be1d-afbed2954562), log id: 3d6b27fd
2016-03-24 13:39:13,243 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] 
(DefaultQuartzScheduler_Worker-89) FINISH, 
GetGlusterVolumeAdvancedDetailsVDSCommand, return: 
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@11d55805, 
log id: 3d6b27fd
2016-03-24 13:39:13,278 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] 
(DefaultQuartzScheduler_Worker-89) START, 
GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 2hp1, HostId = 
45f76a0f-9616-420a-be1d-afbed2954562), log id: 3ae0e479
2016-03-24 13:39:13,349 INFO 
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
(DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock 
EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e 
value: GLUSTER

, sharedLocks= ]
2016-03-24 13:39:13,444 INFO 
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
(DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock 
EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e 
value: GLUSTER

, sharedLocks= ]
2016-03-24 13:39:13,801 INFO 
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
(DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock 
EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e 
value: GLUSTER

, sharedLocks= ]
2016-03-24 13:39:14,646 INFO 
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
(DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock 
EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e 
value: GLUSTER

, sharedLocks= ]
2016-03-24 13:39:15,630 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] 
(DefaultQuartzScheduler_Worker-89) FINISH, 
GetGlusterVolumeAdvancedDetailsVDSCommand, return: 
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@cab5100f, 
log id: 3ae0e479
2016-03-24 13:39:15,656 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler_Worker-43) START, 
GlusterVolumesListVDSCommand(HostName = 1hp2, HostId = 
184ebfaa-51a9-43e4-a57b-9d4f03e85b47), log id: 5756f325
2016-03-24 13:39:16,105 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler_Worker-43) FINISH, GlusterVolumesListVDSCommand, 
return: 
{e4121610-6128-4ecc-86d3-1429ab3b8356=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@edd4741e, 
d3d260cd-455f-42d6-9580-d88ae6df0519=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@84ea7412}, 
log id: 5756f325
2016-03-24 13:39:21,161 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler_Worker-81) START, 
GlusterVolumesListVDSCommand(HostName = 1hp2, HostId = 
184ebfaa-51a9-43e4-a57b-9d4f03e85b47), log id: 6e2f6c69
2016-03-24 13:39:21,667 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler_Worker-81) FINISH, GlusterVolumesListVDSCommand, 
return: 
{e4121610-6128-4ecc-86d3-1429ab3b8356=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@2ad03bfa, 
d3d260cd-455f-42d6-9580-d88ae6df0519=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@6a99fa09}, 
log id: 6e2f6c69


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] delete storage definition

2016-03-24 Thread p...@email.cz

Hi Maor,

it helped, great !

thx a lot , U spare my time .
regs.Pa.

On 24.3.2016 12:21, Maor Lipchuk wrote:



On Thu, Mar 24, 2016 at 12:20 PM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


Hi folks,
how can I delete the last storage  definition from oVirt database
if the last  volume has been deleted from bricks commandline ( rm
-rf  < path to that volume > ) directly ?
In oVirt DB exists this storage last record and blocking create
new storage operation ( ovirt offering " delete datacenter", but
this is not the right way for me, now )
regs. Pavel

___
Users mailing list
Users@ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users


Hi Pavel,

What are your plans regarding that Data Center?
In case you want to keep that Data Center to be used with other 
storage domains you can try to add a new storage domain without 
attaching it to any Data Center and try to re-initialize the Data 
Center with this new storage domain. Once that Data Center will be 
re-initialized you can try to remove the old Storage Domain (or force 
remove it you encounter any problem)
Please let me know if this helps you or is there anything else that 
you were trying to do


Regards,
Maor


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] delete storage definition

2016-03-24 Thread p...@email.cz

Hi folks,
how can I delete the last storage  definition from oVirt database if the 
last  volume has been deleted from bricks commandline ( rm -rf < path to 
that volume > ) directly ?
In oVirt DB exists this storage last record and blocking create new 
storage operation ( ovirt offering " delete datacenter", but this is not 
the right way for me, now )

regs. Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] question mark on VM ( DB status 8 )

2016-03-19 Thread p...@email.cz

I used that, but lock active in a few seconds again.
And oVirt do not update any VM's status

Pa.

On 17.3.2016 10:26, Eli Mesika wrote:





*From: *p...@email.cz <mailto:p...@email.cz>
*To: *"users" <users@ovirt.org <mailto:users@ovirt.org>>
*Sent: *Thursday, March 17, 2016 9:27:11 AM
*Subject: *[ovirt-users] question mark on VM ( DB status 8 )

Hello,
during backup VM hanged with question mark in ovirt and status 8
in DB, snapshot file ( for backup )is locked.
How to clean snapshot locking a wake up this VM from "unknow"
state ???


Try using the unlock_entity.sh utility (run with --help for usage)



regs.
pavel

___
Users mailing list
Users@ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt - rescan new copied disks to storage

2016-03-19 Thread p...@email.cz

Hello,
how can I put a copy of VM disks ( outside ovirt envir. copy ) to ovirt 
inventory ?? ( available  for " attach disk"  option visibility  )


regs.
Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] oVirt - rescan new copied disks to storage

2016-03-18 Thread p...@email.cz

Hi Nir,
thx for info.

YES I mean import raw disk (  raw image , not raw device, not v2v utils )
such as  " 
/rhev/data-center/mnt/glusterSD/localhost:DOMAIN/aaa/image//cc 
" file which represents the VM disk.


regs.
Pavel

On 17.3.2016 16:03, Nir Soffer wrote:

On Thu, Mar 17, 2016 at 4:37 PM, p...@email.cz <p...@email.cz> wrote:

Hello,
how can I put a copy of VM disks ( outside ovirt envir. copy ) to ovirt
inventory ?? ( available  for " attach disk"  option visibility  )

Do you mean how to import existing vm disk into ovirt?

You can use v2v (in ovirt-3.6) to import vms directly from vmware, or
ova created by vmware.

If you need to import qcow or raw disk, it is not automated yet, we
are working on it for 4.0.

Maybe Shahar can add more info what available today and what are the workarounds
to import disks in 3.6.

In 4.0 you will be able to upload disks directly using http,
see http://www.ovirt.org/develop/release-management/features/image-upload/

Nir


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fwd: Re: question mark on VM ( DB status 8 )

2016-03-18 Thread p...@email.cz


URGENT

 Forwarded Message 
Subject:Re: [ovirt-users] question mark on VM ( DB status 8 )
Date:   Thu, 17 Mar 2016 16:43:54 +0200
From:   Nir Soffer <nsof...@redhat.com>
To: p...@email.cz <p...@email.cz>



Can you send this to the users list?

This looks like virt issue, so it should be checked by the guys
working on this pars of the code.

Thanks,
Nir

On Thu, Mar 17, 2016 at 4:07 PM, p...@email.cz <p...@email.cz> wrote:

Hi Nir,
look at piece of logs which are repeated in cycle.

The main issue happened about 3-5AM today ( 17.Mar)

CSA_EBSDB_TEST2 - was shutted down from OS , but  status was not updated in
oVirt GUI ( changed manually in DB ( status 1 )) ,  still one other VM is in
status "8"  due snapshot locked file ( sf-sh-s07)  .

engine.log
==

repeately hours by hours ... continually

2016-03-17 14:38:21,146 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler_Worker-20) START,
GlusterVolumesListVDSCommand(HostName = 2kvm1, HostId =
4c3a2622-14d5-43c8-8e15-99cb66104b5a), log id: 5a34e053
2016-03-17 14:38:21,830 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler_Worker-20) FINISH, GlusterVolumesListVDSCommand,
return:
{a5a8ccbc-edee-4e49-9e2a-4d2ee5767f76=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@240192c6,
18310aeb-639f-4b6d-9ef4-9ef560d6175c=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@753f6685,
4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@79a21b20,
f410c6a9-9a51-42b3-89bb-c20ac72a0461=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@a4634e44,
62c89345-fd61-4b67-b8b4-69296eb7d217=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@fd990620,
aa2d607d-3c6c-4f13-8205-aae09dcc9d35=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@57883869,
b4356604-4404-428a-9da6-f1636115e2fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@3b458bc8,
9745551f-4696-4a6c-820a-619e359a61fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@80f225de,
25a5ec22-660e-42a0-aa00-45211d341738=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@ec4c19bd,
6060ff77-d552-4d94-97bf-5a32982e7d8a=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@947dc2e4,
cbf142f8-a40b-4cf4-ad29-2243c81d30c1=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f773ab98},
log id: 5a34e053
2016-03-17 14:38:27,131 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler_Worker-79) START,
GlusterVolumesListVDSCommand(HostName = 2kvm1, HostId =
4c3a2622-14d5-43c8-8e15-99cb66104b5a), log id: 24e7703f
2016-03-17 14:38:27,801 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler_Worker-79) FINISH, GlusterVolumesListVDSCommand,
return:
{a5a8ccbc-edee-4e49-9e2a-4d2ee5767f76=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72f0f4,
18310aeb-639f-4b6d-9ef4-9ef560d6175c=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@89bfd4dd,
4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f6cb25b,
f410c6a9-9a51-42b3-89bb-c20ac72a0461=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f4bb56bf,
62c89345-fd61-4b67-b8b4-69296eb7d217=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@e0121f88,
aa2d607d-3c6c-4f13-8205-aae09dcc9d35=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@435fc00f,
b4356604-4404-428a-9da6-f1636115e2fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7b23bf23,
9745551f-4696-4a6c-820a-619e359a61fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@1f8e886,
25a5ec22-660e-42a0-aa00-45211d341738=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@1fbbe1c1,
6060ff77-d552-4d94-97bf-5a32982e7d8a=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@87c991cd,
cbf142f8-a40b-4cf4-ad29-2243c81d30c1=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@2fc8ef3e},
log id: 24e7703f
2016-03-17 14:38:33,097 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler_Worker-15) START,
GlusterVolumesListVDSCommand(HostName = 2kvm1, HostId =
4c3a2622-14d5-43c8-8e15-99cb66104b5a), log id: 2e987652
2016-03-17 14:38:33,809 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler_Worker-15) FINISH, GlusterVolumesListVDSCommand,
return:
{a5a8ccbc-edee-4e49-9e2a-4d2ee5767f76=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@22f57

Re: [ovirt-users] [Gluster-users] open error -13 = sanlock

2016-03-03 Thread p...@email.cz

OK,
will extend replica 2 to replica 3 ( arbiter )  ASAP .

If is deleted "untouching" ids file on brick , healing of this file 
doesn't work .


regs.Pa.

On 3.3.2016 12:19, Nir Soffer wrote:
On Thu, Mar 3, 2016 at 11:23 AM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


This is replica 2, only , with following settings


Replica 2 is not supported. Even if you "fix" this now, you will have 
the same issue

soon.


Options Reconfigured:
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: fixed
cluster.server-quorum-type: none
storage.owner-uid: 36
storage.owner-gid: 36
cluster.quorum-count: 1
cluster.self-heal-daemon: enable

If I'll create "ids" file manually (  eg. " sanlock direct init -s

3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0
" ) on both bricks,
vdsm is writing only to half of them ( that with 2 links = correct )
"ids" file has correct permittions, owner, size  on both bricks.
brick 1:  -rw-rw 1 vdsm kvm 1048576  2. bře 18.56
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
- not updated
brick 2:  -rw-rw 2 vdsm kvm 1048576  3. bře 10.16
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
- is continually updated

What happens when I'll restart vdsm ? Will oVirt storages go to
"disable " state ??? = disconnect VMs storages ?


 Nothing will happen, the vms will continue to run normally.

On block storage, stopping vdsm will prevent automatic extending of vm 
disks
when the disk become too full, but on file based storage (like 
gluster) there is no issue.



regs.Pa.


On 3.3.2016 02:02, Ravishankar N wrote:

On 03/03/2016 12:43 AM, Nir Soffer wrote:


PS:  # find /STORAGES -samefile
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
-print
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
= missing "shadowfile" in " .gluster " dir.
How can I fix it ?? - online !


Ravi?

Is this the case in all 3 bricks of the replica?
BTW, you can just stat the file on the brick and see the link
count (it must be 2) instead of running the more expensive find
command.






___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] open error -13 = sanlock

2016-03-03 Thread p...@email.cz

This is replica 2, only , with following settings

Options Reconfigured:
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: fixed
cluster.server-quorum-type: none
storage.owner-uid: 36
storage.owner-gid: 36
cluster.quorum-count: 1
cluster.self-heal-daemon: enable

If I'll create "ids" file manually (  eg. " sanlock direct init -s 
3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0 
" ) on both bricks,

vdsm is writing only to half of them ( that with 2 links = correct )
"ids" file has correct permittions, owner, size  on both bricks.
brick 1:  -rw-rw 1 vdsm kvm 1048576  2. bře 18.56 
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids - 
not updated
brick 2:  -rw-rw 2 vdsm kvm 1048576  3. bře 10.16 
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids - 
is continually updated


What happens when I'll restart vdsm ? Will oVirt storages go to "disable 
" state ??? = disconnect VMs storages ?


regs.Pa.

On 3.3.2016 02:02, Ravishankar N wrote:

On 03/03/2016 12:43 AM, Nir Soffer wrote:


PS:  # find /STORAGES -samefile
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
-print
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
= missing "shadowfile" in " .gluster " dir.
How can I fix it ?? - online !


Ravi?

Is this the case in all 3 bricks of the replica?
BTW, you can just stat the file on the brick and see the link count 
(it must be 2) instead of running the more expensive find command.




___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] open error -13 = sanlock

2016-03-02 Thread p...@email.cz

UPDATE:

all "ids"  file have permittion fixed to 660 now

#  find /STORAGES -name ids -exec ls -l {} \;
-rw-rw 2 vdsm kvm 0 24. úno 07.41 
/STORAGES/g1r5p1/GFS/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md/ids
-rw-rw 2 vdsm kvm 0 24. úno 07.43 
/STORAGES/g1r5p2/GFS/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids
-rw-rw 2 vdsm kvm 0 24. úno 07.43 
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
-rw-rw 2 vdsm kvm 0 24. úno 07.44 
/STORAGES/g1r5p4/GFS/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids
-rw-rw 2 vdsm kvm 1048576 24. úno 13.03 
/STORAGES/g1r5p5/GFS/3b24d023-fd35-4666-af2f-f5e1d19531ad/dom_md/ids
-rw-rw 2 vdsm kvm 1048576  2. bře 17.47 
/STORAGES/g2r5p1/GFS/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids


SPM is and  was  running continually ...

I tried to update "ids" file - ONLINE  ( offline not possible yet )
# sanlock direct init -s 
3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0


#  find /STORAGES -name ids -exec ls -l {} \; | grep p3
-rw-rw 1 vdsm kvm 1048576  2. bře 18.32 
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids


The storage ids file has correct permittions, size, owners , but is not 
checking by sanlock = the same access time

What's wrong ??

regs.
Pa.
PS:  # find /STORAGES -samefile 
/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -print

/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids
= missing "shadowfile" in " .gluster " dir.
How can I fix it ?? - online !



On 2.3.2016 08:16, Ravishankar N wrote:

On 03/02/2016 12:02 PM, Sahina Bose wrote:



On 03/02/2016 03:45 AM, Nir Soffer wrote:

On Tue, Mar 1, 2016 at 10:51 PM, p...@email.cz <p...@email.cz> wrote:
>
> HI,
> requested output:
>
> # ls -lh /rhev/data-center/mnt/glusterSD/localhost:*/*/dom_md
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md:

> total 2,1M
> -rw-rw 1 vdsm kvm 1,0M  1. bře 21.28 ids  <-- good
> -rw-rw 1 vdsm kvm  16M  7. lis 22.16 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 22.17 leases
> -rw-r--r-- 1 vdsm kvm  335  7. lis 22.17 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 22.16 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.41 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M  7. lis 00.14 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 03.56 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 03.56 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.14 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M  7. lis 00.15 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 22.14 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 22.14 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.15 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P3/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M 23. úno 22.51 inbox
> -rw-rw 1 vdsm kvm 2,0M 23. úno 23.12 leases
> -rw-r--r-- 1 vdsm kvm  998 25. úno 00.35 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.16 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.44 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M  7. lis 00.17 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 00.18 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 00.18 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.17 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P1/42d710a9-b844-43dc-be41-77002d1cd553/dom_md:

> total 1,1M
> -rw-rw-r-- 1 vdsm kvm0 24. úno 07.32 ids  <-- bad (other can read)
> -rw-rw 1 vdsm kvm  16M  7. lis 22.18 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 22.18 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 22.18 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 22.18 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md:

> total 3,0M
> -rw-rw-r-- 1 vdsm kvm 1,0M  1. bře 21.28 ids  <-- bad (other can read)
> -rw-rw 1 vdsm kvm  16M 25. úno 00.42 inbox
> -rw-rw 1 vdsm kvm 2,0M 25. úno 00.44 leases
> -rw-r--r-- 1 vdsm kvm  997 24. úno 02.46 metadata
> -rw-rw 1 vdsm kvm  16M 25. úno 00.44 outbox
>
> 
/rhev/data-center/mnt/glusterS

[ovirt-users] SUCCESS / error

2016-03-02 Thread p...@email.cz

Hi, next explanation
VDSM log give me following message
Is this live checking for storage availability ? If "SUCCESS"  then why 
" " ???


regs.
Pavel

Thread-233::DEBUG::2016-03-02 
17:31:55,275::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: 
 = '0+1 records in\n0+1 records out\n346 bytes (346 B) copied, 
0.00022692 s, 1.5 MB/s\n';  = 0
Thread-299::DEBUG::2016-03-02 
17:31:57,870::fileSD::262::Storage.Misc.excCmd::(getReadDelay) 
/usr/bin/dd 
if=/rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97/dom_md/metadata 
iflag=direct of=/dev/null bs=4096 count=1 (cwd None)
Thread-299::DEBUG::2016-03-02 
17:31:57,886::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: 
 = '0+1 records in\n0+1 records out\n734 bytes (734 B) copied, 
0.000391403 s, 1.9 MB/s\n';  = 0
Thread-209::DEBUG::2016-03-02 
17:31:58,301::fileSD::262::Storage.Misc.excCmd::(getReadDelay) 
/usr/bin/dd 
if=/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/metadata 
iflag=direct of=/dev/null bs=4096 count=1 (cwd None)
Thread-209::DEBUG::2016-03-02 
17:31:58,318::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: 
 = '0+1 records in\n0+1 records out\n335 bytes (335 B) copied, 
0.00048095 s, 697 kB/s\n';  = 0
Thread-198::DEBUG::2016-03-02 
17:31:58,885::fileSD::262::Storage.Misc.excCmd::(getReadDelay) 
/usr/bin/dd 
if=/rhev/data-center/mnt/installshare:_DATA3/7f334c9d-ac01-4b2a-9f29-af1d2f24c3e0/dom_md/metadata 
iflag=direct of=/dev/null bs=4096 count=1 (cwd None)


 dist = RHEL - 7 - 2.1511
kernel = 3.10.0 - 327.10.1.el7.x86_64
KVM = 2.3.0 - 29.1.el7
libvirt = libvirt-1.2.17-13.el7_2.3
vdsm = vdsm-4.16.30-0.el7
GlusterFS = glusterfs-3.7.8-1.el7
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] open error -13 = sanlock

2016-03-02 Thread p...@email.cz

Yes we have had "ids" split brains + some other VM's files
Split brains was fixed by healing with preffered ( source ) brick.

eg: " # gluster volume heal 1KVM12-P1 split-brain source-brick 
16.0.0.161:/STORAGES/g1r5p1/GFS "


Pavel


Okay, so what I understand from the output above is you have different 
gluster volumes mounted and some of them have incorrect permissions 
for the 'ids' file. The way to fix it is to do it from the mount like 
Nir said.
Why did you delete the file from the .glusterfs in the brick(s)? Was 
there a gfid split brain?


-Ravi



___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] [Gluster-users] open error -13 = sanlock

2016-03-02 Thread p...@email.cz

Hi guys,
thx a lot for your support ...at first.

Because we had been under huge time pressure, we found "google 
workaround"  which delete both files . It helped, probabbly at first 
steps of recover .
eg: " #  find /STORAGES/g1r5p5/GFS/ -samefile 
/STORAGES/g1r5p5/GFS/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids 
-print -delete "


-->
Well at first I'll  fix permittions from mount points  to 660 .
If "ids"  file will be writeable , can't  became gluster colaps ??

regs.Pavel


On 2.3.2016 08:16, Ravishankar N wrote:

On 03/02/2016 12:02 PM, Sahina Bose wrote:



On 03/02/2016 03:45 AM, Nir Soffer wrote:

On Tue, Mar 1, 2016 at 10:51 PM, p...@email.cz <p...@email.cz> wrote:
>
> HI,
> requested output:
>
> # ls -lh /rhev/data-center/mnt/glusterSD/localhost:*/*/dom_md
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md:

> total 2,1M
> -rw-rw 1 vdsm kvm 1,0M  1. bře 21.28 ids  <-- good
> -rw-rw 1 vdsm kvm  16M  7. lis 22.16 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 22.17 leases
> -rw-r--r-- 1 vdsm kvm  335  7. lis 22.17 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 22.16 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.41 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M  7. lis 00.14 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 03.56 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 03.56 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.14 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M  7. lis 00.15 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 22.14 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 22.14 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.15 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P3/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M 23. úno 22.51 inbox
> -rw-rw 1 vdsm kvm 2,0M 23. úno 23.12 leases
> -rw-r--r-- 1 vdsm kvm  998 25. úno 00.35 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.16 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md:

> total 1,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.44 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M  7. lis 00.17 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 00.18 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 00.18 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 00.17 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P1/42d710a9-b844-43dc-be41-77002d1cd553/dom_md:

> total 1,1M
> -rw-rw-r-- 1 vdsm kvm0 24. úno 07.32 ids  <-- bad (other can read)
> -rw-rw 1 vdsm kvm  16M  7. lis 22.18 inbox
> -rw-rw 1 vdsm kvm 2,0M  7. lis 22.18 leases
> -rw-r--r-- 1 vdsm kvm  333  7. lis 22.18 metadata
> -rw-rw 1 vdsm kvm  16M  7. lis 22.18 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md:

> total 3,0M
> -rw-rw-r-- 1 vdsm kvm 1,0M  1. bře 21.28 ids  <-- bad (other can read)
> -rw-rw 1 vdsm kvm  16M 25. úno 00.42 inbox
> -rw-rw 1 vdsm kvm 2,0M 25. úno 00.44 leases
> -rw-r--r-- 1 vdsm kvm  997 24. úno 02.46 metadata
> -rw-rw 1 vdsm kvm  16M 25. úno 00.44 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P3/ef010d08-aed1-41c4-ba9a-e6d9bdecb4b4/dom_md:

> total 2,1M
> -rw-r--r-- 1 vdsm kvm0 24. úno 07.34 ids  <-- bad (sanlock 
cannot write, other can read)

> -rw-rw 1 vdsm kvm  16M 23. úno 22.35 inbox
> -rw-rw 1 vdsm kvm 2,0M 23. úno 22.38 leases
> -rw-r--r-- 1 vdsm kvm 1,1K 24. úno 19.07 metadata
> -rw-rw 1 vdsm kvm  16M 23. úno 22.27 outbox
>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97/dom_md:

> total 3,0M
> -rw-rw-r-- 1 vdsm kvm 1,0M  1. bře 21.28 ids  <-- bad (other can read)
> -rw-rw-r-- 1 vdsm kvm  16M  6. lis 23.50 inbox  <-- bad (other can 
read)
> -rw-rw-r-- 1 vdsm kvm 2,0M  6. lis 23.51 leases  <-- bad 
(other can read)
> -rw-rw-r-- 1 vdsm kvm  734  7. lis 02.13 metadata<-- bad 
(group can write, other can read)
> -rw-rw-r-- 1 vdsm kvm  16M  6. lis 16.55 outbox  <-- bad (other 
can read)

>
> 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P5/1ca56b45-701e-4c22-9f59-3aebea4d8477/dom_md:

> total 1,1

[ovirt-users] open error -13 = sanlock

2016-03-01 Thread p...@email.cz

Hello,  can anybody  explain this error no.13 ( open file ) in sanlock.log .

The size of  "ids" file is zero (0)

2016-02-28 03:25:46+0100 269626 [1951]: open error -13 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids
2016-02-28 03:25:46+0100 269626 [1951]: s187985 open_disk 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids 
error -13
2016-02-28 03:25:56+0100 269636 [11304]: s187992 lockspace 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0


If the main problem is about zero file size, can I regenerate  this file 
online securely , with no VM dependence  



dist = RHEL - 7 - 2.1511
kernel = 3.10.0 - 327.10.1.el7.x86_64
KVM = 2.3.0 - 29.1.el7
libvirt = libvirt-1.2.17-13.el7_2.3
vdsm = vdsm-4.16.30-0.el7
GlusterFS = glusterfs-3.7.8-1.el7


regs.
Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Fwd: Re: ovirt - can't attach master domain II

2016-02-24 Thread p...@email.cz

used replica2 with volume option

Volume Name: 2KVM12-P2
Type: Replicate
Volume ID: 9745551f-4696-4a6c-820a-619e359a61fd
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 16.0.0.164:/STORAGES/g1r5p2/GFS
Brick2: 16.0.0.163:/STORAGES/g1r5p2/GFS
Options Reconfigured:
storage.owner-uid: 36
storage.owner-gid: 36
performance.io-cache: off
performance.read-ahead: off
network.remote-dio: enable
cluster.eager-lock: enable
performance.stat-prefetch: off
performance.quick-read: off
cluster.quorum-count: 1
cluster.server-quorum-type: none
cluster.quorum-type: fixed

was runnig over year with no problems ( reboots, ..etc... )


On 24.2.2016 12:34, Ravishankar N wrote:

On 02/24/2016 04:48 PM, p...@email.cz wrote:



prereq: 2KVM12-P2 = master domain
-
YES - I'm using gluster.fuse NFS
localhost:/2KVM12-P2 on 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2 type 
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)

-
Healing
==
# gluster volume heal 2KVM12-P2 info
Brick 16.0.0.164:/STORAGES/g1r5p2/GFS
Number of entries: 0

Brick 16.0.0.163:/STORAGES/g1r5p2/GFS
Number of entries: 0

# while true; do for vol in `gluster volume list`; do gluster volume 
heal $vol info | sort | grep "Number of entries" | awk -F: '{tot+=$2} 
END { printf("Heal entries for '"$vol"': %d\n", $tot);}'; done; sleep 
120; echo -e "\n==\n"; done

Heal entries for 1KVM12-BCK: 1
Heal entries for 1KVM12-P1: 1
Heal entries for 1KVM12-P2: 0
Heal entries for 1KVM12-P3: 0
Heal entries for 1KVM12-P4: 0
Heal entries for 1KVM12-P5: 0
Heal entries for 2KVM12-P1: 1
Heal entries for 2KVM12-P2: 0
Heal entries for 2KVM12-P3: 0
Heal entries for 2KVM12-P5: 0
Heal entries for 2KVM12_P4: 1

# gluster volume heal 1KVM12-BCK info split-brain
Brick 16.0.0.161:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

# gluster volume heal 1KVM12-P1 info split-brain
Brick 16.0.0.161:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

etc..


YES - in split brain , but NOT master domain ( will solve later, 
after master - if possible  )


I'm not sure if it is related, but you could try to resolve the 
split-brain first and see if it helps. Also, I see that you are using 
replica-2. It is recommended to use replica-3 or arbiter volumes to 
avoid split-brains.


-Ravi



---
vdsm.log
=

Thread-461::DEBUG::2016-02-24 
11:12:45,328::fileSD::262::Storage.Misc.excCmd::(getReadDelay) 
SUCCESS:  = '0+1 records in\n0+1 records out\n333 bytes (333 B) 
copied, 0.000724379 s, 460 kB/s\n';  = 0
Thread-461::INFO::2016-02-24 
11:12:45,331::clusterlock::219::Storage.SANLock::(acquireHostId) 
Acquiring host id for domain 88adbd49-62d6-45b1-9992-b04464a04112 (id: 3)
Thread-461::DEBUG::2016-02-24 
11:12:45,331::clusterlock::237::Storage.SANLock::(acquireHostId) Host 
id for domain 88adbd49-62d6-45b1-9992-b04464a04112 successfully 
acquired (id: 3)
Thread-33186::DEBUG::2016-02-24 
11:12:46,067::__init__::481::jsonrpc.JsonRpcServer::(_serveRequest) 
Calling 'GlusterVolume.list' in bridge with {}
Thread-33186::DEBUG::2016-02-24 
11:12:46,204::__init__::514::jsonrpc.JsonRpcServer::(_serveRequest) 
Return 'GlusterVolume.list' in bridge with {'volumes': {'2KVM12-P5': 
{'transportType': ['TCP'], 'uuid': 
'4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d', 'bricks': 
['16.0.0.164:/STORAGES/g1r5p5/GFS', 
'16.0.0.163:/STORAGES/g1r5p5/GFS'], 'volumeName': '2KVM12-P5', 
'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 
'distCount': '2', 'volumeStatus': 'ONLINE', 'stripeCount': '1', 
'bricksInfo': [{'name': '16.0.0.164:/STORAGES/g1r5p5/GFS', 
'hostUuid': '06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}, {'name': 
'16.0.0.163:/STORAGES/g1r5p5/GFS', 'hostUuid': 
'6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}], 'options': 
{'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 
'enable', 'performance.stat-prefetch': 'off', 'cluster.quorum-type': 
'fixed', 'performance.quick-read': 'off', 'network.remote-dio': 
'enable', 'cluster.quorum-count': '1', 'performance.io-cache': 'off', 
'storage.owner-uid': '36', 'performance.read-ahead': 'off', 
'storage.owner-gid': '36'}}, '2KVM12_P4': {'transportType': ['TCP'], 
'uuid': '18310aeb-639f-4b6d-9ef4-9ef560d6175c', 'bricks': 
['16.0.0.163:/STORAGES/g1r5p4/GFS', 
'16.0.0.164:/STORAGES/g1r5p4/GFS'], 'volumeName': '2KVM12_P4', 
'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 
'distCount': '2', 'vo

[ovirt-users] Fwd: Re: ovirt - can't attach master domain III

2016-02-24 Thread p...@email.cz


hi,
after a lot of test will get

2016-02-24 11:38:05+0100 7406 [25824]: cmd_add_lockspace 3,10 
ff71b47b-0f72-4528-9bfe-c3da888e47f0:4:/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 11:38:05+0100 7406 [25824]: s2256 lockspace 
ff71b47b-0f72-4528-9bfe-c3da888e47f0:4:/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids:0

2016-02-24 11:38:05+0100 7406 [25824]: cmd_add_lockspace 3,10 async done 0
2016-02-24 11:38:05+0100 7406 [26186]: open error -2 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids
2016-02-24 11:38:05+0100 7406 [26186]: s2256 open_disk 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids 
error -2


what's wrong ??
thx.
Pa.

On 24.2.2016 08:14, Nir Soffer wrote:
On Wed, Feb 24, 2016 at 8:53 AM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


Hi,
it seems that sanlock daemon has problem with reading empty "ids"
file .
How can I regenarate this "ids" file to get 2k rows of datas ??
It's the base problem to get up "master domain" following
"datacenter"


You should understand why the ids files is empty and fix the root cause.

To recover your ids files, you can follow the instructions here:
http://lists.ovirt.org/pipermail/users/2016-February/038046.html

Nir


regs.
Pa.

___
Users mailing list
Users@ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users






___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Fwd: Re: ovirt - can't attach master domain II

2016-02-24 Thread p...@email.cz
gt; 
state preparing
Thread-35634::INFO::2016-02-24 
11:18:20,860::logUtils::44::dispatcher::(wrapper) Run and protect: 
getSpmStatus(spUUID=u'0002-0002-0002-0002-021e', options=None)
Thread-35634::INFO::2016-02-24 
11:18:20,867::logUtils::47::dispatcher::(wrapper) Run and protect: 
getSpmStatus, Return response: {'spm_st': {'spmId': -1, 'spmStatus': 
'Free', 'spmLver': -1}}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::task::1191::Storage.TaskManager.Task::(prepare) 
Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::finished: {'spm_st': 
{'spmId': -1, 'spmStatus': 'Free', 'spmLver': -1}}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::task::595::Storage.TaskManager.Task::(_updateState) 
Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::moving from state preparing 
-> state finished
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) 
Owner.releaseAll requests {} resources {}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) 
Owner.cancelAll requests {}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::task::993::Storage.TaskManager.Task::(_decref) 
Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::ref 0 aborting False

---

these blocks generated in cycle for each domain

Any IDEA ??
regs.
Pa.



On 24.2.2016 10:54, Ravishankar N wrote:

Hi,

On 02/24/2016 06:43 AM, p...@email.cz wrote:

Hi,
I found the main ( maybe ) problem with IO error ( -5 ) for "ids" 
file access

This file is not accessable via NFS, locally yes

How is NFS coming into the picture? Are you not using gluster fuse mount?

.
How can I fix it ??
Can you run `gluster volume heal volname info` and `gluster volume 
heal volname info split-brain` to see if the "ids" file is in 
split-brain? A file in split-brain returns EIO when accessed from the 
mount.

Regards,
Ravi



regs.
Pavel

# sanlock client log_dump

0 flags 1 timeout 0
2016-02-24 02:01:10+0100 3828 [12111]: s1316 lockspace 
88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0
2016-02-24 02:01:10+0100 3828 [12111]: cmd_add_lockspace 4,15 async 
done 0
2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire begin 
88adbd49-62d6-45b1-9992-b04464a04112:1
2016-02-24 02:01:10+0100 3828 [19556]: 88adbd49 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:10+0100 3828 [19556]: read_sectors delta_leader 
offset 0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids
2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire 
leader_read1 error -5

2016-02-24 02:01:11+0100 3829 [12111]: s1316 add_lockspace fail result -5
2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:12+0100 3831 [12116]: s1317 lockspace 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0
2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 async 
done 0
2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire begin 
7f52b697-c199-4f58-89aa-102d44327124:1
2016-02-24 02:01:12+0100 3831 [19562]: 7f52b697 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:12+0100 3831 [19562]: read_sectors delta_leader 
offset 0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids
2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire 
leader_read1 error -5
2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:13+0100 3831 [1321]: s1318 lockspace 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0

2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire begin 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1
2016-02-24 02:01:13+0100 3831 [19564]: 0fcad888 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458201000 result -5:0 match res
2016-02-24 02:01:13+0100 3831 [19564]: read_sectors delta_leader 
offset 0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire 
leader_read1 error -5

2016-02-24 02:01:13+0100 3832 [12116]: s1317 add_lockspace fail result -5
2016-02-24 02:01:14+0100 3832 [1321]: s1318 add_lockspace fail result -5

[ovirt-users] Fwd: Re: ovirt - can't attach master domain III

2016-02-24 Thread p...@email.cz


Hi Nir,

it  doesn't running, or failed by me

1) no trafic od any storage in gluster

two ways
A:  - will stop ( maintenance not allowed )master domain ( 2KVM12-P2 ) 
from  GUI

- try to mount locally to one node
# mount -t glusterfs localhost:/2KVM12-P2 /mnt ==> error -19

B: go to 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/f71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/

- remove empty ids file
- sanlock direct init -s :0:ids:0 - from manual
- restart sanlockd

2) no way was successfull

regs.Pa.



On 24.2.2016 08:14, Nir Soffer wrote:
On Wed, Feb 24, 2016 at 8:53 AM, p...@email.cz <mailto:p...@email.cz> 
<p...@email.cz <mailto:p...@email.cz>> wrote:


Hi,
it seems that sanlock daemon has problem with reading empty "ids"
file .
How can I regenarate this "ids" file to get 2k rows of datas ??
It's the base problem to get up "master domain" following
"datacenter"


You should understand why the ids files is empty and fix the root cause.

To recover your ids files, you can follow the instructions here:
http://lists.ovirt.org/pipermail/users/2016-February/038046.html

Nir


regs.
Pa.

___
Users mailing list
Users@ovirt.org <mailto:Users@ovirt.org>
http://lists.ovirt.org/mailman/listinfo/users






___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] ovirt - can't attach master domain II

2016-02-24 Thread p...@email.cz
reparing
Thread-35634::INFO::2016-02-24 
11:18:20,860::logUtils::44::dispatcher::(wrapper) Run and protect: 
getSpmStatus(spUUID=u'0002-0002-0002-0002-021e', options=None)
Thread-35634::INFO::2016-02-24 
11:18:20,867::logUtils::47::dispatcher::(wrapper) Run and protect: 
getSpmStatus, Return response: {'spm_st': {'spmId': -1, 'spmStatus': 
'Free', 'spmLver': -1}}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::task::1191::Storage.TaskManager.Task::(prepare) 
Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::finished: {'spm_st': 
{'spmId': -1, 'spmStatus': 'Free', 'spmLver': -1}}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::task::595::Storage.TaskManager.Task::(_updateState) 
Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::moving from state preparing 
-> state finished
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) 
Owner.releaseAll requests {} resources {}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) 
Owner.cancelAll requests {}
Thread-35634::DEBUG::2016-02-24 
11:18:20,867::task::993::Storage.TaskManager.Task::(_decref) 
Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::ref 0 aborting False

---

these blocks generated in cycle for each domain

Any IDEA ??
regs.
Pa.



On 24.2.2016 10:54, Ravishankar N wrote:

Hi,

On 02/24/2016 06:43 AM, p...@email.cz wrote:

Hi,
I found the main ( maybe ) problem with IO error ( -5 ) for "ids" 
file access

This file is not accessable via NFS, locally yes

How is NFS coming into the picture? Are you not using gluster fuse mount?

.
How can I fix it ??
Can you run `gluster volume heal volname info` and `gluster volume 
heal volname info split-brain` to see if the "ids" file is in 
split-brain? A file in split-brain returns EIO when accessed from the 
mount.

Regards,
Ravi



regs.
Pavel

# sanlock client log_dump

0 flags 1 timeout 0
2016-02-24 02:01:10+0100 3828 [12111]: s1316 lockspace 
88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0
2016-02-24 02:01:10+0100 3828 [12111]: cmd_add_lockspace 4,15 async 
done 0
2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire begin 
88adbd49-62d6-45b1-9992-b04464a04112:1
2016-02-24 02:01:10+0100 3828 [19556]: 88adbd49 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:10+0100 3828 [19556]: read_sectors delta_leader 
offset 0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids
2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire 
leader_read1 error -5

2016-02-24 02:01:11+0100 3829 [12111]: s1316 add_lockspace fail result -5
2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:12+0100 3831 [12116]: s1317 lockspace 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0
2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 async 
done 0
2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire begin 
7f52b697-c199-4f58-89aa-102d44327124:1
2016-02-24 02:01:12+0100 3831 [19562]: 7f52b697 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:12+0100 3831 [19562]: read_sectors delta_leader 
offset 0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids
2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire 
leader_read1 error -5
2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:13+0100 3831 [1321]: s1318 lockspace 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0

2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire begin 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1
2016-02-24 02:01:13+0100 3831 [19564]: 0fcad888 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458201000 result -5:0 match res
2016-02-24 02:01:13+0100 3831 [19564]: read_sectors delta_leader 
offset 0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire 
leader_read1 error -5

2016-02-24 02:01:13+0100 3832 [12116]: s1317 add_lockspace fail result -5
2016-02-24 02:01:14+0100 3832 [1321]: s1318 add_lockspace fail result -5
2016-02-24 0

[ovirt-users] ovirt - can't attach master domain III

2016-02-23 Thread p...@email.cz

Hi,
it seems that sanlock daemon has problem with reading empty "ids" file .
How can I regenarate this "ids" file to get 2k rows of datas ??
It's the base problem to get up "master domain"  following "datacenter"

regs.
Pa.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] ovirt - can't attach master domain II

2016-02-23 Thread p...@email.cz

Hi,
I found the main ( maybe ) problem with IO error ( -5 ) for "ids" file 
access

This file is not accessable via NFS, locally yes.
How can I fix it ??
regs.
Pavel

# sanlock client log_dump

0 flags 1 timeout 0
2016-02-24 02:01:10+0100 3828 [12111]: s1316 lockspace 
88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0

2016-02-24 02:01:10+0100 3828 [12111]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire begin 
88adbd49-62d6-45b1-9992-b04464a04112:1
2016-02-24 02:01:10+0100 3828 [19556]: 88adbd49 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:10+0100 3828 [19556]: read_sectors delta_leader offset 
0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids
2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire leader_read1 
error -5

2016-02-24 02:01:11+0100 3829 [12111]: s1316 add_lockspace fail result -5
2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:12+0100 3831 [12116]: s1317 lockspace 
7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0

2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire begin 
7f52b697-c199-4f58-89aa-102d44327124:1
2016-02-24 02:01:12+0100 3831 [19562]: 7f52b697 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:12+0100 3831 [19562]: read_sectors delta_leader offset 
0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids
2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire leader_read1 
error -5
2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:13+0100 3831 [1321]: s1318 lockspace 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0

2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire begin 
0fcad888-d573-47be-bef3-0bc0b7a99fb7:1
2016-02-24 02:01:13+0100 3831 [19564]: 0fcad888 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458201000 result -5:0 match res
2016-02-24 02:01:13+0100 3831 [19564]: read_sectors delta_leader offset 
0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire leader_read1 
error -5

2016-02-24 02:01:13+0100 3832 [12116]: s1317 add_lockspace fail result -5
2016-02-24 02:01:14+0100 3832 [1321]: s1318 add_lockspace fail result -5
2016-02-24 02:01:19+0100 3838 [12106]: cmd_add_lockspace 4,15 
3da46e07-d1ea-4f10-9250-6cbbb7b94d80:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P5/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:19+0100 3838 [12106]: s1319 lockspace 
3da46e07-d1ea-4f10-9250-6cbbb7b94d80:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P5/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids:0

2016-02-24 02:01:19+0100 3838 [12106]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:19+0100 3838 [19638]: s1319 delta_acquire begin 
3da46e07-d1ea-4f10-9250-6cbbb7b94d80:1
2016-02-24 02:01:19+0100 3838 [19638]: 3da46e07 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res
2016-02-24 02:01:19+0100 3838 [19638]: read_sectors delta_leader offset 
0 rv -5 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P5/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids
2016-02-24 02:01:19+0100 3838 [19638]: s1319 delta_acquire leader_read1 
error -5

2016-02-24 02:01:20+0100 3839 [12106]: s1319 add_lockspace fail result -5
2016-02-24 02:01:20+0100 3839 [1320]: cmd_add_lockspace 4,15 
88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0 
flags 1 timeout 0
2016-02-24 02:01:20+0100 3839 [1320]: s1320 lockspace 
88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0

2016-02-24 02:01:20+0100 3839 [1320]: cmd_add_lockspace 4,15 async done 0
2016-02-24 02:01:20+0100 3839 [19658]: s1320 delta_acquire begin 
88adbd49-62d6-45b1-9992-b04464a04112:1
2016-02-24 02:01:20+0100 3839 [19658]: 88adbd49 aio collect 0 
0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 

[ovirt-users] ovirt - can't attach master domain

2016-02-23 Thread p...@email.cz

Hi,
due cleaning envir. I restarted all nodes and ovirt mgm server too,
Now I can't  get master domain ON. Tried to move to diff. storage with 
same result.

All storages are OFF and DataCenter OFF too.
All NFS mount points are connected and writeable via ovirt mountpoint.
Any idea how to fix thi - quicky

Rhel 7.1 /  KVM 2.1.2 / libvirt 1.2.8-16 / vdsm 4.16.27 / gluster 
3.7.5.1 / ovirt 3.5.2


VDSM log

Thread-22688::DEBUG::2016-02-24 
01:32:34,852::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) 
Owner.releaseAll requests {} resources {}
Thread-22688::DEBUG::2016-02-24 
01:32:34,852::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) 
Owner.cancelAll requests {}
Thread-22688::DEBUG::2016-02-24 
01:32:34,852::task::993::Storage.TaskManager.Task::(_decref) 
Task=`492b6d3f-47b7-47fe-a713-6e47d5f2d3c1`::ref 0 aborting False
Thread-22688::DEBUG::2016-02-24 
01:32:34,855::stompReactor::162::yajsonrpc.StompServer::(send) Sending 
response
Thread-22697::DEBUG::2016-02-24 
01:32:34,967::stompReactor::162::yajsonrpc.StompServer::(send) Sending 
response
Thread-200::DEBUG::2016-02-24 
01:32:35,001::fileSD::153::Storage.StorageDomain::(__init__) Reading 
domain in path 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97
Thread-200::ERROR::2016-02-24 
01:32:35,007::domainMonitor::256::Storage.DomainMonitorThread::(_monitorDomain) 
Error while collecting domain 300e9ac8-3c2f-4703-9bb1-1df2130c7c97 
monitoring information

Traceback (most recent call last):
  File "/usr/share/vdsm/storage/domainMonitor.py", line 221, in 
_monitorDomain

self.domain = sdCache.produce(self.sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
domain.getRealDomain()
  File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
return self._cache._realProduce(self._sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
domain = self._findDomain(sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
dom = findMethod(sdUUID)
  File "/usr/share/vdsm/storage/glusterSD.py", line 32, in findDomain
return 
GlusterStorageDomain(GlusterStorageDomain.findDomainPath(sdUUID))

  File "/usr/share/vdsm/storage/fileSD.py", line 160, in __init__
validateFileSystemFeatures(sdUUID, self.mountpoint)
  File "/usr/share/vdsm/storage/fileSD.py", line 89, in 
validateFileSystemFeatures

oop.getProcessPool(sdUUID).directTouch(testFilePath)
  File "/usr/share/vdsm/storage/outOfProcess.py", line 351, in directTouch
ioproc.touch(path, flags, mode)
  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 
507, in touch

self.timeout)
  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 
391, in _sendCommand

raise OSError(errcode, errstr)
OSError: [Errno 5] Input/output error
Thread-130::DEBUG::2016-02-24 
01:32:35,013::fileSD::153::Storage.StorageDomain::(__init__) Reading 
domain in path 
/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4
Thread-130::ERROR::2016-02-24 
01:32:35,017::domainMonitor::256::Storage.DomainMonitorThread::(_monitorDomain) 
Error while collecting domain 553d9b92-e4a0-4042-a579-4cabeb55ded4 
monitoring information

Traceback (most recent call last):
  File "/usr/share/vdsm/storage/domainMonitor.py", line 221, in 
_monitorDomain

self.domain = sdCache.produce(self.sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
domain.getRealDomain()
  File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
return self._cache._realProduce(self._sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
domain = self._findDomain(sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
dom = findMethod(sdUUID)
  File "/usr/share/vdsm/storage/glusterSD.py", line 32, in findDomain
return 
GlusterStorageDomain(GlusterStorageDomain.findDomainPath(sdUUID))

  File "/usr/share/vdsm/storage/fileSD.py", line 160, in __init__
validateFileSystemFeatures(sdUUID, self.mountpoint)
  File "/usr/share/vdsm/storage/fileSD.py", line 89, in 
validateFileSystemFeatures

oop.getProcessPool(sdUUID).directTouch(testFilePath)
  File "/usr/share/vdsm/storage/outOfProcess.py", line 351, in directTouch
ioproc.touch(path, flags, mode)
  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 
507, in touch

self.timeout)
  File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 
391, in _sendCommand

raise OSError(errcode, errstr)
OSError: [Errno 5] Input/output error

thx Pavel

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] Can't fine engine postgres password

2016-02-19 Thread p...@email.cz

Hello,

it's running well on ovirt 3.6, but exists the same procedure for ovirt 3.5 ??
Didn't find that.  Or can I use this from ovirt 3.6 on 3.5 version ??

regs.
Pavel




===
Hi,

database configuration is stored in the VM/host where engine is installed
in "/etc/ovirt-engine/engine.conf.d/10-setup-database.conf". If you used
automatic database setup, then whole db configuration is generated
automatically. And please bear in mind that db user password is
different from webadmin user (admin@internal) password which you can
specify during installation.

Martin Perina

- Original Message -


From: "Adam Verslype"
To:users@ovirt.org
Sent: Monday, February 15, 2016 2:50:21 AM
Subject: [ovirt-users] Can't fine engine postgres password

Hi,

After deploying hosted engine using the hosted-engine-appliance I need to run
the unlock_entity utility on the engine vm to unlock a disk that's stuck. So
I ssh on the vm and run:

sudo PGPASSWORD=x ./unlock_entity.sh -t disk -u engine -q
psql: FATAL: password authentication failed for user "engine"
FATAL: Cannot execute sql command: --file=./unlock_entity.sql
where x is the password I supplied for everything during deployment.
Does it have a different password that came with the appliance? Is this
documented somewhere?

Thanks,

Adam Verslype
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] delete hang task

2016-02-18 Thread p...@email.cz

Hello,
I'm testing oVirt 3.6 for failover and have total issue.
Snapshot VM  will hang on ZFS filesystem.
But the main questionis is : how can I cancel any unfinished tasks in 
ovirt ???
I didn't find any "normal" solution, except deleting record from ovirt 
DB manually


any idea ? - no one is missing this functionality ???
regs.
Pa.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] question mark in oVirt - "cannot acquire state change lock"

2016-02-17 Thread p...@email.cz

Hello,
some requested outputs

Vdsm generate a huge count of records in time, so I'm not shure if 
"hang" time is still included .

Problems occured 17-feb-2016 17:00-18:00 when VM shutdown was requested.
But some performance problems occured a day before ( I'm not shure about 
application problem = JAVA = running EBS )


In mail before  I attached pointers to bugs with any workaround

killall -9 libvirtd
rm /var/run/libvirtd.pid
/etc/init.d/libvirt-bin restart

which would solve current situation.
Is this  workaround safe enought?? Will stay VMs running ???

THX for any help, because I can't reboot hypervizor due huge DB traffic 
and no VM migration possibility via oVirt engine. ( mem synchro 
overloaded ?? )


Pa.
PS: logs sent directly to RH.


On 17.2.2016 08:22, Nir Soffer wrote:

On Tue, Feb 16, 2016 at 9:58 PM, p...@email.cz <p...@email.cz> wrote:

Hello all,
how can I eliminate question mark from "hanged" VM? ( it happened in
shutdown process via oVirt  env. )
Best solution without hypervizor reboot, meaning "power off" , "kill" or
another workaround for that VM .
Will help removing task DB record from oVirt mgmt server - if exists ,
ofcourse ??
Running  gluster 3.7.5 / oVirt 3.5.2 /
libvirt-daemon-kvm-1.2.8-16.el7_1.5.x86_64  on centos 7.1.1503 .

thx a lot.
Pa.

PS: I found similar problem in libvirt library a many years ago ( redhat 5.6
)
https://bugzilla.redhat.com/show_bug.cgi?id=676205
https://bugzilla.redhat.com/show_bug.cgi?id=668438
problem explanation:
http://linux.india365.org/2012/01/libvirt-virsh-cannot-acquire-state-change-lock-error-explained/

This is very interesting, I have seen this error in testing
environment on rhel 7.1 or 7.2.

Can you provide vdsm logs showing the time of the error?

Nir


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] question mark in oVirt - "cannot acquire state change lock"

2016-02-16 Thread p...@email.cz

Hello all,
how can I eliminate question mark from "hanged" VM? ( it happened in shutdown 
process via oVirt  env. )
Best solution without hypervizor reboot, meaning "power off" , "kill" or 
another workaround for that VM .
Will help removing task DB record from oVirt mgmt server - if exists , ofcourse 
??
Running  gluster 3.7.5 / oVirt 3.5.2 / 
libvirt-daemon-kvm-1.2.8-16.el7_1.5.x86_64  on centos 7.1.1503 .

thx a lot.
Pa.

PS: I found similar problem in libvirt library a many years ago ( redhat 5.6 )
   https://bugzilla.redhat.com/show_bug.cgi?id=676205
   https://bugzilla.redhat.com/show_bug.cgi?id=668438
problem explanation:
http://linux.india365.org/2012/01/libvirt-virsh-cannot-acquire-state-change-lock-error-explained/

  So it means that this bug was not fixed to this time ??

regs.

___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] VM database update ( VMrestore)

2015-12-23 Thread p...@email.cz

Hello,

1) if I've got full VM's datastore backup ( eg. via rsync ) and restore 
one VM's  files to different store and empty oVirt database, will that 
database be updated automaticaly or any more actions needed ??
2) Are VM's file stored in oVirt database in names only ? ( no check 
sum, or  other  IDs ) - meaning can I replace VM file by another one 
with different size  ??

regs.
Paf1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


Re: [ovirt-users] timeouts

2015-11-27 Thread p...@email.cz

Hi,
all glusterd daemons was runnig correctly at this time, no 
firewalls/iptables restrictions

But  "not connected" bricks are changing during the time without any touch .
It looks that glusterd  has non-stable cross  communication , especially 
with different LAN range  as nodes in Ovirt environmet

( Volumes bricks in 16.0.0.0 net and ovirt nodes in 172.0.0.0 net )
So I desided reinstall whole cluster, but I'm afraid that these problems 
will occure again - will you know


regs.for your answers
Pavel

On 27.11.2015 10:16, knarra wrote:

On 11/27/2015 11:04 AM, knarra wrote:

Hi Paf1,

Looks like when you reboot the nodes, glusterd does not start up 
in one node and due to this the node gets disconnected from other 
node(that is what i see from logs). After reboot, once your systems 
are up and running , can you check if glusterd is running on all the 
nodes? Can you please let me know which build of gluster are you using ?


For more info please read, 
http://www.gluster.org/pipermail/gluster-users.old/2015-June/022377.html 
- (please ignore this line)




Thanks
kasturi

On 11/27/2015 10:52 AM, Sahina Bose wrote:

[+ gluster-users]

On 11/26/2015 08:37 PM, p...@email.cz wrote:

Hello,
can anybody  help me with this timeouts ??
Volumes are not active yes ( bricks down )

desc. of gluster bellow ...

*/var/log/glusterfs/**etc-glusterfs-glusterd.vol.log*
[2015-11-26 14:44:47.174221] I [MSGID: 106004] 
[glusterd-handler.c:5065:__glusterd_peer_rpc_notify] 0-management: 
Peer <1hp1-SAN> (<87fc7db8-aba8-41f2-a1cd-b77e83b17436>), in state 
, has disconnected from glusterd.
[2015-11-26 14:44:47.174354] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P1 not held
[2015-11-26 14:44:47.17] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P3 not held
[2015-11-26 14:44:47.174521] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P1 not held
[2015-11-26 14:44:47.174662] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P3 not held
[2015-11-26 14:44:47.174532] W [MSGID: 106118] 
[glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: 
Lock not released for 2HP12-P1
[2015-11-26 14:44:47.174675] W [MSGID: 106118] 
[glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: 
Lock not released for 2HP12-P3
[2015-11-26 14:44:49.423334] I [MSGID: 106488] 
[glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 
0-glusterd: Received get vol req
The message "I [MSGID: 106488] 
[glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 
0-glusterd: Received get vol req" repeated 4 times between 
[2015-11-26 14:44:49.423334] and [2015-11-26 14:44:49.429781]
[2015-11-26 14:44:51.148711] I [MSGID: 106163] 
[glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 
0-management: using the op-version 30702
[2015-11-26 14:44:52.177266] W [socket.c:869:__socket_keepalive] 
0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 12, 
Invalid argument
[2015-11-26 14:44:52.177291] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-26 14:44:53.180426] W [socket.c:869:__socket_keepalive] 
0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 17, 
Invalid argument
[2015-11-26 14:44:53.180447] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-26 14:44:52.395468] I [MSGID: 106163] 
[glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 
0-management: using the op-version 30702
[2015-11-26 14:44:54.851958] I [MSGID: 106488] 
[glusterd-ha

[ovirt-users] timeouts

2015-11-26 Thread p...@email.cz

Hello,
can anybody  help me with this timeouts ??
Volumes are not active yes ( bricks down )

desc. of gluster bellow ...

*/var/log/glusterfs/**etc-glusterfs-glusterd.vol.log*
[2015-11-26 14:44:47.174221] I [MSGID: 106004] 
[glusterd-handler.c:5065:__glusterd_peer_rpc_notify] 0-management: Peer 
<1hp1-SAN> (<87fc7db8-aba8-41f2-a1cd-b77e83b17436>), in state Cluster>, has disconnected from glusterd.
[2015-11-26 14:44:47.174354] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P1 not held
[2015-11-26 14:44:47.17] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P3 not held
[2015-11-26 14:44:47.174521] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P1 not held
[2015-11-26 14:44:47.174662] W 
[glusterd-locks.c:681:glusterd_mgmt_v3_unlock] 
(-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) 
[0x7fb7039d44dc] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) 
[0x7fb7039de542] 
-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) 
[0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P3 not held
[2015-11-26 14:44:47.174532] W [MSGID: 106118] 
[glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: Lock 
not released for 2HP12-P1
[2015-11-26 14:44:47.174675] W [MSGID: 106118] 
[glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: Lock 
not released for 2HP12-P3
[2015-11-26 14:44:49.423334] I [MSGID: 106488] 
[glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: 
Received get vol req
The message "I [MSGID: 106488] 
[glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: 
Received get vol req" repeated 4 times between [2015-11-26 
14:44:49.423334] and [2015-11-26 14:44:49.429781]
[2015-11-26 14:44:51.148711] I [MSGID: 106163] 
[glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 
0-management: using the op-version 30702
[2015-11-26 14:44:52.177266] W [socket.c:869:__socket_keepalive] 
0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 12, Invalid 
argument
[2015-11-26 14:44:52.177291] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-26 14:44:53.180426] W [socket.c:869:__socket_keepalive] 
0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 17, Invalid 
argument
[2015-11-26 14:44:53.180447] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-26 14:44:52.395468] I [MSGID: 106163] 
[glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 
0-management: using the op-version 30702
[2015-11-26 14:44:54.851958] I [MSGID: 106488] 
[glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: 
Received get vol req
[2015-11-26 14:44:57.183969] W [socket.c:869:__socket_keepalive] 
0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 19, Invalid 
argument
[2015-11-26 14:44:57.183990] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument


After volumes creation all works fine ( volumes up ) , but then, after 
several reboots ( yum updates) volumes failed due timeouts .


Gluster description:

4 nodes with 4 volumes replica 2
oVirt 3.6 - the last
gluster 3.7.6 - the last
vdsm 4.17.999 - from git repo
oVirt - mgmt.nodes 172.16.0.0
oVirt - bricks 16.0.0.0 ( "SAN" - defined as "gluster" net)
Network works fine, no lost packets

# gluster volume status
Staging failed on 2hp1-SAN. Please check log file for details.
Staging failed on 1hp2-SAN. Please check log file for details.
Staging failed on 2hp2-SAN. Please check log file for details.

# gluster volume info

Volume Name: 1HP12-P1
Type: Replicate
Volume ID: 6991e82c-9745-4203-9b0a-df202060f455
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 1hp1-SAN:/STORAGE/p1/G
Brick2: 1hp2-SAN:/STORAGE/p1/G
Options Reconfigured:
performance.readdir-ahead: on

Volume Name: 1HP12-P3

[ovirt-users] remove volume from DB

2015-11-18 Thread p...@email.cz

Hello,
howto remove volume definition from oVirt DB ( & from nodes gluster 
config ) if volume totaly cleaned in background in running mode ??


regs.
Paf1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] volume parameters

2015-11-07 Thread p...@email.cz

Hello,
would U recommend me set of functional params for volume replica 2 ?
Old ones was  ( for 3.5.2 gluster version )
storage.owner-uid   36
storage.owner-gid   36
performance.io-cache  off
performance.read-ahead  off
network.remote-dio enable
cluster.eager-lock enable
performance.stat-prefetch off
performance.quick-read off
cluster.quorum-count 1
cluster.server-quorum-type none
cluster.quorum-type  fixed

after upgrade to 3.5.7 version and setting default recommendation, 
volumes became inaccessable ( permission denied - fixed by owner uid/gui 
settings to 36)

Why the defaults have been changed  ?
Just still Error / Critical messages occure  ( examples follow )

*E* - list of grep  etc-glusterfs-glusterd.vol.log
[2015-11-07 10:49:10.883564] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-07 10:49:10.886152] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-07 10:49:15.954942] E [rpc-clnt.c:362:saved_frames_unwind] (--> 
/lib64/libglusterfs.so.0(_gf_log_callingfn+0x196)[0x7fa88b014a66] (--> 
/lib64/libgfrpc.so.0(saved_frames_unwind+0x1de)[0x7fa88addf9be] (--> 
/lib64/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fa88addface] (--> 
/lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x9c)[0x7fa88ade148c] 
(--> /lib64/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7fa88ade1c98] ) 
0-management: forced unwinding frame type(Peer mgmt) op(--(2)) called at 
2015-11-07 10:49:10.918764 (xid=0x5)
[2015-11-07 10:49:26.719176] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-07 10:54:59.738232] E [MSGID: 106243] [glusterd.c:1623:init] 
0-management: creation of 1 listeners failed, continuing with succeeded 
transport
[2015-11-07 10:55:01.860991] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-07 10:55:01.863932] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument
[2015-11-07 10:55:01.866779] E [socket.c:2965:socket_connect] 
0-management: Failed to set keep-alive: Invalid argument


*C* - list of grep  etc-glusterfs-glusterd.vol.log
[2015-11-07 10:49:16.045778] C [MSGID: 106003] 
[glusterd-server-quorum.c:346:glusterd_do_volume_quorum_action] 
0-management: Server quorum regained for volume 1KVM12-P4. Starting 
local bricks.
[2015-11-07 10:49:16.049319] C [MSGID: 106003] 
[glusterd-server-quorum.c:346:glusterd_do_volume_quorum_action] 
0-management: Server quorum regained for volume 1KVM12-P5. Starting 
local bricks.


regs.Paf1


___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] can oVirt 3.6 manage 3.5 hypervizor

2015-11-06 Thread p...@email.cz

Hi,
can oVirt 3.6 manage  hypervizors with 3.5 version ?
Meaning during cluster upgrade step by step.
(   A)  oVirt mgmt , B) 1st.hypervizor, C)  2nd hypervizor,  .. )
If oVirt DB converted from 3.5.5 -> 3.5.5.upg.3.6 -> final 3.6
regs. Paf1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] ovirt DB 3.6 to empty 3.6 DB

2015-11-05 Thread p...@email.cz

Hello,
can anybody help me with importing oVirt DB v.3.5 to new clean oVIrt 3.6 
database ??


# engine-backup --mode=restore --file=./backup --log=./restore.log
Preparing to restore:
- Unpacking file './backup'
FATAL: Backup version '3.6' doesn't match installed version

How to FIX it ??
Regs. Paf1
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] oVirt timeouts

2015-06-09 Thread p...@email.cz

Hello,

pls, howto  change oVirt timeouts for status  :
1) node ( brick) is power down / up
2) volume status for node(brick) is up/down

They are too long ( I'm expecting a few sec. not a lot of minutes )
If it has some spacial reason, let me know about, pls.

regs.
Pa.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] Replica2 stripe2 hang on write to VM disk

2015-05-25 Thread p...@email.cz

Hello,
can anybody help me with hanging  replica2 stripe2 datastore on 4 nodes 
cluster ??


oVirt -  ovirt-engine-lib-3.5.2.1-1.el7.centos.noarch
gluster - glusterfs-server-3.7.0-2.el7.x86_64
VM - Centos 7.1

If I use any bigger write to VM disk ( eg 2-5GB ) storage hosted virtual 
disk will hang = I/O error

created by :
gluster volume create 12KVM12SC4 replica 2 stripe 2 
16.0.0.161:/STORAGES/SlowClass/p4/GFS1 
16.0.0.162:/STORAGES/SlowClass/p4/GFS1 
16.0.0.163:/STORAGES/SlowClass/p4/GFS1 
16.0.0.164:/STORAGES/SlowClass/p4/GFS1


rhev-data-center-mnt-glusterSD-localhost:_12KVM12SC4.log
---
[2015-05-25 14:47:24.205609] I [rpc-clnt.c:1807:rpc_clnt_reconfig] 
0-12KVM12SC4-client-3: changing port to 49158 (from 0)
[2015-05-25 14:47:24.210824] I 
[client-handshake.c:1405:select_server_supported_programs] 
0-12KVM12SC4-client-3: Using Program GlusterFS 3.3, Num (1298437), 
Version (330)
[2015-05-25 14:47:24.211204] I 
[client-handshake.c:1193:client_setvolume_cbk] 0-12KVM12SC4-client-3: 
Connected to 12KVM12SC4-client-3, attached to remote volume 
'/STORAGES/SlowClass/p4/GFS1'.
[2015-05-25 14:47:24.211225] I 
[client-handshake.c:1203:client_setvolume_cbk] 0-12KVM12SC4-client-3: 
Server and Client lk-version numbers are not same, reopening the fds
[2015-05-25 14:47:24.211275] I [MSGID: 108005] 
[afr-common.c:3880:afr_notify] 0-12KVM12SC4-replicate-1: Subvolume 
'12KVM12SC4-client-3' came back up; going online.
[2015-05-25 14:47:24.216465] I [fuse-bridge.c:5077:fuse_graph_setup] 
0-fuse: switched to graph 0
[2015-05-25 14:47:24.216556] I 
[client-handshake.c:187:client_set_lk_version_cbk] 
0-12KVM12SC4-client-3: Server lk version = 1
[2015-05-25 14:47:24.216643] I [fuse-bridge.c:4007:fuse_init] 
0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 
kernel 7.22
[2015-05-25 14:47:24.217998] I 
[afr-common.c:1673:afr_local_discovery_cbk] 0-12KVM12SC4-replicate-0: 
selecting local read_child 12KVM12SC4-client-0
[2015-05-25 14:47:29.737732] W [fuse-bridge.c:1080:fuse_setattr_cbk] 
0-glusterfs-fuse: 40: SETATTR() /__DIRECT_IO_TEST__ = -1 (Read-only 
file system)
[2015-05-25 14:49:18.266212] E 
[client-handshake.c:1488:client_query_portmap_cbk] 
0-12KVM12SC4-client-2: failed to get the port number for remote 
subvolume. Please run 'gluster volume status' on server to see if brick 
process is running.
[2015-05-25 14:49:18.266274] I [client.c:2086:client_rpc_notify] 
0-12KVM12SC4-client-2: disconnected from 12KVM12SC4-client-2. Client 
process will keep trying to connect to glusterd until brick's port is 
available
[2015-05-25 14:49:19.346555] I [rpc-clnt.c:1807:rpc_clnt_reconfig] 
0-12KVM12SC4-client-2: changing port to 49158 (from 0)
[2015-05-25 14:49:19.351812] I 
[client-handshake.c:1405:select_server_supported_programs] 
0-12KVM12SC4-client-2: Using Program GlusterFS 3.3, Num (1298437), 
Version (330)
[2015-05-25 14:49:19.352169] I 
[client-handshake.c:1193:client_setvolume_cbk] 0-12KVM12SC4-client-2: 
Connected to 12KVM12SC4-client-2, attached to remote volume 
'/STORAGES/SlowClass/p4/GFS1'.
[2015-05-25 14:49:19.352191] I 
[client-handshake.c:1203:client_setvolume_cbk] 0-12KVM12SC4-client-2: 
Server and Client lk-version numbers are not same, reopening the fds
[2015-05-25 14:49:19.352242] I [MSGID: 108002] 
[afr-common.c:3959:afr_notify] 0-12KVM12SC4-replicate-1: Client-quorum 
is met
[2015-05-25 14:49:19.352353] I 
[client-handshake.c:187:client_set_lk_version_cbk] 
0-12KVM12SC4-client-2: Server lk version = 1
[2015-05-25 14:49:27.843616] W [fuse-bridge.c:1263:fuse_err_cbk] 
0-glusterfs-fuse: 151: REMOVEXATTR() /__DIRECT_IO_TEST__ = -1 (No data 
available)
[2015-05-25 14:49:58.356900] W [fuse-bridge.c:1263:fuse_err_cbk] 
0-glusterfs-fuse: 327: REMOVEXATTR() /__DIRECT_IO_TEST__ = -1 (No data 
available)



# gluster volume status
Status of volume: 12KVM12SC4
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick 16.0.0.161:/STORAGES/SlowClass/p4/GFS
1   49173 0  Y   17678
Brick 16.0.0.162:/STORAGES/SlowClass/p4/GFS
1   49158 0  Y   19184
Brick 16.0.0.163:/STORAGES/SlowClass/p4/GFS
1   49158 0  Y   9784
Brick 16.0.0.164:/STORAGES/SlowClass/p4/GFS
1   49158 0  Y   9327
NFS Server on localhost 2049 0  Y   17697
Self-heal Daemon on localhost   N/A N/AY   17708
NFS Server on 16.0.0.1622049 0  Y   19205
Self-heal Daemon on 16.0.0.162  N/A N/AY   19215
NFS Server on 16.0.0.1632049 0  Y   9806
Self-heal Daemon on 16.0.0.163  N/A N/AY   9813
NFS Server 

[ovirt-users] VM pause when host crash/poweroff

2015-05-23 Thread p...@email.cz

Hello,
how to solve situation if  VM going to pause mode  when one of replica 
2 node is stopped ( power down or crash, etc. .. )
VM is not running on stopped node and OVirt message get  VM xyz has 
paused due to unknown storage error

Where the error is ??

regs.
Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] snapshot deleting

2015-05-22 Thread p...@email.cz

Hello,
my question is about deleting default snapshots generated via disk 
moving between storages.
The snapshot called Auto-generated for Live Storage Migration is NOT 
automaticaly deleted after  movement done.

Is this default or any bug ??

regs.
Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] disk format hang

2015-05-20 Thread p...@email.cz

Hello,

during VM instalation I needed add 300GB new disk and format as EXT3.  
Easy task, but 

Used  striped mirror volume for that  - created by:
# gluster volume create 12KVM12SC4 replica 2 stripe 2 1kvm1-SAN:/p4/GFS1 
1kvm2-SAN:/p4/GFS1 2kvm1-SAN /p4/GFS1 2kvm2-SAN:/p4/GFS1 force

- added disk throught  oVisrt GUI
- VM# partprobe
- VM# fdisk - make one partition, type linux
- VM# mkfs.ext3 /dev/vdb1
it hangs after cca 30% of format

oVirt alert:
*VM has been paused due to a storage I/O error*

so I tried to copy some  data directly from hypervizor ( cca 80GB ) to 
default gluster-ovirt mountpoint ( 
/rhev/data-center/mnt/glusterSD/localhost:_12KVM12SC4 ) with no hangs.

The same whole operation on volume distribute replica type succesfully done.

Can anybody help me with it ??

Maybe continual error on /var/log/messages occure will get right view on 
it :


May 20 15:47:00 1kvm2 virsh: All-whitespace username.
May 20 15:47:00 1kvm2 journal: End of file while reading data: 
Input/output error

May 20 15:47:00 1kvm2 virsh: All-whitespace username.
May 20 15:47:00 1kvm2 journal: End of file while reading data: 
Input/output error

May 20 15:48:00 1kvm2 virsh: All-whitespace username.
May 20 15:48:00 1kvm2 journal: End of file while reading data: 
Input/output error

May 20 15:48:00 1kvm2 virsh: All-whitespace username.
May 20 15:48:00 1kvm2 journal: End of file while reading data: 
Input/output error


regs to ALL

!! URGENT !!

Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] disk moving

2015-05-11 Thread p...@email.cz

hello, would I explicitly define interface for only disk moving ??

reg.
Pa.
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] simple virtual disk moving

2015-05-11 Thread p...@email.cz

hello, would I explicitly define interface for only disk moving ??

Move process takes the first interface from list  ( VM assigned ), but 
any  interface for moving virtuals ( explicitly defined ).


eg.
VM moving from A to B - it takes interface defined for moving ( 
movingLAN )

but
moving only simple disk of VM from A to B - it takes default interface 
from list of assigned LANs


any idea ??
regs.
Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] gluster storage ( wanted / supported / recommended configuration )

2015-05-11 Thread p...@email.cz

Hello dears,

is anybody here for serious storage conversation ??
I've got any ideas and a lot of errors - gluster filesystem concept checking

let me know, will sent diagram and questions 

regs.
Pavel
___
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


[ovirt-users] ovirt - import detached gluter volumes

2015-04-27 Thread p...@email.cz

Hello dears,
i've got some troubles with reattaching gluster volumes with data.

1) Base on a lot of tests I decided clear oVirt database ( # 
engine-cleanup ; # yum remove ovirt-engine;  # yum -y install 
ovirt-engine; #  engine-setup)

2) clearing sucessfully done and start with empty oVirt envir.
3) then I added networks, nodes and make basic network adjustment = all 
works fine
4) time to attach  volumes/ domains with original data ( a lot of 
virtuals , ISO files ,  )


So, main question is about HOWTO attach this volumes if I haven't 
defined any domain and can't clearly  import them ??


Current status of nodes are without glusterfs NFS mounted, but bricks are OK

# gluster volume info

Volume Name: VGFS1
Type: Replicate
Volume ID: b9a1c347-6ffd-4122-8756-d513fe3f40b9
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 1kvm2:/FastClass/p1/GFS1
Brick2: 1kvm1:/FastClass/p1/GFS1
Options Reconfigured:
storage.owner-gid: 36
storage.owner-uid: 36

Volume Name: VGFS2
Type: Replicate
Volume ID: b65bb689-ecc8-4c33-a4e7-11dea6028f83
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 1kvm2:/FastClass/p2/GFS1
Brick2: 1kvm1:/FastClass/p2/GFS1
Options Reconfigured:
storage.owner-uid: 36
storage.owner-gid: 36


[root@1kvm1 glusterfs]# gluster volume status
Status of volume: VGFS1
Gluster process PortOnline  Pid
--
Brick 1kvm1:/FastClass/p1/GFS1 N/A N   N/A
NFS Server on localhost N/A N   N/A
Self-heal Daemon on localhost N/A N   N/A

Task Status of Volume VGFS1
--
There are no active volume tasks

Status of volume: VGFS2
Gluster process PortOnline  Pid
--
Brick 1kvm1:/FastClass/p2/GFS1 N/A N   N/A
NFS Server on localhost N/A N   N/A
Self-heal Daemon on localhost N/A N   N/A

Task Status of Volume VGFS2
--
There are no active volume tasks

[root@1kvm1 glusterfs]# gluster volume start VGFS1
volume start: VGFS1: failed: Volume VGFS1 already started



# mount | grep mapper # base XFS mounting
/dev/mapper/3600605b0099f9e601cb1b5bf0e9765e8p1 on /FastClass/p1 type 
xfs (rw,relatime,seclabel,attr2,inode64,noquota)
/dev/mapper/3600605b0099f9e601cb1b5bf0e9765e8p2 on /FastClass/p2 type 
xfs (rw,relatime,seclabel,attr2,inode64,noquota)



*5)* import screen
/VGFS1 dir exists  iptables flushed


# cat rhev-data-center-mnt-glusterSD-1kvm1:_VGFS1.log
[2015-04-22 15:21:50.204521] I [MSGID: 100030] [glusterfsd.c:2018:main] 
0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.6.2 
(args: /usr/sbin/glusterfs --volfile-server=1kvm1 --volfile-id=/VGFS1 
/rhev/data-center/mnt/glusterSD/1kvm1:_VGFS1)
[2015-04-22 15:21:50.220383] I [dht-shared.c:337:dht_init_regex] 
0-VGFS1-dht: using regex rsync-hash-regex = ^\.(.+)\.[^.]+$
[2015-04-22 15:21:50.55] I [client.c:2280:notify] 0-VGFS1-client-1: 
parent translators are ready, attempting connect on transport
[2015-04-22 15:21:50.224528] I [client.c:2280:notify] 0-VGFS1-client-2: 
parent translators are ready, attempting connect on transport

Final graph:
+--+
  1: volume VGFS1-client-1
  2: type protocol/client
  3: option ping-timeout 42
  4: option remote-host 1kvm2
  5: option remote-subvolume /FastClass/p1/GFS1
  6: option transport-type socket
  7: option username 52f1efd1-60dc-4fb1-b94f-572945d6eb66
  8: option password 34bac9cd-0b4f-41c6-973b-7af568784d7b
  9: option send-gids true
 10: end-volume
 11:
 12: volume VGFS1-client-2
 13: type protocol/client
 14: option ping-timeout 42
 15: option remote-host 1kvm1
 16: option remote-subvolume /FastClass/p1/GFS1
 17: option transport-type socket
 18: option username 52f1efd1-60dc-4fb1-b94f-572945d6eb66
 19: option password 34bac9cd-0b4f-41c6-973b-7af568784d7b
 20: option send-gids true
 21: end-volume
 22:
 23: volume VGFS1-replicate-0
 24: type cluster/replicate
 25: subvolumes VGFS1-client-1 VGFS1-client-2
 26: end-volume
 27:
 28: volume VGFS1-dht
 29: type cluster/distribute
 30: subvolumes VGFS1-replicate-0
 31: end-volume
 32:
 33: volume VGFS1-write-behind
 34: type performance/write-behind
 35: subvolumes VGFS1-dht
 36: end-volume
 37:
 38: volume VGFS1-read-ahead
 39: type performance/read-ahead
 40: subvolumes VGFS1-write-behind
 41: end-volume
 42:
 43: volume VGFS1-io-cache
 44: type performance/io-cache
 45: subvolumes VGFS1-read-ahead
 46: end-volume
 47:
 48: volume VGFS1-quick-read
 49: type performance/quick-read
 50: subvolumes VGFS1-io-cache
 51: end-volume
 52:
 53: volume VGFS1-open-behind
 54: