[ovirt-users] Re: gluster on iSCSI devices in ovirt environment
Hello, yes, U're right, but only as separate domain ( ? ( mirror realized by any clever storage on background ) , but if mirror needed over two locations ? I had an idea realize mirror over two loacation via gluster ( with iscsi bricks ) Pa. On 1/18/24 09:21, Strahil Nikolov wrote: Hi, Why would you do that? Ovirt already supports iSCSI. Best Regards, Strahil Nikolov On Thu, Jan 18, 2024 at 10:20, p...@email.cz wrote: hello dears, can anybody explain me HOWTO realize 2 nodes + aribiter gluster from two (three) locations on block iSCSI devices ? Something like this: gluster volume create TEST replica 3 arbiter 1 < location-three-host3 - /dev/sda5 e.g. > - ALL applied on multinode ovirt cluster thx a lot for any help regs. Pa. ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/VBP7TKZNWLOCY7IAQNEAHWBQXRSQBPE5/ ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/C2HOVCHXS2IWYFYTOPCE2EMKFUKOC4F6/
[ovirt-users] gluster on iSCSI devices in ovirt environment
hello dears, can anybody explain me HOWTO realize 2 nodes + aribiter gluster from two (three) locations on block iSCSI devices ? Something like this: gluster volume create TEST replica 3 arbiter 1 iSCSI target > < location-three-host3 - /dev/sda5 e.g. > - ALL applied on multinode ovirt cluster thx a lot for any help regs. Pa.___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/VBP7TKZNWLOCY7IAQNEAHWBQXRSQBPE5/
[ovirt-users] Re: ovirt 4.2.7-1 - adding virtual host ( nested virt. )
Hello, problem was fixed now. Missing nested virtualization = enabled option. ( set on VM config > CPU -> hardware virtualization ) Thx for response. Pa. On 1/10/19 4:39 PM, Shani Leviim wrote: Hi Paul, Can you please attach vdsm & engine logs? *Regards, * *Shani Leviim * On Mon, Jan 7, 2019 at 8:44 PM p...@email.cz <mailto:p...@email.cz> mailto:p...@email.cz>> wrote: Hello guys, I've got problem with adding new host (ESX-virtual) to ovirt 4.2.7-1 ( gluster included) Is this feature supported ??? 2019-01-07 19:38:30,168+01 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler1) [15a4029b] Error while refreshing server data for cluster 'MID' from database: null regs. Paul ___ Users mailing list -- users@ovirt.org <mailto:users@ovirt.org> To unsubscribe send an email to users-le...@ovirt.org <mailto:users-le...@ovirt.org> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TREULSOFHYWLGLK7H557NT4WEYPBUXPU/ ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/RFH43FBXQOFRA6WTSLOMTQTS2N6OY5VA/
[ovirt-users] ovirt 4.2.7-1 - adding virtual host ( nested virt. )
Hello guys, I've got problem with adding new host (ESX-virtual) to ovirt 4.2.7-1 ( gluster included) Is this feature supported ??? 2019-01-07 19:38:30,168+01 ERROR [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler1) [15a4029b] Error while refreshing server data for cluster 'MID' from database: null regs. Paul ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/TREULSOFHYWLGLK7H557NT4WEYPBUXPU/
Re: [ovirt-users] VDSM hang
0,00 0,000,000,00 0,00 0,00 0,00 0,000,000,000,00 0,00 0,00 dm-17 0,00 0,000,000,00 0,00 0,00 0,00 0,000,000,000,00 0,00 0,00 dm-18 0,00 0,000,000,00 0,00 0,00 0,00 0,000,000,000,00 0,00 0,00 look w_await ( write queue ) concurent tasks: - The data import is about 1,5Gb over then 10Gb LAN - VM clone on the same volume ( 300G ) - VM snapshot deleting on the same volume ( cca 1TB ) - VM with Oracle DB start/stop - VM seq.write ( 500GB ) Yes - it could happen in one time regards Paf1 On 03/13/2017 06:01 AM, Krutika Dhananjay wrote: Hi, Could you please share your volume info output? -Krutika On Fri, Mar 10, 2017 at 6:41 PM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: freez / freezing IO operations are paused from any reasons available posibilities are 1) net - any tcp framework collapse 2) gluster interconnect due gluster daemon - process hang ?? 3) VSD - pause managed services 4) XFS - RW issues 5) swap overfulled - any processes are killed - but why swap is full if max 30% of mem (196 GB ) is used by VMs ? ( unmanaged process forking ) regs On 03/10/2017 01:56 PM, Nir Soffer wrote: On Fri, Mar 10, 2017 at 1:07 PM,p...@email.cz <mailto:p...@email.cz> <p...@email.cz> <mailto:p...@email.cz> wrote: Hello everybody, for production usage i'm testing ovirt with gluster. All components seems to be running fine but whenever I'm testing huge workload, then node freez. Not the main OS, but VDSM mgmt and attached services, VMs eg. What do you mean by freez? mgmt oVirt - 4.1.0.4 centos 7.3-1611 nodes ( installed from ovirt image "ovirt-node-ng-installer-ovirt-4.1-2017030804.iso" ) OS Version: == RHEL - 7 - 3.1611.el7.centos OS Description:== oVirt Node 4.1.0 Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64 KVM Version:== 2.6.0 - 28.el7_3.3.1 LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5 VDSM Version:== vdsm-4.19.4-1.el7.centos SPICE Version:== 0.12.4 - 20.el7_3 GlusterFS Version:== glusterfs-3.8.9-1.el7 ( LVM thinprovisioning in replica 2 - created from ovirt GUI ) concurently running - huge import from export domain( net workload ) - sequential write to VMs local disk ( gluster replica sequential workload ) - VMs database huge select ( random IOps ) - huge old snapshot delete ( random IOps ) In this configuration / workload is runnig one hour eg, with no exceptions , with 70-80% disk load, but in some point VDSM freez all jobs for a timeout and VM's are in "uknown" status . The whole system revitalize then automaticaly in cca 20min time frame ( except the import and snapshot deleting(rollback) ) engine.log - focus 10:39:07 time ( Failed in 'HSMGetAllTasksStatusesVDS' method ) n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508' type:'RemoveSnapshotSingleDiskLive' to complete 2017-03-10 10:39:01,727+01 INFO [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback] (DefaultQuartzScheduler2) [759c8e1f] Command 'RemoveSnapshotSingleDiskLive' (id: 'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id: '33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete 2017-03-10 10:39:03,929+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, GlusterServersListVDSCommand(HostName = 2kvm1, VdsIdVDSCommandParametersBase:{runAsync='true', hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3 2017-03-10 10:39:04,343+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED <http://172.16.5.163/24:CONNECTED>, 16.0.0.164:CONNECTED], log id: 446d0cd3 2017-03-10 10:39:04,353+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, GlusterVolumesListVDSCommand(HostName = 2kvm1, GlusterVolumesListVDSParameters:{runAsync='true', hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda 2017-03-10 10:39:05,128+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, GlusterVolumesListVDSCommand, return: {8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.glus
Re: [ovirt-users] VDSM hang
freez / freezing IO operations are paused from any reasons available posibilities are 1) net - any tcp framework collapse 2) gluster interconnect due gluster daemon - process hang ?? 3) VSD - pause managed services 4) XFS - RW issues 5) swap overfulled - any processes are killed - but why swap is full if max 30% of mem (196 GB ) is used by VMs ? ( unmanaged process forking ) regs On 03/10/2017 01:56 PM, Nir Soffer wrote: On Fri, Mar 10, 2017 at 1:07 PM, p...@email.cz <p...@email.cz> wrote: Hello everybody, for production usage i'm testing ovirt with gluster. All components seems to be running fine but whenever I'm testing huge workload, then node freez. Not the main OS, but VDSM mgmt and attached services, VMs eg. What do you mean by freez? mgmt oVirt - 4.1.0.4 centos 7.3-1611 nodes ( installed from ovirt image "ovirt-node-ng-installer-ovirt-4.1-2017030804.iso" ) OS Version: == RHEL - 7 - 3.1611.el7.centos OS Description:== oVirt Node 4.1.0 Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64 KVM Version:== 2.6.0 - 28.el7_3.3.1 LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5 VDSM Version:== vdsm-4.19.4-1.el7.centos SPICE Version:== 0.12.4 - 20.el7_3 GlusterFS Version:== glusterfs-3.8.9-1.el7 ( LVM thinprovisioning in replica 2 - created from ovirt GUI ) concurently running - huge import from export domain( net workload ) - sequential write to VMs local disk ( gluster replica sequential workload ) - VMs database huge select ( random IOps ) - huge old snapshot delete ( random IOps ) In this configuration / workload is runnig one hour eg, with no exceptions , with 70-80% disk load, but in some point VDSM freez all jobs for a timeout and VM's are in "uknown" status . The whole system revitalize then automaticaly in cca 20min time frame ( except the import and snapshot deleting(rollback) ) engine.log - focus 10:39:07 time ( Failed in 'HSMGetAllTasksStatusesVDS' method ) n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508' type:'RemoveSnapshotSingleDiskLive' to complete 2017-03-10 10:39:01,727+01 INFO [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback] (DefaultQuartzScheduler2) [759c8e1f] Command 'RemoveSnapshotSingleDiskLive' (id: 'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id: '33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete 2017-03-10 10:39:03,929+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, GlusterServersListVDSCommand(HostName = 2kvm1, VdsIdVDSCommandParametersBase:{runAsync='true', hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3 2017-03-10 10:39:04,343+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id: 446d0cd3 2017-03-10 10:39:04,353+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, GlusterVolumesListVDSCommand(HostName = 2kvm1, GlusterVolumesListVDSParameters:{runAsync='true', hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda 2017-03-10 10:39:05,128+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, GlusterVolumesListVDSCommand, return: {8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad}, log id: 69ea1fda 2017-03-10 10:39:07,163+01 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand] (DefaultQuartzScheduler2) [759c8e1f] Failed in 'HSMGetAllTasksStatusesVDS' method 2017-03-10 10:39:07,178+01 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (DefaultQuartzScheduler2) [759c8e1f] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: VDSM 2kvm2 command HSMGetAllTasksStatusesVDS failed: Connection timed out 2017-03-10 10:39:07,182+01 INFO [org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) [759c8e1f] BaseAsyncTask::onTaskEndSuccess: Task 'f594bf69-619b-4d1b-8f6d-a9826997e478' (Parent Command 'ImportVm', Parameters Type 'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') ended successfully. 2017-03-10 10:39:07,182+01 INFO [org.ovirt.engine.core.bll.CommandMultiAsyncTasks] (DefaultQuartzScheduler2) [759c8e1f] Task with DB Task ID 'a05c7c07-9b98-4ab2-ac7b-9e70a75ba7b7' and VDSM Task ID '7c60369f-70a3-4a6a-80c9-4753ac9ed372' is in state Polling. End action for command 8deb3fe3-4a83-4605-816c-ffdc63fd9ac1 will proceed when all the entity's tasks are completed. 2017-03-10 10:39:07,182+01 INFO [org.ovirt.engine.c
[ovirt-users] VDSM hang
Hello everybody, for production usage i'm testing ovirt with gluster. All components seems to be running fine but whenever I'm testing huge workload, then node freez. Not the main OS, but VDSM mgmt and attached services, VMs eg. *mgmt * oVirt - 4.1.0.4 centos 7.3-1611 *nodes* ( installed from ovirt image /"ovirt-node-ng-installer-ovirt-4.1-2017030804.iso" )/ OS Version:== RHEL - 7 - 3.1611.el7.centos OS Description:== oVirt Node 4.1.0 Kernel Version:== 3.10.0 - 514.10.2.el7.x86_64 KVM Version:== 2.6.0 - 28.el7_3.3.1 LIBVIRT Version:== libvirt-2.0.0-10.el7_3.5 VDSM Version:== vdsm-4.19.4-1.el7.centos SPICE Version:== 0.12.4 - 20.el7_3 GlusterFS Version:== glusterfs-3.8.9-1.el7 ( LVM thinprovisioning in replica 2 - created from ovirt GUI ) concurently running - huge import from export domain( net workload ) - sequential write to VMs local disk ( gluster replica sequential workload ) - VMs database huge select ( random IOps ) - huge old snapshot delete ( random IOps ) In this configuration / workload is runnig one hour eg, with no exceptions , with 70-80% disk load, but in some point VDSM freez all jobs for a timeout and VM's are in "uknown" status . The whole system revitalize then automaticaly in cca 20min time frame ( except the import and snapshot deleting(rollback) ) engine.log - focus 10:39:07 time ( Failed in 'HSMGetAllTasksStatusesVDS' method ) n child command id: 'a8a3a4d5-cf7d-4423-8243-022911232508' type:'RemoveSnapshotSingleDiskLive' to complete 2017-03-10 10:39:01,727+01 INFO [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback] (DefaultQuartzScheduler2) [759c8e1f] Command 'RemoveSnapshotSingleDiskLive' (id: 'a8a3a4d5-cf7d-4423-8243-022911232508') waiting on child command id: '33df2c1e-6ce3-44fd-a39b-d111883b4c4e' type:'DestroyImage' to complete 2017-03-10 10:39:03,929+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, GlusterServersListVDSCommand(HostName = 2kvm1, VdsIdVDSCommandParametersBase:{runAsync='true', hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 446d0cd3 2017-03-10 10:39:04,343+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, GlusterServersListVDSCommand, return: [172.16.5.163/24:CONNECTED, 16.0.0.164:CONNECTED], log id: 446d0cd3 2017-03-10 10:39:04,353+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] START, GlusterVolumesListVDSCommand(HostName = 2kvm1, GlusterVolumesListVDSParameters:{runAsync='true', hostId='86876b79-71d8-4ae1-883b-ba010cd270e7'}), log id: 69ea1fda 2017-03-10 10:39:05,128+01 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler5) [fde51205-3e8b-4b84-a478-352dc444ccc4] FINISH, GlusterVolumesListVDSCommand, return: {8ded4083-2f31-489e-a60d-a315a5eb9b22=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7765e4ad}, log id: 69ea1fda 2017-03-10 10:39:07,163+01 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand] (DefaultQuartzScheduler2) [759c8e1f] Failed in 'HSMGetAllTasksStatusesVDS' method 2017-03-10 10:39:07,178+01 ERROR [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (DefaultQuartzScheduler2) [759c8e1f] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: null, Call Stack: null, Custom Event ID: -1, Message: VDSM 2kvm2 command HSMGetAllTasksStatusesVDS failed: Connection timed out 2017-03-10 10:39:07,182+01 INFO [org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) [759c8e1f] BaseAsyncTask::onTaskEndSuccess: Task 'f594bf69-619b-4d1b-8f6d-a9826997e478' (Parent Command 'ImportVm', Parameters Type 'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') ended successfully. 2017-03-10 10:39:07,182+01 INFO [org.ovirt.engine.core.bll.CommandMultiAsyncTasks] (DefaultQuartzScheduler2) [759c8e1f] Task with DB Task ID 'a05c7c07-9b98-4ab2-ac7b-9e70a75ba7b7' and VDSM Task ID '7c60369f-70a3-4a6a-80c9-4753ac9ed372' is in state Polling. End action for command 8deb3fe3-4a83-4605-816c-ffdc63fd9ac1 will proceed when all the entity's tasks are completed. 2017-03-10 10:39:07,182+01 INFO [org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) [759c8e1f] SPMAsyncTask::PollTask: Polling task 'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent Command 'DestroyImage', Parameters Type 'org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters') returned status 'finished', result 'cleanSuccess'. 2017-03-10 10:39:07,182+01 ERROR [org.ovirt.engine.core.bll.tasks.SPMAsyncTask] (DefaultQuartzScheduler2) [759c8e1f] BaseAsyncTask::logEndTaskFailure: Task 'f351e8f6-6dd7-49aa-bf54-650d84fc6352' (Parent
[ovirt-users] GUI node detail long delay
Hello everybody, We are using oVirt Engine Version: 4.0.6.3-1.el7.centos on centos 7.3 with gluster replica 3 arbiter = (1+1)+1 I'm confused with GUI delaying - if node details are wanted ( cluster -> nodes -> node detail = click on node raw ) then request generate over 10 min delay to display details. This unexpected mistake didn't occure initially, but later - not specified when . The followed partial list of "engine.log" shows requests to "arbiter node" ( 16.0.0.159) connectivity. This requested 3rd node of gluster(arbiter) is NOT included in oVirt environment and will NOT. Maybe this is that problem, but I'm not shure, especially how to fix this. 2017-02-06 13:20:03,924 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler3) [49cebf0] START, GlusterServersListVDSCommand(HostName = 1kvm2, VdsIdVDSCommandParametersBase:{runAsync='true', hostId='258decac-46f4-4c15-b855-ad97b570ee60'}), log id: 6873151 2017-02-06 13:20:04,796 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] (DefaultQuartzScheduler3) [49cebf0] FINISH, GlusterServersListVDSCommand, return: [172.16.5.162/24:CONNECTED, 172.16.5.161:CONNECTED, 16.0.0.159:CONNECTED], log id: 6873151 2017-02-06 13:20:04,814 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler3) [49cebf0] START, GlusterVolumesListVDSCommand(HostName = 1kvm2, GlusterVolumesListVDSParameters:{runAsync='true', hostId='258decac-46f4-4c15-b855-ad97b570ee60'}), log id: 381ae630 2017-02-06 13:20:05,970 WARN [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] (DefaultQuartzScheduler3) [49cebf0] Could not add brick '16.0.0.159:/GLUSTER/1KVM12-sda2/GFS' to volume '19c27787-f1c9-4dee-8415-c6d1c81e3aa2' - server uuid 'f7670ea9-2204-4310-96a6-243c2c6a00de' not found in cluster '587fa2d8-017d-03b3-0003-030d' 2017-02-06 13:20:05,987 WARN [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturnForXmlRpc] (DefaultQuartzScheduler3) [49cebf0] Could not add brick '16.0.0.159:/GLUSTER/1KVM12-sda1/GFS' to volume '96adac2a-0dc4-4bd8-ad79-23dd3448f73b' - server uuid 'f7670ea9-2204-4310-96a6-243c2c6a00de' not found in cluster '587fa2d8-017d-03b3-0003-030d' 2017-02-06 13:20:05,987 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler3) [49cebf0] FINISH, GlusterVolumesListVDSCommand, return: {19c27787-f1c9-4dee-8415-c6d1c81e3aa2=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@b9f51962, 96adac2a-0dc4-4bd8-ad79-23dd3448f73b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@86597dda}, log id: 381ae630 repeatelly occured several times per minute, so huge filling logs OS Version:RHEL - 7 - 3.1611.el7.centos OS Description:CentOS Linux 7 (Core) Kernel Version:3.10.0 - 514.6.1.el7.x86_64 KVM Version:2.6.0 - 28.el7_3.3.1 LIBVIRT Version:libvirt-2.0.0-10.el7_3.4 VDSM Version:vdsm-4.18.21-1.el7.centos SPICE Version:0.12.4 - 19.el7 GlusterFS Version:glusterfs-3.8.8-1.el7 CEPH Version:librbd1-0.94.5-1.el7 regards Paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] fast import to ovirt
Hello, how can I import Vm from different ovirt envir..? There is no common mgmt ovirt. ( ovirt 3.5 -> 4.0 ) Gluster FS used. Will ovirt accept "rsync" file migrations , meaning will update oVirt DB automaticaly ? I'd prefer more quickly method then export-umount oV1-mount oV2-import . regards paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] ovirt with glusterfs - big test - unwanted results
Hello Yaniv, we tried another small test - reboot two nodes from replica3 a1 ( 1HP12-R3A1P1 ) which keep master domain. All domains went down = master down, but master domain didn't move to another available domain ( eg. 2HP12-R3A1P1 ). It looks that "master domain" management isn't correct ( has a bug ?? ) regs. Pavel On 31.3.2016 14:30, Yaniv Kaul wrote: Hi Pavel, Thanks for the report. Can you begin with a more accurate description of your environment? Begin with host, oVirt and Gluster versions. Then continue with the exact setup (what are 'A', 'B', 'C' - domains? Volumes? What is the mapping between domains and volumes?). Are there any logs you can share with us? I'm sure with more information, we'd be happy to look at the issue. Y. On Thu, Mar 31, 2016 at 3:09 PM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: Hello, we tried the following test - with unwanted results input: 5 node gluster A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 ) B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 ) C = distributed replica 3 arbiter 1 ( node1+node2, node3+node4, each arbiter on node 5) node 5 has only arbiter replica ( 4x ) TEST: 1) directly reboot one node - OK ( is not important which ( data node or arbiter node )) 2) directly reboot two nodes - OK ( if nodes are not from the same replica ) 3) directly reboot three nodes - yes, this is the main problem and a questions - rebooted all three nodes from replica "B" ( not so possible, but who knows ... ) - all VMs with data on this replica was paused ( no data access ) - OK - all VMs running on replica "B" nodes lost ( started manually, later )( datas on other replicas ) - acceptable BUT - !!! all oVIrt domains went down !! - master domain is on replica "A" which lost only one member from three !!! so we are not expecting that all domain will go down, especially master with 2 live members. Results: - the whole cluster unreachable until at all domains up - depent of all nodes up !!! - all paused VMs started back - OK - rest of all VMs rebooted and runnig - OK Questions: 1) why all domains down if master domain ( on replica "A" ) has two runnig members ( 2 of 3 ) ?? 2) how to fix that colaps without waiting to all nodes up ? ( in worste case if node has HW error eg. ) ?? 3) which oVirt cluster policy can prevent that situation ?? ( if any ) regs. Pavel ___ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] ovirt with glusterfs - big test - unwanted results
Hi, rest of logs: www.uschovna.cz/en/zasilka/HYGXR57CNHM3TP39-L3W <http://www.uschovna.cz/en/zasilka/HYGXR57CNHM3TP39-L3W> The TEST is the last big event in logs TEST TIME : about 14:00-14:30 CET regs.Pavel On 31.3.2016 14:30, Yaniv Kaul wrote: Hi Pavel, Thanks for the report. Can you begin with a more accurate description of your environment? Begin with host, oVirt and Gluster versions. Then continue with the exact setup (what are 'A', 'B', 'C' - domains? Volumes? What is the mapping between domains and volumes?). Are there any logs you can share with us? I'm sure with more information, we'd be happy to look at the issue. Y. On Thu, Mar 31, 2016 at 3:09 PM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: Hello, we tried the following test - with unwanted results input: 5 node gluster A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 ) B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 ) C = distributed replica 3 arbiter 1 ( node1+node2, node3+node4, each arbiter on node 5) node 5 has only arbiter replica ( 4x ) TEST: 1) directly reboot one node - OK ( is not important which ( data node or arbiter node )) 2) directly reboot two nodes - OK ( if nodes are not from the same replica ) 3) directly reboot three nodes - yes, this is the main problem and a questions - rebooted all three nodes from replica "B" ( not so possible, but who knows ... ) - all VMs with data on this replica was paused ( no data access ) - OK - all VMs running on replica "B" nodes lost ( started manually, later )( datas on other replicas ) - acceptable BUT - !!! all oVIrt domains went down !! - master domain is on replica "A" which lost only one member from three !!! so we are not expecting that all domain will go down, especially master with 2 live members. Results: - the whole cluster unreachable until at all domains up - depent of all nodes up !!! - all paused VMs started back - OK - rest of all VMs rebooted and runnig - OK Questions: 1) why all domains down if master domain ( on replica "A" ) has two runnig members ( 2 of 3 ) ?? 2) how to fix that colaps without waiting to all nodes up ? ( in worste case if node has HW error eg. ) ?? 3) which oVirt cluster policy can prevent that situation ?? ( if any ) regs. Pavel ___ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] ovirt with glusterfs - big test - unwanted results
Hello, some envir. answers : * OS = RHEL - 7 - 2.151 kernel = 3.10.0 - 327.10.1.el7.x86_64 KVM = 2.3.0 - 31.el7_2.7.1 libvirt = libvirt-1.2.17-13.el7_2.3 vdsm = vdsm-4.17.23.2-0.el7 glusterfs = glusterfs-3.7.9-1.el7 ovirt = 3.5.6.2-1 * # gluster peer status Number of Peers: 4 Hostname: 1hp2 Uuid: 8e87cf18-8958-41b7-8d24-7ee420a1ef9f State: Peer in Cluster (Connected) Hostname: 2hp2 Uuid: b1d987d8-0b42-4ce4-b85f-83b4072e0990 State: Peer in Cluster (Connected) Hostname: 2hp1 Uuid: a1cbe1a8-88ad-4e89-8a0e-d2bb2b6786d8 State: Peer in Cluster (Connected) Hostname: kvmarbiter Uuid: bb1d63f1-7757-4c07-b70d-aa2f68449e21 State: Peer in Cluster (Connected) * == "C" == Volume Name: 12HP12-D2R3A1P2 Type: Distributed-Replicate Volume ID: 3c22d3dc-7c6e-4e37-9e0b-78410873ed6d Status: Started Number of Bricks: 2 x (2 + 1) = 6 Transport-type: tcp Bricks: Brick1: 1hp1:/STORAGES/P2/GFS Brick2: 1hp2:/STORAGES/P2/GFS Brick3: kvmarbiter:/STORAGES/P2-1/GFS (arbiter) Brick4: 2hp1:/STORAGES/P2/GFS Brick5: 2hp2:/STORAGES/P2/GFS Brick6: kvmarbiter:/STORAGES/P2-2/GFS (arbiter) Options Reconfigured: performance.readdir-ahead: on * == "A" == Volume Name: 1HP12-R3A1P1 Type: Replicate Volume ID: e4121610-6128-4ecc-86d3-1429ab3b8356 Status: Started Number of Bricks: 1 x (2 + 1) = 3 Transport-type: tcp Bricks: Brick1: 1hp1:/STORAGES/P1/GFS Brick2: 1hp2:/STORAGES/P1/GFS Brick3: kvmarbiter:/STORAGES/P1-1/GFS (arbiter) Options Reconfigured: performance.readdir-ahead: on performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on features.shard-block-size: 512MB cluster.data-self-heal-algorithm: full performance.write-behind: on performance.low-prio-threads: 32 performance.write-behind-window-size: 128MB network.ping-timeout: 10 * == "B" == Volume Name: 2HP12-R3A1P1 Type: Replicate Volume ID: d3d260cd-455f-42d6-9580-d88ae6df0519 Status: Started Number of Bricks: 1 x (2 + 1) = 3 Transport-type: tcp Bricks: Brick1: 2hp1:/STORAGES/P1/GFS Brick2: 2hp2:/STORAGES/P1/GFS Brick3: kvmarbiter:/STORAGES/P1-2/GFS (arbiter) Options Reconfigured: performance.readdir-ahead: on performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server storage.owner-uid: 36 storage.owner-gid: 36 features.shard: on features.shard-block-size: 512MB cluster.data-self-heal-algorithm: full performance.write-behind: on performance.low-prio-threads: 32 performance.write-behind-window-size: 128MB network.ping-timeout: 10 The oVirt volumes(storages) have the same name as gluster volumes ( eg: "B" = 2HP12-R3A1P1( ovirt storage ) = 2HP12-R3A1P1( gluster volume name ) ) In the test the master volume was "A" = 1HP12-R3A1P1 regs. Pavel PS: logs will follow as webstore pointer ... this takes some time On 31.3.2016 14:30, Yaniv Kaul wrote: Hi Pavel, Thanks for the report. Can you begin with a more accurate description of your environment? Begin with host, oVirt and Gluster versions. Then continue with the exact setup (what are 'A', 'B', 'C' - domains? Volumes? What is the mapping between domains and volumes?). Are there any logs you can share with us? I'm sure with more information, we'd be happy to look at the issue. Y. On Thu, Mar 31, 2016 at 3:09 PM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: Hello, we tried the following test - with unwanted results input: 5 node gluster A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 ) B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 ) C = distributed replica 3 arbiter 1 ( node1+node2, node3+node4, each arbiter on node 5) node 5 has only arbiter replica ( 4x ) TEST: 1) directly reboot one node - OK ( is not important which ( data node or arbiter node )) 2) directly reboot two nodes - OK ( if nodes are not from the same replica ) 3) directly reboot three nodes - yes, this is the main problem and a questions - rebooted all three nodes from replica "B" ( not so possible, but who knows ... ) - all VMs with data on this replica was paused ( no data access ) - OK - all VMs running on replica "B" nodes lost ( started manually, later )( datas on other replicas ) - acceptable BUT - !!! all oVIrt domains
[ovirt-users] ovirt with glusterfs - big test - unwanted results
Hello, we tried the following test - with unwanted results input: 5 node gluster A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 ) B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 ) C = distributed replica 3 arbiter 1 ( node1+node2, node3+node4, each arbiter on node 5) node 5 has only arbiter replica ( 4x ) TEST: 1) directly reboot one node - OK ( is not important which ( data node or arbiter node )) 2) directly reboot two nodes - OK ( if nodes are not from the same replica ) 3) directly reboot three nodes - yes, this is the main problem and a questions - rebooted all three nodes from replica "B" ( not so possible, but who knows ... ) - all VMs with data on this replica was paused ( no data access ) - OK - all VMs running on replica "B" nodes lost ( started manually, later )( datas on other replicas ) - acceptable BUT - !!! all oVIrt domains went down !! - master domain is on replica "A" which lost only one member from three !!! so we are not expecting that all domain will go down, especially master with 2 live members. Results: - the whole cluster unreachable until at all domains up - depent of all nodes up !!! - all paused VMs started back - OK - rest of all VMs rebooted and runnig - OK Questions: 1) why all domains down if master domain ( on replica "A" ) has two runnig members ( 2 of 3 ) ?? 2) how to fix that colaps without waiting to all nodes up ? ( in worste case if node has HW error eg. ) ?? 3) which oVirt cluster policy can prevent that situation ?? ( if any ) regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] ovirt - unsupported GFS config ??
- Yes, database will run in VM , but with RMAN online backups. So VM takes over 1,4TB . This VM wil run 5-6x :o( - I have no special issues with 512MB shard size, looks OK but not tested for such databases yet ( Waiting for 3.7.10 gluster (dupl.healing fix)) regs. Pavel On 28.3.2016 08:30, Sahina Bose wrote: On 03/27/2016 10:24 PM, p...@email.cz wrote: What's the recommended shard size for databases ( espacially Oracle ) , I'm afraid that 512M is too large. I found that someones using about 16MB , but it generate a lot of files for healing if volumes splitted. (eg. for 500GB DB in worse case ) Would the database be running within the guest VM? Did you run into any specific issue with 512M shard size? What we have noticed is that with smaller shard sizes , like 4MB, the number of entries in the .shard directory is too high, and hence affects the performance when it comes to healing entries in directory. The 256M/512M shard size is a good balance between the number of entries created and data size to heal. Pa. On 27.3.2016 16:57, Sahina Bose wrote: Stripe is not supported. What you need to do instead is turn on sharding for the volume. So: gluster volume create 12HP12-S2R3A1P2 replica 3 arbiter 1 1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-1/GFS force gluster volume set 12HP12-S2R3A1P2 features.shard on gluster volume set 12HP12-S2R3A1P2 features.shard-block-size 512MB If you want to utilize the additional nodes as well, you can change this to a distributed replicate volume - instead of the volume creation in step above , use below gluster volume create 12HP12-S2R3A1P2 replica 3 arbiter 1 1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-1/GFS 2hp1:/STORAGES/P2/GFS 2hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-2/GFS force On 03/24/2016 07:49 PM, p...@email.cz wrote: Hello, I tried create stripe 2 replica 3 arbiter1 gluster volume for testing. So , glusterFS such type from commandline was successfull, but domain creation looks to be unsupported. with oVirt message "Error while executing action AddGlusterFsStorageDomain: Storage Domain target is unsupported". Cam U tell me if is it error or really unsuported ?? exam: gluster volume create 12HP12-S2R3A1P2 stripe 2 replica 3 arbiter 1 1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-1/GFS 2hp1:/STORAGES/P2/GFS 2hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-2/GFS force RHEL 7-2.1511 vdsm - vdsm-4.17.23-1.el7 gluster - glusterfs-3.7.9-1.el7 ovirt - 3.5.6.2-1 regs.Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] ovirt - unsupported GFS config ??
Hello, I tried create stripe 2 replica 3 arbiter1 gluster volume for testing. So , glusterFS such type from commandline was successfull, but domain creation looks to be unsupported. with oVirt message "Error while executing action AddGlusterFsStorageDomain: Storage Domain target is unsupported". Cam U tell me if is it error or really unsuported ?? exam: gluster volume create 12HP12-S2R3A1P2 stripe 2 replica 3 arbiter 1 1hp1:/STORAGES/P2/GFS 1hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-1/GFS 2hp1:/STORAGES/P2/GFS 2hp2:/STORAGES/P2/GFS kvmarbiter:/STORAGES/P2-2/GFS force RHEL 7-2.1511 vdsm - vdsm-4.17.23-1.el7 gluster - glusterfs-3.7.9-1.el7 ovirt - 3.5.6.2-1 regs.Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] engine.log messages in cycle
Hello, can anybody explain why is engine.log filled with following messages, even thought this is fresh installation ?? Especially START / FINISH rows in cycle, ... in cycle, in cycle - it takes a lot of space and is really needed ?? RHEL 7-2.1511 vdsm - vdsm-4.17.23-1.el7 gluster - glusterfs-3.7.9-1.el7 ovirt - 3.5.6.2-1 regs.Pavel 2016-03-24 13:39:10,758 INFO [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler_Worker-89) START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 2hp1, HostId = 45f76a0f-9616-420a-be1d-afbed2954562), log id: 3d6b27fd 2016-03-24 13:39:13,243 INFO [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler_Worker-89) FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@11d55805, log id: 3d6b27fd 2016-03-24 13:39:13,278 INFO [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler_Worker-89) START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = 2hp1, HostId = 45f76a0f-9616-420a-be1d-afbed2954562), log id: 3ae0e479 2016-03-24 13:39:13,349 INFO [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e value: GLUSTER , sharedLocks= ] 2016-03-24 13:39:13,444 INFO [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e value: GLUSTER , sharedLocks= ] 2016-03-24 13:39:13,801 INFO [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e value: GLUSTER , sharedLocks= ] 2016-03-24 13:39:14,646 INFO [org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler_Worker-43) Failed to acquire lock and wait lock EngineLock [exclusiveLocks= key: 0001-0001-0001-0001-022e value: GLUSTER , sharedLocks= ] 2016-03-24 13:39:15,630 INFO [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler_Worker-89) FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@cab5100f, log id: 3ae0e479 2016-03-24 13:39:15,656 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-43) START, GlusterVolumesListVDSCommand(HostName = 1hp2, HostId = 184ebfaa-51a9-43e4-a57b-9d4f03e85b47), log id: 5756f325 2016-03-24 13:39:16,105 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-43) FINISH, GlusterVolumesListVDSCommand, return: {e4121610-6128-4ecc-86d3-1429ab3b8356=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@edd4741e, d3d260cd-455f-42d6-9580-d88ae6df0519=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@84ea7412}, log id: 5756f325 2016-03-24 13:39:21,161 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-81) START, GlusterVolumesListVDSCommand(HostName = 1hp2, HostId = 184ebfaa-51a9-43e4-a57b-9d4f03e85b47), log id: 6e2f6c69 2016-03-24 13:39:21,667 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-81) FINISH, GlusterVolumesListVDSCommand, return: {e4121610-6128-4ecc-86d3-1429ab3b8356=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@2ad03bfa, d3d260cd-455f-42d6-9580-d88ae6df0519=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@6a99fa09}, log id: 6e2f6c69 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] delete storage definition
Hi Maor, it helped, great ! thx a lot , U spare my time . regs.Pa. On 24.3.2016 12:21, Maor Lipchuk wrote: On Thu, Mar 24, 2016 at 12:20 PM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: Hi folks, how can I delete the last storage definition from oVirt database if the last volume has been deleted from bricks commandline ( rm -rf < path to that volume > ) directly ? In oVirt DB exists this storage last record and blocking create new storage operation ( ovirt offering " delete datacenter", but this is not the right way for me, now ) regs. Pavel ___ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users Hi Pavel, What are your plans regarding that Data Center? In case you want to keep that Data Center to be used with other storage domains you can try to add a new storage domain without attaching it to any Data Center and try to re-initialize the Data Center with this new storage domain. Once that Data Center will be re-initialized you can try to remove the old Storage Domain (or force remove it you encounter any problem) Please let me know if this helps you or is there anything else that you were trying to do Regards, Maor ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] delete storage definition
Hi folks, how can I delete the last storage definition from oVirt database if the last volume has been deleted from bricks commandline ( rm -rf < path to that volume > ) directly ? In oVirt DB exists this storage last record and blocking create new storage operation ( ovirt offering " delete datacenter", but this is not the right way for me, now ) regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] question mark on VM ( DB status 8 )
I used that, but lock active in a few seconds again. And oVirt do not update any VM's status Pa. On 17.3.2016 10:26, Eli Mesika wrote: *From: *p...@email.cz <mailto:p...@email.cz> *To: *"users" <users@ovirt.org <mailto:users@ovirt.org>> *Sent: *Thursday, March 17, 2016 9:27:11 AM *Subject: *[ovirt-users] question mark on VM ( DB status 8 ) Hello, during backup VM hanged with question mark in ovirt and status 8 in DB, snapshot file ( for backup )is locked. How to clean snapshot locking a wake up this VM from "unknow" state ??? Try using the unlock_entity.sh utility (run with --help for usage) regs. pavel ___ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] oVirt - rescan new copied disks to storage
Hello, how can I put a copy of VM disks ( outside ovirt envir. copy ) to ovirt inventory ?? ( available for " attach disk" option visibility ) regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] oVirt - rescan new copied disks to storage
Hi Nir, thx for info. YES I mean import raw disk ( raw image , not raw device, not v2v utils ) such as " /rhev/data-center/mnt/glusterSD/localhost:DOMAIN/aaa/image//cc " file which represents the VM disk. regs. Pavel On 17.3.2016 16:03, Nir Soffer wrote: On Thu, Mar 17, 2016 at 4:37 PM, p...@email.cz <p...@email.cz> wrote: Hello, how can I put a copy of VM disks ( outside ovirt envir. copy ) to ovirt inventory ?? ( available for " attach disk" option visibility ) Do you mean how to import existing vm disk into ovirt? You can use v2v (in ovirt-3.6) to import vms directly from vmware, or ova created by vmware. If you need to import qcow or raw disk, it is not automated yet, we are working on it for 4.0. Maybe Shahar can add more info what available today and what are the workarounds to import disks in 3.6. In 4.0 you will be able to upload disks directly using http, see http://www.ovirt.org/develop/release-management/features/image-upload/ Nir ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Fwd: Re: question mark on VM ( DB status 8 )
URGENT Forwarded Message Subject:Re: [ovirt-users] question mark on VM ( DB status 8 ) Date: Thu, 17 Mar 2016 16:43:54 +0200 From: Nir Soffer <nsof...@redhat.com> To: p...@email.cz <p...@email.cz> Can you send this to the users list? This looks like virt issue, so it should be checked by the guys working on this pars of the code. Thanks, Nir On Thu, Mar 17, 2016 at 4:07 PM, p...@email.cz <p...@email.cz> wrote: Hi Nir, look at piece of logs which are repeated in cycle. The main issue happened about 3-5AM today ( 17.Mar) CSA_EBSDB_TEST2 - was shutted down from OS , but status was not updated in oVirt GUI ( changed manually in DB ( status 1 )) , still one other VM is in status "8" due snapshot locked file ( sf-sh-s07) . engine.log == repeately hours by hours ... continually 2016-03-17 14:38:21,146 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-20) START, GlusterVolumesListVDSCommand(HostName = 2kvm1, HostId = 4c3a2622-14d5-43c8-8e15-99cb66104b5a), log id: 5a34e053 2016-03-17 14:38:21,830 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-20) FINISH, GlusterVolumesListVDSCommand, return: {a5a8ccbc-edee-4e49-9e2a-4d2ee5767f76=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@240192c6, 18310aeb-639f-4b6d-9ef4-9ef560d6175c=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@753f6685, 4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@79a21b20, f410c6a9-9a51-42b3-89bb-c20ac72a0461=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@a4634e44, 62c89345-fd61-4b67-b8b4-69296eb7d217=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@fd990620, aa2d607d-3c6c-4f13-8205-aae09dcc9d35=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@57883869, b4356604-4404-428a-9da6-f1636115e2fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@3b458bc8, 9745551f-4696-4a6c-820a-619e359a61fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@80f225de, 25a5ec22-660e-42a0-aa00-45211d341738=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@ec4c19bd, 6060ff77-d552-4d94-97bf-5a32982e7d8a=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@947dc2e4, cbf142f8-a40b-4cf4-ad29-2243c81d30c1=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f773ab98}, log id: 5a34e053 2016-03-17 14:38:27,131 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-79) START, GlusterVolumesListVDSCommand(HostName = 2kvm1, HostId = 4c3a2622-14d5-43c8-8e15-99cb66104b5a), log id: 24e7703f 2016-03-17 14:38:27,801 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-79) FINISH, GlusterVolumesListVDSCommand, return: {a5a8ccbc-edee-4e49-9e2a-4d2ee5767f76=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72f0f4, 18310aeb-639f-4b6d-9ef4-9ef560d6175c=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@89bfd4dd, 4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f6cb25b, f410c6a9-9a51-42b3-89bb-c20ac72a0461=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f4bb56bf, 62c89345-fd61-4b67-b8b4-69296eb7d217=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@e0121f88, aa2d607d-3c6c-4f13-8205-aae09dcc9d35=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@435fc00f, b4356604-4404-428a-9da6-f1636115e2fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7b23bf23, 9745551f-4696-4a6c-820a-619e359a61fd=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@1f8e886, 25a5ec22-660e-42a0-aa00-45211d341738=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@1fbbe1c1, 6060ff77-d552-4d94-97bf-5a32982e7d8a=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@87c991cd, cbf142f8-a40b-4cf4-ad29-2243c81d30c1=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@2fc8ef3e}, log id: 24e7703f 2016-03-17 14:38:33,097 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-15) START, GlusterVolumesListVDSCommand(HostName = 2kvm1, HostId = 4c3a2622-14d5-43c8-8e15-99cb66104b5a), log id: 2e987652 2016-03-17 14:38:33,809 INFO [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] (DefaultQuartzScheduler_Worker-15) FINISH, GlusterVolumesListVDSCommand, return: {a5a8ccbc-edee-4e49-9e2a-4d2ee5767f76=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@22f57
Re: [ovirt-users] [Gluster-users] open error -13 = sanlock
OK, will extend replica 2 to replica 3 ( arbiter ) ASAP . If is deleted "untouching" ids file on brick , healing of this file doesn't work . regs.Pa. On 3.3.2016 12:19, Nir Soffer wrote: On Thu, Mar 3, 2016 at 11:23 AM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: This is replica 2, only , with following settings Replica 2 is not supported. Even if you "fix" this now, you will have the same issue soon. Options Reconfigured: performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: fixed cluster.server-quorum-type: none storage.owner-uid: 36 storage.owner-gid: 36 cluster.quorum-count: 1 cluster.self-heal-daemon: enable If I'll create "ids" file manually ( eg. " sanlock direct init -s 3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0 " ) on both bricks, vdsm is writing only to half of them ( that with 2 links = correct ) "ids" file has correct permittions, owner, size on both bricks. brick 1: -rw-rw 1 vdsm kvm 1048576 2. bře 18.56 /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids - not updated brick 2: -rw-rw 2 vdsm kvm 1048576 3. bře 10.16 /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids - is continually updated What happens when I'll restart vdsm ? Will oVirt storages go to "disable " state ??? = disconnect VMs storages ? Nothing will happen, the vms will continue to run normally. On block storage, stopping vdsm will prevent automatic extending of vm disks when the disk become too full, but on file based storage (like gluster) there is no issue. regs.Pa. On 3.3.2016 02:02, Ravishankar N wrote: On 03/03/2016 12:43 AM, Nir Soffer wrote: PS: # find /STORAGES -samefile /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -print /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids = missing "shadowfile" in " .gluster " dir. How can I fix it ?? - online ! Ravi? Is this the case in all 3 bricks of the replica? BTW, you can just stat the file on the brick and see the link count (it must be 2) instead of running the more expensive find command. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] [Gluster-users] open error -13 = sanlock
This is replica 2, only , with following settings Options Reconfigured: performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: fixed cluster.server-quorum-type: none storage.owner-uid: 36 storage.owner-gid: 36 cluster.quorum-count: 1 cluster.self-heal-daemon: enable If I'll create "ids" file manually ( eg. " sanlock direct init -s 3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0 " ) on both bricks, vdsm is writing only to half of them ( that with 2 links = correct ) "ids" file has correct permittions, owner, size on both bricks. brick 1: -rw-rw 1 vdsm kvm 1048576 2. bře 18.56 /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids - not updated brick 2: -rw-rw 2 vdsm kvm 1048576 3. bře 10.16 /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids - is continually updated What happens when I'll restart vdsm ? Will oVirt storages go to "disable " state ??? = disconnect VMs storages ? regs.Pa. On 3.3.2016 02:02, Ravishankar N wrote: On 03/03/2016 12:43 AM, Nir Soffer wrote: PS: # find /STORAGES -samefile /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -print /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids = missing "shadowfile" in " .gluster " dir. How can I fix it ?? - online ! Ravi? Is this the case in all 3 bricks of the replica? BTW, you can just stat the file on the brick and see the link count (it must be 2) instead of running the more expensive find command. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] [Gluster-users] open error -13 = sanlock
UPDATE: all "ids" file have permittion fixed to 660 now # find /STORAGES -name ids -exec ls -l {} \; -rw-rw 2 vdsm kvm 0 24. úno 07.41 /STORAGES/g1r5p1/GFS/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md/ids -rw-rw 2 vdsm kvm 0 24. úno 07.43 /STORAGES/g1r5p2/GFS/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids -rw-rw 2 vdsm kvm 0 24. úno 07.43 /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -rw-rw 2 vdsm kvm 0 24. úno 07.44 /STORAGES/g1r5p4/GFS/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids -rw-rw 2 vdsm kvm 1048576 24. úno 13.03 /STORAGES/g1r5p5/GFS/3b24d023-fd35-4666-af2f-f5e1d19531ad/dom_md/ids -rw-rw 2 vdsm kvm 1048576 2. bře 17.47 /STORAGES/g2r5p1/GFS/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids SPM is and was running continually ... I tried to update "ids" file - ONLINE ( offline not possible yet ) # sanlock direct init -s 3c34ad63-6c66-4e23-ab46-084f3d70b147:0:/STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids:0 # find /STORAGES -name ids -exec ls -l {} \; | grep p3 -rw-rw 1 vdsm kvm 1048576 2. bře 18.32 /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids The storage ids file has correct permittions, size, owners , but is not checking by sanlock = the same access time What's wrong ?? regs. Pa. PS: # find /STORAGES -samefile /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids -print /STORAGES/g1r5p3/GFS/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md/ids = missing "shadowfile" in " .gluster " dir. How can I fix it ?? - online ! On 2.3.2016 08:16, Ravishankar N wrote: On 03/02/2016 12:02 PM, Sahina Bose wrote: On 03/02/2016 03:45 AM, Nir Soffer wrote: On Tue, Mar 1, 2016 at 10:51 PM, p...@email.cz <p...@email.cz> wrote: > > HI, > requested output: > > # ls -lh /rhev/data-center/mnt/glusterSD/localhost:*/*/dom_md > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md: > total 2,1M > -rw-rw 1 vdsm kvm 1,0M 1. bře 21.28 ids <-- good > -rw-rw 1 vdsm kvm 16M 7. lis 22.16 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 22.17 leases > -rw-r--r-- 1 vdsm kvm 335 7. lis 22.17 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 22.16 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.41 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 00.14 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 03.56 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 03.56 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.14 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 00.15 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 22.14 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 22.14 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.15 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P3/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 23. úno 22.51 inbox > -rw-rw 1 vdsm kvm 2,0M 23. úno 23.12 leases > -rw-r--r-- 1 vdsm kvm 998 25. úno 00.35 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.16 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.44 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 00.17 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 00.18 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 00.18 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.17 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P1/42d710a9-b844-43dc-be41-77002d1cd553/dom_md: > total 1,1M > -rw-rw-r-- 1 vdsm kvm0 24. úno 07.32 ids <-- bad (other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 22.18 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 22.18 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 22.18 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 22.18 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md: > total 3,0M > -rw-rw-r-- 1 vdsm kvm 1,0M 1. bře 21.28 ids <-- bad (other can read) > -rw-rw 1 vdsm kvm 16M 25. úno 00.42 inbox > -rw-rw 1 vdsm kvm 2,0M 25. úno 00.44 leases > -rw-r--r-- 1 vdsm kvm 997 24. úno 02.46 metadata > -rw-rw 1 vdsm kvm 16M 25. úno 00.44 outbox > > /rhev/data-center/mnt/glusterS
[ovirt-users] SUCCESS / error
Hi, next explanation VDSM log give me following message Is this live checking for storage availability ? If "SUCCESS" then why " " ??? regs. Pavel Thread-233::DEBUG::2016-03-02 17:31:55,275::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: = '0+1 records in\n0+1 records out\n346 bytes (346 B) copied, 0.00022692 s, 1.5 MB/s\n'; = 0 Thread-299::DEBUG::2016-03-02 17:31:57,870::fileSD::262::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd if=/rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97/dom_md/metadata iflag=direct of=/dev/null bs=4096 count=1 (cwd None) Thread-299::DEBUG::2016-03-02 17:31:57,886::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: = '0+1 records in\n0+1 records out\n734 bytes (734 B) copied, 0.000391403 s, 1.9 MB/s\n'; = 0 Thread-209::DEBUG::2016-03-02 17:31:58,301::fileSD::262::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd if=/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/metadata iflag=direct of=/dev/null bs=4096 count=1 (cwd None) Thread-209::DEBUG::2016-03-02 17:31:58,318::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: = '0+1 records in\n0+1 records out\n335 bytes (335 B) copied, 0.00048095 s, 697 kB/s\n'; = 0 Thread-198::DEBUG::2016-03-02 17:31:58,885::fileSD::262::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd if=/rhev/data-center/mnt/installshare:_DATA3/7f334c9d-ac01-4b2a-9f29-af1d2f24c3e0/dom_md/metadata iflag=direct of=/dev/null bs=4096 count=1 (cwd None) dist = RHEL - 7 - 2.1511 kernel = 3.10.0 - 327.10.1.el7.x86_64 KVM = 2.3.0 - 29.1.el7 libvirt = libvirt-1.2.17-13.el7_2.3 vdsm = vdsm-4.16.30-0.el7 GlusterFS = glusterfs-3.7.8-1.el7 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] [Gluster-users] open error -13 = sanlock
Yes we have had "ids" split brains + some other VM's files Split brains was fixed by healing with preffered ( source ) brick. eg: " # gluster volume heal 1KVM12-P1 split-brain source-brick 16.0.0.161:/STORAGES/g1r5p1/GFS " Pavel Okay, so what I understand from the output above is you have different gluster volumes mounted and some of them have incorrect permissions for the 'ids' file. The way to fix it is to do it from the mount like Nir said. Why did you delete the file from the .glusterfs in the brick(s)? Was there a gfid split brain? -Ravi ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] [Gluster-users] open error -13 = sanlock
Hi guys, thx a lot for your support ...at first. Because we had been under huge time pressure, we found "google workaround" which delete both files . It helped, probabbly at first steps of recover . eg: " # find /STORAGES/g1r5p5/GFS/ -samefile /STORAGES/g1r5p5/GFS/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids -print -delete " --> Well at first I'll fix permittions from mount points to 660 . If "ids" file will be writeable , can't became gluster colaps ?? regs.Pavel On 2.3.2016 08:16, Ravishankar N wrote: On 03/02/2016 12:02 PM, Sahina Bose wrote: On 03/02/2016 03:45 AM, Nir Soffer wrote: On Tue, Mar 1, 2016 at 10:51 PM, p...@email.cz <p...@email.cz> wrote: > > HI, > requested output: > > # ls -lh /rhev/data-center/mnt/glusterSD/localhost:*/*/dom_md > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md: > total 2,1M > -rw-rw 1 vdsm kvm 1,0M 1. bře 21.28 ids <-- good > -rw-rw 1 vdsm kvm 16M 7. lis 22.16 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 22.17 leases > -rw-r--r-- 1 vdsm kvm 335 7. lis 22.17 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 22.16 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.41 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 00.14 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 03.56 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 03.56 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.14 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 00.15 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 22.14 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 22.14 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.15 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P3/3c34ad63-6c66-4e23-ab46-084f3d70b147/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.43 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 23. úno 22.51 inbox > -rw-rw 1 vdsm kvm 2,0M 23. úno 23.12 leases > -rw-r--r-- 1 vdsm kvm 998 25. úno 00.35 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.16 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md: > total 1,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.44 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 00.17 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 00.18 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 00.18 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 00.17 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P1/42d710a9-b844-43dc-be41-77002d1cd553/dom_md: > total 1,1M > -rw-rw-r-- 1 vdsm kvm0 24. úno 07.32 ids <-- bad (other can read) > -rw-rw 1 vdsm kvm 16M 7. lis 22.18 inbox > -rw-rw 1 vdsm kvm 2,0M 7. lis 22.18 leases > -rw-r--r-- 1 vdsm kvm 333 7. lis 22.18 metadata > -rw-rw 1 vdsm kvm 16M 7. lis 22.18 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md: > total 3,0M > -rw-rw-r-- 1 vdsm kvm 1,0M 1. bře 21.28 ids <-- bad (other can read) > -rw-rw 1 vdsm kvm 16M 25. úno 00.42 inbox > -rw-rw 1 vdsm kvm 2,0M 25. úno 00.44 leases > -rw-r--r-- 1 vdsm kvm 997 24. úno 02.46 metadata > -rw-rw 1 vdsm kvm 16M 25. úno 00.44 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P3/ef010d08-aed1-41c4-ba9a-e6d9bdecb4b4/dom_md: > total 2,1M > -rw-r--r-- 1 vdsm kvm0 24. úno 07.34 ids <-- bad (sanlock cannot write, other can read) > -rw-rw 1 vdsm kvm 16M 23. úno 22.35 inbox > -rw-rw 1 vdsm kvm 2,0M 23. úno 22.38 leases > -rw-r--r-- 1 vdsm kvm 1,1K 24. úno 19.07 metadata > -rw-rw 1 vdsm kvm 16M 23. úno 22.27 outbox > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97/dom_md: > total 3,0M > -rw-rw-r-- 1 vdsm kvm 1,0M 1. bře 21.28 ids <-- bad (other can read) > -rw-rw-r-- 1 vdsm kvm 16M 6. lis 23.50 inbox <-- bad (other can read) > -rw-rw-r-- 1 vdsm kvm 2,0M 6. lis 23.51 leases <-- bad (other can read) > -rw-rw-r-- 1 vdsm kvm 734 7. lis 02.13 metadata<-- bad (group can write, other can read) > -rw-rw-r-- 1 vdsm kvm 16M 6. lis 16.55 outbox <-- bad (other can read) > > /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P5/1ca56b45-701e-4c22-9f59-3aebea4d8477/dom_md: > total 1,1
[ovirt-users] open error -13 = sanlock
Hello, can anybody explain this error no.13 ( open file ) in sanlock.log . The size of "ids" file is zero (0) 2016-02-28 03:25:46+0100 269626 [1951]: open error -13 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids 2016-02-28 03:25:46+0100 269626 [1951]: s187985 open_disk /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids error -13 2016-02-28 03:25:56+0100 269636 [11304]: s187992 lockspace 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 If the main problem is about zero file size, can I regenerate this file online securely , with no VM dependence dist = RHEL - 7 - 2.1511 kernel = 3.10.0 - 327.10.1.el7.x86_64 KVM = 2.3.0 - 29.1.el7 libvirt = libvirt-1.2.17-13.el7_2.3 vdsm = vdsm-4.16.30-0.el7 GlusterFS = glusterfs-3.7.8-1.el7 regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Fwd: Re: ovirt - can't attach master domain II
used replica2 with volume option Volume Name: 2KVM12-P2 Type: Replicate Volume ID: 9745551f-4696-4a6c-820a-619e359a61fd Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 16.0.0.164:/STORAGES/g1r5p2/GFS Brick2: 16.0.0.163:/STORAGES/g1r5p2/GFS Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 performance.io-cache: off performance.read-ahead: off network.remote-dio: enable cluster.eager-lock: enable performance.stat-prefetch: off performance.quick-read: off cluster.quorum-count: 1 cluster.server-quorum-type: none cluster.quorum-type: fixed was runnig over year with no problems ( reboots, ..etc... ) On 24.2.2016 12:34, Ravishankar N wrote: On 02/24/2016 04:48 PM, p...@email.cz wrote: prereq: 2KVM12-P2 = master domain - YES - I'm using gluster.fuse NFS localhost:/2KVM12-P2 on /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2 type fuse.glusterfs (rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072) - Healing == # gluster volume heal 2KVM12-P2 info Brick 16.0.0.164:/STORAGES/g1r5p2/GFS Number of entries: 0 Brick 16.0.0.163:/STORAGES/g1r5p2/GFS Number of entries: 0 # while true; do for vol in `gluster volume list`; do gluster volume heal $vol info | sort | grep "Number of entries" | awk -F: '{tot+=$2} END { printf("Heal entries for '"$vol"': %d\n", $tot);}'; done; sleep 120; echo -e "\n==\n"; done Heal entries for 1KVM12-BCK: 1 Heal entries for 1KVM12-P1: 1 Heal entries for 1KVM12-P2: 0 Heal entries for 1KVM12-P3: 0 Heal entries for 1KVM12-P4: 0 Heal entries for 1KVM12-P5: 0 Heal entries for 2KVM12-P1: 1 Heal entries for 2KVM12-P2: 0 Heal entries for 2KVM12-P3: 0 Heal entries for 2KVM12-P5: 0 Heal entries for 2KVM12_P4: 1 # gluster volume heal 1KVM12-BCK info split-brain Brick 16.0.0.161:/STORAGES/g2r5p1/GFS /0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids Number of entries in split-brain: 1 Brick 16.0.0.162:/STORAGES/g2r5p1/GFS /0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids Number of entries in split-brain: 1 # gluster volume heal 1KVM12-P1 info split-brain Brick 16.0.0.161:/STORAGES/g1r5p1/GFS /__DIRECT_IO_TEST__ Number of entries in split-brain: 1 Brick 16.0.0.162:/STORAGES/g1r5p1/GFS /__DIRECT_IO_TEST__ Number of entries in split-brain: 1 etc.. YES - in split brain , but NOT master domain ( will solve later, after master - if possible ) I'm not sure if it is related, but you could try to resolve the split-brain first and see if it helps. Also, I see that you are using replica-2. It is recommended to use replica-3 or arbiter volumes to avoid split-brains. -Ravi --- vdsm.log = Thread-461::DEBUG::2016-02-24 11:12:45,328::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: = '0+1 records in\n0+1 records out\n333 bytes (333 B) copied, 0.000724379 s, 460 kB/s\n'; = 0 Thread-461::INFO::2016-02-24 11:12:45,331::clusterlock::219::Storage.SANLock::(acquireHostId) Acquiring host id for domain 88adbd49-62d6-45b1-9992-b04464a04112 (id: 3) Thread-461::DEBUG::2016-02-24 11:12:45,331::clusterlock::237::Storage.SANLock::(acquireHostId) Host id for domain 88adbd49-62d6-45b1-9992-b04464a04112 successfully acquired (id: 3) Thread-33186::DEBUG::2016-02-24 11:12:46,067::__init__::481::jsonrpc.JsonRpcServer::(_serveRequest) Calling 'GlusterVolume.list' in bridge with {} Thread-33186::DEBUG::2016-02-24 11:12:46,204::__init__::514::jsonrpc.JsonRpcServer::(_serveRequest) Return 'GlusterVolume.list' in bridge with {'volumes': {'2KVM12-P5': {'transportType': ['TCP'], 'uuid': '4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d', 'bricks': ['16.0.0.164:/STORAGES/g1r5p5/GFS', '16.0.0.163:/STORAGES/g1r5p5/GFS'], 'volumeName': '2KVM12-P5', 'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 'distCount': '2', 'volumeStatus': 'ONLINE', 'stripeCount': '1', 'bricksInfo': [{'name': '16.0.0.164:/STORAGES/g1r5p5/GFS', 'hostUuid': '06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}, {'name': '16.0.0.163:/STORAGES/g1r5p5/GFS', 'hostUuid': '6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}], 'options': {'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 'enable', 'performance.stat-prefetch': 'off', 'cluster.quorum-type': 'fixed', 'performance.quick-read': 'off', 'network.remote-dio': 'enable', 'cluster.quorum-count': '1', 'performance.io-cache': 'off', 'storage.owner-uid': '36', 'performance.read-ahead': 'off', 'storage.owner-gid': '36'}}, '2KVM12_P4': {'transportType': ['TCP'], 'uuid': '18310aeb-639f-4b6d-9ef4-9ef560d6175c', 'bricks': ['16.0.0.163:/STORAGES/g1r5p4/GFS', '16.0.0.164:/STORAGES/g1r5p4/GFS'], 'volumeName': '2KVM12_P4', 'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 'distCount': '2', 'vo
[ovirt-users] Fwd: Re: ovirt - can't attach master domain III
hi, after a lot of test will get 2016-02-24 11:38:05+0100 7406 [25824]: cmd_add_lockspace 3,10 ff71b47b-0f72-4528-9bfe-c3da888e47f0:4:/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids:0 flags 1 timeout 0 2016-02-24 11:38:05+0100 7406 [25824]: s2256 lockspace ff71b47b-0f72-4528-9bfe-c3da888e47f0:4:/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids:0 2016-02-24 11:38:05+0100 7406 [25824]: cmd_add_lockspace 3,10 async done 0 2016-02-24 11:38:05+0100 7406 [26186]: open error -2 /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids 2016-02-24 11:38:05+0100 7406 [26186]: s2256 open_disk /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/ff71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ids error -2 what's wrong ?? thx. Pa. On 24.2.2016 08:14, Nir Soffer wrote: On Wed, Feb 24, 2016 at 8:53 AM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: Hi, it seems that sanlock daemon has problem with reading empty "ids" file . How can I regenarate this "ids" file to get 2k rows of datas ?? It's the base problem to get up "master domain" following "datacenter" You should understand why the ids files is empty and fix the root cause. To recover your ids files, you can follow the instructions here: http://lists.ovirt.org/pipermail/users/2016-February/038046.html Nir regs. Pa. ___ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Fwd: Re: ovirt - can't attach master domain II
gt; state preparing Thread-35634::INFO::2016-02-24 11:18:20,860::logUtils::44::dispatcher::(wrapper) Run and protect: getSpmStatus(spUUID=u'0002-0002-0002-0002-021e', options=None) Thread-35634::INFO::2016-02-24 11:18:20,867::logUtils::47::dispatcher::(wrapper) Run and protect: getSpmStatus, Return response: {'spm_st': {'spmId': -1, 'spmStatus': 'Free', 'spmLver': -1}} Thread-35634::DEBUG::2016-02-24 11:18:20,867::task::1191::Storage.TaskManager.Task::(prepare) Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::finished: {'spm_st': {'spmId': -1, 'spmStatus': 'Free', 'spmLver': -1}} Thread-35634::DEBUG::2016-02-24 11:18:20,867::task::595::Storage.TaskManager.Task::(_updateState) Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::moving from state preparing -> state finished Thread-35634::DEBUG::2016-02-24 11:18:20,867::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-35634::DEBUG::2016-02-24 11:18:20,867::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-35634::DEBUG::2016-02-24 11:18:20,867::task::993::Storage.TaskManager.Task::(_decref) Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::ref 0 aborting False --- these blocks generated in cycle for each domain Any IDEA ?? regs. Pa. On 24.2.2016 10:54, Ravishankar N wrote: Hi, On 02/24/2016 06:43 AM, p...@email.cz wrote: Hi, I found the main ( maybe ) problem with IO error ( -5 ) for "ids" file access This file is not accessable via NFS, locally yes How is NFS coming into the picture? Are you not using gluster fuse mount? . How can I fix it ?? Can you run `gluster volume heal volname info` and `gluster volume heal volname info split-brain` to see if the "ids" file is in split-brain? A file in split-brain returns EIO when accessed from the mount. Regards, Ravi regs. Pavel # sanlock client log_dump 0 flags 1 timeout 0 2016-02-24 02:01:10+0100 3828 [12111]: s1316 lockspace 88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0 2016-02-24 02:01:10+0100 3828 [12111]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire begin 88adbd49-62d6-45b1-9992-b04464a04112:1 2016-02-24 02:01:10+0100 3828 [19556]: 88adbd49 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:10+0100 3828 [19556]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids 2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire leader_read1 error -5 2016-02-24 02:01:11+0100 3829 [12111]: s1316 add_lockspace fail result -5 2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:12+0100 3831 [12116]: s1317 lockspace 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire begin 7f52b697-c199-4f58-89aa-102d44327124:1 2016-02-24 02:01:12+0100 3831 [19562]: 7f52b697 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:12+0100 3831 [19562]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids 2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire leader_read1 error -5 2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:13+0100 3831 [1321]: s1318 lockspace 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire begin 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1 2016-02-24 02:01:13+0100 3831 [19564]: 0fcad888 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458201000 result -5:0 match res 2016-02-24 02:01:13+0100 3831 [19564]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids 2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire leader_read1 error -5 2016-02-24 02:01:13+0100 3832 [12116]: s1317 add_lockspace fail result -5 2016-02-24 02:01:14+0100 3832 [1321]: s1318 add_lockspace fail result -5
[ovirt-users] Fwd: Re: ovirt - can't attach master domain III
Hi Nir, it doesn't running, or failed by me 1) no trafic od any storage in gluster two ways A: - will stop ( maintenance not allowed )master domain ( 2KVM12-P2 ) from GUI - try to mount locally to one node # mount -t glusterfs localhost:/2KVM12-P2 /mnt ==> error -19 B: go to /rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2/f71b47b-0f72-4528-9bfe-c3da888e47f0/dom_md/ - remove empty ids file - sanlock direct init -s :0:ids:0 - from manual - restart sanlockd 2) no way was successfull regs.Pa. On 24.2.2016 08:14, Nir Soffer wrote: On Wed, Feb 24, 2016 at 8:53 AM, p...@email.cz <mailto:p...@email.cz> <p...@email.cz <mailto:p...@email.cz>> wrote: Hi, it seems that sanlock daemon has problem with reading empty "ids" file . How can I regenarate this "ids" file to get 2k rows of datas ?? It's the base problem to get up "master domain" following "datacenter" You should understand why the ids files is empty and fix the root cause. To recover your ids files, you can follow the instructions here: http://lists.ovirt.org/pipermail/users/2016-February/038046.html Nir regs. Pa. ___ Users mailing list Users@ovirt.org <mailto:Users@ovirt.org> http://lists.ovirt.org/mailman/listinfo/users ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] ovirt - can't attach master domain II
reparing Thread-35634::INFO::2016-02-24 11:18:20,860::logUtils::44::dispatcher::(wrapper) Run and protect: getSpmStatus(spUUID=u'0002-0002-0002-0002-021e', options=None) Thread-35634::INFO::2016-02-24 11:18:20,867::logUtils::47::dispatcher::(wrapper) Run and protect: getSpmStatus, Return response: {'spm_st': {'spmId': -1, 'spmStatus': 'Free', 'spmLver': -1}} Thread-35634::DEBUG::2016-02-24 11:18:20,867::task::1191::Storage.TaskManager.Task::(prepare) Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::finished: {'spm_st': {'spmId': -1, 'spmStatus': 'Free', 'spmLver': -1}} Thread-35634::DEBUG::2016-02-24 11:18:20,867::task::595::Storage.TaskManager.Task::(_updateState) Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::moving from state preparing -> state finished Thread-35634::DEBUG::2016-02-24 11:18:20,867::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-35634::DEBUG::2016-02-24 11:18:20,867::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-35634::DEBUG::2016-02-24 11:18:20,867::task::993::Storage.TaskManager.Task::(_decref) Task=`e887fd8b-6961-40f1-b3a0-917ffbea25c0`::ref 0 aborting False --- these blocks generated in cycle for each domain Any IDEA ?? regs. Pa. On 24.2.2016 10:54, Ravishankar N wrote: Hi, On 02/24/2016 06:43 AM, p...@email.cz wrote: Hi, I found the main ( maybe ) problem with IO error ( -5 ) for "ids" file access This file is not accessable via NFS, locally yes How is NFS coming into the picture? Are you not using gluster fuse mount? . How can I fix it ?? Can you run `gluster volume heal volname info` and `gluster volume heal volname info split-brain` to see if the "ids" file is in split-brain? A file in split-brain returns EIO when accessed from the mount. Regards, Ravi regs. Pavel # sanlock client log_dump 0 flags 1 timeout 0 2016-02-24 02:01:10+0100 3828 [12111]: s1316 lockspace 88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0 2016-02-24 02:01:10+0100 3828 [12111]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire begin 88adbd49-62d6-45b1-9992-b04464a04112:1 2016-02-24 02:01:10+0100 3828 [19556]: 88adbd49 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:10+0100 3828 [19556]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids 2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire leader_read1 error -5 2016-02-24 02:01:11+0100 3829 [12111]: s1316 add_lockspace fail result -5 2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:12+0100 3831 [12116]: s1317 lockspace 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire begin 7f52b697-c199-4f58-89aa-102d44327124:1 2016-02-24 02:01:12+0100 3831 [19562]: 7f52b697 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:12+0100 3831 [19562]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids 2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire leader_read1 error -5 2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:13+0100 3831 [1321]: s1318 lockspace 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire begin 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1 2016-02-24 02:01:13+0100 3831 [19564]: 0fcad888 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458201000 result -5:0 match res 2016-02-24 02:01:13+0100 3831 [19564]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids 2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire leader_read1 error -5 2016-02-24 02:01:13+0100 3832 [12116]: s1317 add_lockspace fail result -5 2016-02-24 02:01:14+0100 3832 [1321]: s1318 add_lockspace fail result -5 2016-02-24 0
[ovirt-users] ovirt - can't attach master domain III
Hi, it seems that sanlock daemon has problem with reading empty "ids" file . How can I regenarate this "ids" file to get 2k rows of datas ?? It's the base problem to get up "master domain" following "datacenter" regs. Pa. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] ovirt - can't attach master domain II
Hi, I found the main ( maybe ) problem with IO error ( -5 ) for "ids" file access This file is not accessable via NFS, locally yes. How can I fix it ?? regs. Pavel # sanlock client log_dump 0 flags 1 timeout 0 2016-02-24 02:01:10+0100 3828 [12111]: s1316 lockspace 88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0 2016-02-24 02:01:10+0100 3828 [12111]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire begin 88adbd49-62d6-45b1-9992-b04464a04112:1 2016-02-24 02:01:10+0100 3828 [19556]: 88adbd49 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:10+0100 3828 [19556]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids 2016-02-24 02:01:10+0100 3828 [19556]: s1316 delta_acquire leader_read1 error -5 2016-02-24 02:01:11+0100 3829 [12111]: s1316 add_lockspace fail result -5 2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:12+0100 3831 [12116]: s1317 lockspace 7f52b697-c199-4f58-89aa-102d44327124:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids:0 2016-02-24 02:01:12+0100 3831 [12116]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire begin 7f52b697-c199-4f58-89aa-102d44327124:1 2016-02-24 02:01:12+0100 3831 [19562]: 7f52b697 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:12+0100 3831 [19562]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P4/7f52b697-c199-4f58-89aa-102d44327124/dom_md/ids 2016-02-24 02:01:12+0100 3831 [19562]: s1317 delta_acquire leader_read1 error -5 2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:13+0100 3831 [1321]: s1318 lockspace 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids:0 2016-02-24 02:01:13+0100 3831 [1321]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire begin 0fcad888-d573-47be-bef3-0bc0b7a99fb7:1 2016-02-24 02:01:13+0100 3831 [19564]: 0fcad888 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458201000 result -5:0 match res 2016-02-24 02:01:13+0100 3831 [19564]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-BCK/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids 2016-02-24 02:01:13+0100 3831 [19564]: s1318 delta_acquire leader_read1 error -5 2016-02-24 02:01:13+0100 3832 [12116]: s1317 add_lockspace fail result -5 2016-02-24 02:01:14+0100 3832 [1321]: s1318 add_lockspace fail result -5 2016-02-24 02:01:19+0100 3838 [12106]: cmd_add_lockspace 4,15 3da46e07-d1ea-4f10-9250-6cbbb7b94d80:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P5/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:19+0100 3838 [12106]: s1319 lockspace 3da46e07-d1ea-4f10-9250-6cbbb7b94d80:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P5/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids:0 2016-02-24 02:01:19+0100 3838 [12106]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:19+0100 3838 [19638]: s1319 delta_acquire begin 3da46e07-d1ea-4f10-9250-6cbbb7b94d80:1 2016-02-24 02:01:19+0100 3838 [19638]: 3da46e07 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000 result -5:0 match res 2016-02-24 02:01:19+0100 3838 [19638]: read_sectors delta_leader offset 0 rv -5 /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P5/3da46e07-d1ea-4f10-9250-6cbbb7b94d80/dom_md/ids 2016-02-24 02:01:19+0100 3838 [19638]: s1319 delta_acquire leader_read1 error -5 2016-02-24 02:01:20+0100 3839 [12106]: s1319 add_lockspace fail result -5 2016-02-24 02:01:20+0100 3839 [1320]: cmd_add_lockspace 4,15 88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0 flags 1 timeout 0 2016-02-24 02:01:20+0100 3839 [1320]: s1320 lockspace 88adbd49-62d6-45b1-9992-b04464a04112:1:/rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P2/88adbd49-62d6-45b1-9992-b04464a04112/dom_md/ids:0 2016-02-24 02:01:20+0100 3839 [1320]: cmd_add_lockspace 4,15 async done 0 2016-02-24 02:01:20+0100 3839 [19658]: s1320 delta_acquire begin 88adbd49-62d6-45b1-9992-b04464a04112:1 2016-02-24 02:01:20+0100 3839 [19658]: 88adbd49 aio collect 0 0x7fe4580008c0:0x7fe4580008d0:0x7fe458101000
[ovirt-users] ovirt - can't attach master domain
Hi, due cleaning envir. I restarted all nodes and ovirt mgm server too, Now I can't get master domain ON. Tried to move to diff. storage with same result. All storages are OFF and DataCenter OFF too. All NFS mount points are connected and writeable via ovirt mountpoint. Any idea how to fix thi - quicky Rhel 7.1 / KVM 2.1.2 / libvirt 1.2.8-16 / vdsm 4.16.27 / gluster 3.7.5.1 / ovirt 3.5.2 VDSM log Thread-22688::DEBUG::2016-02-24 01:32:34,852::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-22688::DEBUG::2016-02-24 01:32:34,852::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-22688::DEBUG::2016-02-24 01:32:34,852::task::993::Storage.TaskManager.Task::(_decref) Task=`492b6d3f-47b7-47fe-a713-6e47d5f2d3c1`::ref 0 aborting False Thread-22688::DEBUG::2016-02-24 01:32:34,855::stompReactor::162::yajsonrpc.StompServer::(send) Sending response Thread-22697::DEBUG::2016-02-24 01:32:34,967::stompReactor::162::yajsonrpc.StompServer::(send) Sending response Thread-200::DEBUG::2016-02-24 01:32:35,001::fileSD::153::Storage.StorageDomain::(__init__) Reading domain in path /rhev/data-center/mnt/glusterSD/localhost:_2KVM12__P4/300e9ac8-3c2f-4703-9bb1-1df2130c7c97 Thread-200::ERROR::2016-02-24 01:32:35,007::domainMonitor::256::Storage.DomainMonitorThread::(_monitorDomain) Error while collecting domain 300e9ac8-3c2f-4703-9bb1-1df2130c7c97 monitoring information Traceback (most recent call last): File "/usr/share/vdsm/storage/domainMonitor.py", line 221, in _monitorDomain self.domain = sdCache.produce(self.sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 98, in produce domain.getRealDomain() File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain return self._cache._realProduce(self._sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce domain = self._findDomain(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain dom = findMethod(sdUUID) File "/usr/share/vdsm/storage/glusterSD.py", line 32, in findDomain return GlusterStorageDomain(GlusterStorageDomain.findDomainPath(sdUUID)) File "/usr/share/vdsm/storage/fileSD.py", line 160, in __init__ validateFileSystemFeatures(sdUUID, self.mountpoint) File "/usr/share/vdsm/storage/fileSD.py", line 89, in validateFileSystemFeatures oop.getProcessPool(sdUUID).directTouch(testFilePath) File "/usr/share/vdsm/storage/outOfProcess.py", line 351, in directTouch ioproc.touch(path, flags, mode) File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 507, in touch self.timeout) File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 391, in _sendCommand raise OSError(errcode, errstr) OSError: [Errno 5] Input/output error Thread-130::DEBUG::2016-02-24 01:32:35,013::fileSD::153::Storage.StorageDomain::(__init__) Reading domain in path /rhev/data-center/mnt/glusterSD/localhost:_1KVM12-P1/553d9b92-e4a0-4042-a579-4cabeb55ded4 Thread-130::ERROR::2016-02-24 01:32:35,017::domainMonitor::256::Storage.DomainMonitorThread::(_monitorDomain) Error while collecting domain 553d9b92-e4a0-4042-a579-4cabeb55ded4 monitoring information Traceback (most recent call last): File "/usr/share/vdsm/storage/domainMonitor.py", line 221, in _monitorDomain self.domain = sdCache.produce(self.sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 98, in produce domain.getRealDomain() File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain return self._cache._realProduce(self._sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce domain = self._findDomain(sdUUID) File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain dom = findMethod(sdUUID) File "/usr/share/vdsm/storage/glusterSD.py", line 32, in findDomain return GlusterStorageDomain(GlusterStorageDomain.findDomainPath(sdUUID)) File "/usr/share/vdsm/storage/fileSD.py", line 160, in __init__ validateFileSystemFeatures(sdUUID, self.mountpoint) File "/usr/share/vdsm/storage/fileSD.py", line 89, in validateFileSystemFeatures oop.getProcessPool(sdUUID).directTouch(testFilePath) File "/usr/share/vdsm/storage/outOfProcess.py", line 351, in directTouch ioproc.touch(path, flags, mode) File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 507, in touch self.timeout) File "/usr/lib/python2.7/site-packages/ioprocess/__init__.py", line 391, in _sendCommand raise OSError(errcode, errstr) OSError: [Errno 5] Input/output error thx Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] Can't fine engine postgres password
Hello, it's running well on ovirt 3.6, but exists the same procedure for ovirt 3.5 ?? Didn't find that. Or can I use this from ovirt 3.6 on 3.5 version ?? regs. Pavel === Hi, database configuration is stored in the VM/host where engine is installed in "/etc/ovirt-engine/engine.conf.d/10-setup-database.conf". If you used automatic database setup, then whole db configuration is generated automatically. And please bear in mind that db user password is different from webadmin user (admin@internal) password which you can specify during installation. Martin Perina - Original Message - From: "Adam Verslype"To:users@ovirt.org Sent: Monday, February 15, 2016 2:50:21 AM Subject: [ovirt-users] Can't fine engine postgres password Hi, After deploying hosted engine using the hosted-engine-appliance I need to run the unlock_entity utility on the engine vm to unlock a disk that's stuck. So I ssh on the vm and run: sudo PGPASSWORD=x ./unlock_entity.sh -t disk -u engine -q psql: FATAL: password authentication failed for user "engine" FATAL: Cannot execute sql command: --file=./unlock_entity.sql where x is the password I supplied for everything during deployment. Does it have a different password that came with the appliance? Is this documented somewhere? Thanks, Adam Verslype ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] delete hang task
Hello, I'm testing oVirt 3.6 for failover and have total issue. Snapshot VM will hang on ZFS filesystem. But the main questionis is : how can I cancel any unfinished tasks in ovirt ??? I didn't find any "normal" solution, except deleting record from ovirt DB manually any idea ? - no one is missing this functionality ??? regs. Pa. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] question mark in oVirt - "cannot acquire state change lock"
Hello, some requested outputs Vdsm generate a huge count of records in time, so I'm not shure if "hang" time is still included . Problems occured 17-feb-2016 17:00-18:00 when VM shutdown was requested. But some performance problems occured a day before ( I'm not shure about application problem = JAVA = running EBS ) In mail before I attached pointers to bugs with any workaround killall -9 libvirtd rm /var/run/libvirtd.pid /etc/init.d/libvirt-bin restart which would solve current situation. Is this workaround safe enought?? Will stay VMs running ??? THX for any help, because I can't reboot hypervizor due huge DB traffic and no VM migration possibility via oVirt engine. ( mem synchro overloaded ?? ) Pa. PS: logs sent directly to RH. On 17.2.2016 08:22, Nir Soffer wrote: On Tue, Feb 16, 2016 at 9:58 PM, p...@email.cz <p...@email.cz> wrote: Hello all, how can I eliminate question mark from "hanged" VM? ( it happened in shutdown process via oVirt env. ) Best solution without hypervizor reboot, meaning "power off" , "kill" or another workaround for that VM . Will help removing task DB record from oVirt mgmt server - if exists , ofcourse ?? Running gluster 3.7.5 / oVirt 3.5.2 / libvirt-daemon-kvm-1.2.8-16.el7_1.5.x86_64 on centos 7.1.1503 . thx a lot. Pa. PS: I found similar problem in libvirt library a many years ago ( redhat 5.6 ) https://bugzilla.redhat.com/show_bug.cgi?id=676205 https://bugzilla.redhat.com/show_bug.cgi?id=668438 problem explanation: http://linux.india365.org/2012/01/libvirt-virsh-cannot-acquire-state-change-lock-error-explained/ This is very interesting, I have seen this error in testing environment on rhel 7.1 or 7.2. Can you provide vdsm logs showing the time of the error? Nir ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] question mark in oVirt - "cannot acquire state change lock"
Hello all, how can I eliminate question mark from "hanged" VM? ( it happened in shutdown process via oVirt env. ) Best solution without hypervizor reboot, meaning "power off" , "kill" or another workaround for that VM . Will help removing task DB record from oVirt mgmt server - if exists , ofcourse ?? Running gluster 3.7.5 / oVirt 3.5.2 / libvirt-daemon-kvm-1.2.8-16.el7_1.5.x86_64 on centos 7.1.1503 . thx a lot. Pa. PS: I found similar problem in libvirt library a many years ago ( redhat 5.6 ) https://bugzilla.redhat.com/show_bug.cgi?id=676205 https://bugzilla.redhat.com/show_bug.cgi?id=668438 problem explanation: http://linux.india365.org/2012/01/libvirt-virsh-cannot-acquire-state-change-lock-error-explained/ So it means that this bug was not fixed to this time ?? regs. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] VM database update ( VMrestore)
Hello, 1) if I've got full VM's datastore backup ( eg. via rsync ) and restore one VM's files to different store and empty oVirt database, will that database be updated automaticaly or any more actions needed ?? 2) Are VM's file stored in oVirt database in names only ? ( no check sum, or other IDs ) - meaning can I replace VM file by another one with different size ?? regs. Paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
Re: [ovirt-users] timeouts
Hi, all glusterd daemons was runnig correctly at this time, no firewalls/iptables restrictions But "not connected" bricks are changing during the time without any touch . It looks that glusterd has non-stable cross communication , especially with different LAN range as nodes in Ovirt environmet ( Volumes bricks in 16.0.0.0 net and ovirt nodes in 172.0.0.0 net ) So I desided reinstall whole cluster, but I'm afraid that these problems will occure again - will you know regs.for your answers Pavel On 27.11.2015 10:16, knarra wrote: On 11/27/2015 11:04 AM, knarra wrote: Hi Paf1, Looks like when you reboot the nodes, glusterd does not start up in one node and due to this the node gets disconnected from other node(that is what i see from logs). After reboot, once your systems are up and running , can you check if glusterd is running on all the nodes? Can you please let me know which build of gluster are you using ? For more info please read, http://www.gluster.org/pipermail/gluster-users.old/2015-June/022377.html - (please ignore this line) Thanks kasturi On 11/27/2015 10:52 AM, Sahina Bose wrote: [+ gluster-users] On 11/26/2015 08:37 PM, p...@email.cz wrote: Hello, can anybody help me with this timeouts ?? Volumes are not active yes ( bricks down ) desc. of gluster bellow ... */var/log/glusterfs/**etc-glusterfs-glusterd.vol.log* [2015-11-26 14:44:47.174221] I [MSGID: 106004] [glusterd-handler.c:5065:__glusterd_peer_rpc_notify] 0-management: Peer <1hp1-SAN> (<87fc7db8-aba8-41f2-a1cd-b77e83b17436>), in state , has disconnected from glusterd. [2015-11-26 14:44:47.174354] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P1 not held [2015-11-26 14:44:47.17] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P3 not held [2015-11-26 14:44:47.174521] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P1 not held [2015-11-26 14:44:47.174662] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P3 not held [2015-11-26 14:44:47.174532] W [MSGID: 106118] [glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: Lock not released for 2HP12-P1 [2015-11-26 14:44:47.174675] W [MSGID: 106118] [glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: Lock not released for 2HP12-P3 [2015-11-26 14:44:49.423334] I [MSGID: 106488] [glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: Received get vol req The message "I [MSGID: 106488] [glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: Received get vol req" repeated 4 times between [2015-11-26 14:44:49.423334] and [2015-11-26 14:44:49.429781] [2015-11-26 14:44:51.148711] I [MSGID: 106163] [glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 0-management: using the op-version 30702 [2015-11-26 14:44:52.177266] W [socket.c:869:__socket_keepalive] 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 12, Invalid argument [2015-11-26 14:44:52.177291] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-26 14:44:53.180426] W [socket.c:869:__socket_keepalive] 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 17, Invalid argument [2015-11-26 14:44:53.180447] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-26 14:44:52.395468] I [MSGID: 106163] [glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 0-management: using the op-version 30702 [2015-11-26 14:44:54.851958] I [MSGID: 106488] [glusterd-ha
[ovirt-users] timeouts
Hello, can anybody help me with this timeouts ?? Volumes are not active yes ( bricks down ) desc. of gluster bellow ... */var/log/glusterfs/**etc-glusterfs-glusterd.vol.log* [2015-11-26 14:44:47.174221] I [MSGID: 106004] [glusterd-handler.c:5065:__glusterd_peer_rpc_notify] 0-management: Peer <1hp1-SAN> (<87fc7db8-aba8-41f2-a1cd-b77e83b17436>), in state Cluster>, has disconnected from glusterd. [2015-11-26 14:44:47.174354] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P1 not held [2015-11-26 14:44:47.17] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 1HP12-P3 not held [2015-11-26 14:44:47.174521] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P1 not held [2015-11-26 14:44:47.174662] W [glusterd-locks.c:681:glusterd_mgmt_v3_unlock] (-->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_big_locked_notify+0x4c) [0x7fb7039d44dc] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(__glusterd_peer_rpc_notify+0x162) [0x7fb7039de542] -->/usr/lib64/glusterfs/3.7.6/xlator/mgmt/glusterd.so(glusterd_mgmt_v3_unlock+0x58a) [0x7fb703a79b4a] ) 0-management: Lock for vol 2HP12-P3 not held [2015-11-26 14:44:47.174532] W [MSGID: 106118] [glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: Lock not released for 2HP12-P1 [2015-11-26 14:44:47.174675] W [MSGID: 106118] [glusterd-handler.c:5087:__glusterd_peer_rpc_notify] 0-management: Lock not released for 2HP12-P3 [2015-11-26 14:44:49.423334] I [MSGID: 106488] [glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: Received get vol req The message "I [MSGID: 106488] [glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: Received get vol req" repeated 4 times between [2015-11-26 14:44:49.423334] and [2015-11-26 14:44:49.429781] [2015-11-26 14:44:51.148711] I [MSGID: 106163] [glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 0-management: using the op-version 30702 [2015-11-26 14:44:52.177266] W [socket.c:869:__socket_keepalive] 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 12, Invalid argument [2015-11-26 14:44:52.177291] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-26 14:44:53.180426] W [socket.c:869:__socket_keepalive] 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 17, Invalid argument [2015-11-26 14:44:53.180447] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-26 14:44:52.395468] I [MSGID: 106163] [glusterd-handshake.c:1193:__glusterd_mgmt_hndsk_versions_ack] 0-management: using the op-version 30702 [2015-11-26 14:44:54.851958] I [MSGID: 106488] [glusterd-handler.c:1472:__glusterd_handle_cli_get_volume] 0-glusterd: Received get vol req [2015-11-26 14:44:57.183969] W [socket.c:869:__socket_keepalive] 0-socket: failed to set TCP_USER_TIMEOUT -1000 on socket 19, Invalid argument [2015-11-26 14:44:57.183990] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument After volumes creation all works fine ( volumes up ) , but then, after several reboots ( yum updates) volumes failed due timeouts . Gluster description: 4 nodes with 4 volumes replica 2 oVirt 3.6 - the last gluster 3.7.6 - the last vdsm 4.17.999 - from git repo oVirt - mgmt.nodes 172.16.0.0 oVirt - bricks 16.0.0.0 ( "SAN" - defined as "gluster" net) Network works fine, no lost packets # gluster volume status Staging failed on 2hp1-SAN. Please check log file for details. Staging failed on 1hp2-SAN. Please check log file for details. Staging failed on 2hp2-SAN. Please check log file for details. # gluster volume info Volume Name: 1HP12-P1 Type: Replicate Volume ID: 6991e82c-9745-4203-9b0a-df202060f455 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 1hp1-SAN:/STORAGE/p1/G Brick2: 1hp2-SAN:/STORAGE/p1/G Options Reconfigured: performance.readdir-ahead: on Volume Name: 1HP12-P3
[ovirt-users] remove volume from DB
Hello, howto remove volume definition from oVirt DB ( & from nodes gluster config ) if volume totaly cleaned in background in running mode ?? regs. Paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] volume parameters
Hello, would U recommend me set of functional params for volume replica 2 ? Old ones was ( for 3.5.2 gluster version ) storage.owner-uid 36 storage.owner-gid 36 performance.io-cache off performance.read-ahead off network.remote-dio enable cluster.eager-lock enable performance.stat-prefetch off performance.quick-read off cluster.quorum-count 1 cluster.server-quorum-type none cluster.quorum-type fixed after upgrade to 3.5.7 version and setting default recommendation, volumes became inaccessable ( permission denied - fixed by owner uid/gui settings to 36) Why the defaults have been changed ? Just still Error / Critical messages occure ( examples follow ) *E* - list of grep etc-glusterfs-glusterd.vol.log [2015-11-07 10:49:10.883564] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-07 10:49:10.886152] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-07 10:49:15.954942] E [rpc-clnt.c:362:saved_frames_unwind] (--> /lib64/libglusterfs.so.0(_gf_log_callingfn+0x196)[0x7fa88b014a66] (--> /lib64/libgfrpc.so.0(saved_frames_unwind+0x1de)[0x7fa88addf9be] (--> /lib64/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7fa88addface] (--> /lib64/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x9c)[0x7fa88ade148c] (--> /lib64/libgfrpc.so.0(rpc_clnt_notify+0x88)[0x7fa88ade1c98] ) 0-management: forced unwinding frame type(Peer mgmt) op(--(2)) called at 2015-11-07 10:49:10.918764 (xid=0x5) [2015-11-07 10:49:26.719176] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-07 10:54:59.738232] E [MSGID: 106243] [glusterd.c:1623:init] 0-management: creation of 1 listeners failed, continuing with succeeded transport [2015-11-07 10:55:01.860991] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-07 10:55:01.863932] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument [2015-11-07 10:55:01.866779] E [socket.c:2965:socket_connect] 0-management: Failed to set keep-alive: Invalid argument *C* - list of grep etc-glusterfs-glusterd.vol.log [2015-11-07 10:49:16.045778] C [MSGID: 106003] [glusterd-server-quorum.c:346:glusterd_do_volume_quorum_action] 0-management: Server quorum regained for volume 1KVM12-P4. Starting local bricks. [2015-11-07 10:49:16.049319] C [MSGID: 106003] [glusterd-server-quorum.c:346:glusterd_do_volume_quorum_action] 0-management: Server quorum regained for volume 1KVM12-P5. Starting local bricks. regs.Paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] can oVirt 3.6 manage 3.5 hypervizor
Hi, can oVirt 3.6 manage hypervizors with 3.5 version ? Meaning during cluster upgrade step by step. ( A) oVirt mgmt , B) 1st.hypervizor, C) 2nd hypervizor, .. ) If oVirt DB converted from 3.5.5 -> 3.5.5.upg.3.6 -> final 3.6 regs. Paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] ovirt DB 3.6 to empty 3.6 DB
Hello, can anybody help me with importing oVirt DB v.3.5 to new clean oVIrt 3.6 database ?? # engine-backup --mode=restore --file=./backup --log=./restore.log Preparing to restore: - Unpacking file './backup' FATAL: Backup version '3.6' doesn't match installed version How to FIX it ?? Regs. Paf1 ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] oVirt timeouts
Hello, pls, howto change oVirt timeouts for status : 1) node ( brick) is power down / up 2) volume status for node(brick) is up/down They are too long ( I'm expecting a few sec. not a lot of minutes ) If it has some spacial reason, let me know about, pls. regs. Pa. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] Replica2 stripe2 hang on write to VM disk
Hello, can anybody help me with hanging replica2 stripe2 datastore on 4 nodes cluster ?? oVirt - ovirt-engine-lib-3.5.2.1-1.el7.centos.noarch gluster - glusterfs-server-3.7.0-2.el7.x86_64 VM - Centos 7.1 If I use any bigger write to VM disk ( eg 2-5GB ) storage hosted virtual disk will hang = I/O error created by : gluster volume create 12KVM12SC4 replica 2 stripe 2 16.0.0.161:/STORAGES/SlowClass/p4/GFS1 16.0.0.162:/STORAGES/SlowClass/p4/GFS1 16.0.0.163:/STORAGES/SlowClass/p4/GFS1 16.0.0.164:/STORAGES/SlowClass/p4/GFS1 rhev-data-center-mnt-glusterSD-localhost:_12KVM12SC4.log --- [2015-05-25 14:47:24.205609] I [rpc-clnt.c:1807:rpc_clnt_reconfig] 0-12KVM12SC4-client-3: changing port to 49158 (from 0) [2015-05-25 14:47:24.210824] I [client-handshake.c:1405:select_server_supported_programs] 0-12KVM12SC4-client-3: Using Program GlusterFS 3.3, Num (1298437), Version (330) [2015-05-25 14:47:24.211204] I [client-handshake.c:1193:client_setvolume_cbk] 0-12KVM12SC4-client-3: Connected to 12KVM12SC4-client-3, attached to remote volume '/STORAGES/SlowClass/p4/GFS1'. [2015-05-25 14:47:24.211225] I [client-handshake.c:1203:client_setvolume_cbk] 0-12KVM12SC4-client-3: Server and Client lk-version numbers are not same, reopening the fds [2015-05-25 14:47:24.211275] I [MSGID: 108005] [afr-common.c:3880:afr_notify] 0-12KVM12SC4-replicate-1: Subvolume '12KVM12SC4-client-3' came back up; going online. [2015-05-25 14:47:24.216465] I [fuse-bridge.c:5077:fuse_graph_setup] 0-fuse: switched to graph 0 [2015-05-25 14:47:24.216556] I [client-handshake.c:187:client_set_lk_version_cbk] 0-12KVM12SC4-client-3: Server lk version = 1 [2015-05-25 14:47:24.216643] I [fuse-bridge.c:4007:fuse_init] 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.22 kernel 7.22 [2015-05-25 14:47:24.217998] I [afr-common.c:1673:afr_local_discovery_cbk] 0-12KVM12SC4-replicate-0: selecting local read_child 12KVM12SC4-client-0 [2015-05-25 14:47:29.737732] W [fuse-bridge.c:1080:fuse_setattr_cbk] 0-glusterfs-fuse: 40: SETATTR() /__DIRECT_IO_TEST__ = -1 (Read-only file system) [2015-05-25 14:49:18.266212] E [client-handshake.c:1488:client_query_portmap_cbk] 0-12KVM12SC4-client-2: failed to get the port number for remote subvolume. Please run 'gluster volume status' on server to see if brick process is running. [2015-05-25 14:49:18.266274] I [client.c:2086:client_rpc_notify] 0-12KVM12SC4-client-2: disconnected from 12KVM12SC4-client-2. Client process will keep trying to connect to glusterd until brick's port is available [2015-05-25 14:49:19.346555] I [rpc-clnt.c:1807:rpc_clnt_reconfig] 0-12KVM12SC4-client-2: changing port to 49158 (from 0) [2015-05-25 14:49:19.351812] I [client-handshake.c:1405:select_server_supported_programs] 0-12KVM12SC4-client-2: Using Program GlusterFS 3.3, Num (1298437), Version (330) [2015-05-25 14:49:19.352169] I [client-handshake.c:1193:client_setvolume_cbk] 0-12KVM12SC4-client-2: Connected to 12KVM12SC4-client-2, attached to remote volume '/STORAGES/SlowClass/p4/GFS1'. [2015-05-25 14:49:19.352191] I [client-handshake.c:1203:client_setvolume_cbk] 0-12KVM12SC4-client-2: Server and Client lk-version numbers are not same, reopening the fds [2015-05-25 14:49:19.352242] I [MSGID: 108002] [afr-common.c:3959:afr_notify] 0-12KVM12SC4-replicate-1: Client-quorum is met [2015-05-25 14:49:19.352353] I [client-handshake.c:187:client_set_lk_version_cbk] 0-12KVM12SC4-client-2: Server lk version = 1 [2015-05-25 14:49:27.843616] W [fuse-bridge.c:1263:fuse_err_cbk] 0-glusterfs-fuse: 151: REMOVEXATTR() /__DIRECT_IO_TEST__ = -1 (No data available) [2015-05-25 14:49:58.356900] W [fuse-bridge.c:1263:fuse_err_cbk] 0-glusterfs-fuse: 327: REMOVEXATTR() /__DIRECT_IO_TEST__ = -1 (No data available) # gluster volume status Status of volume: 12KVM12SC4 Gluster process TCP Port RDMA Port Online Pid -- Brick 16.0.0.161:/STORAGES/SlowClass/p4/GFS 1 49173 0 Y 17678 Brick 16.0.0.162:/STORAGES/SlowClass/p4/GFS 1 49158 0 Y 19184 Brick 16.0.0.163:/STORAGES/SlowClass/p4/GFS 1 49158 0 Y 9784 Brick 16.0.0.164:/STORAGES/SlowClass/p4/GFS 1 49158 0 Y 9327 NFS Server on localhost 2049 0 Y 17697 Self-heal Daemon on localhost N/A N/AY 17708 NFS Server on 16.0.0.1622049 0 Y 19205 Self-heal Daemon on 16.0.0.162 N/A N/AY 19215 NFS Server on 16.0.0.1632049 0 Y 9806 Self-heal Daemon on 16.0.0.163 N/A N/AY 9813 NFS Server
[ovirt-users] VM pause when host crash/poweroff
Hello, how to solve situation if VM going to pause mode when one of replica 2 node is stopped ( power down or crash, etc. .. ) VM is not running on stopped node and OVirt message get VM xyz has paused due to unknown storage error Where the error is ?? regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] snapshot deleting
Hello, my question is about deleting default snapshots generated via disk moving between storages. The snapshot called Auto-generated for Live Storage Migration is NOT automaticaly deleted after movement done. Is this default or any bug ?? regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] disk format hang
Hello, during VM instalation I needed add 300GB new disk and format as EXT3. Easy task, but Used striped mirror volume for that - created by: # gluster volume create 12KVM12SC4 replica 2 stripe 2 1kvm1-SAN:/p4/GFS1 1kvm2-SAN:/p4/GFS1 2kvm1-SAN /p4/GFS1 2kvm2-SAN:/p4/GFS1 force - added disk throught oVisrt GUI - VM# partprobe - VM# fdisk - make one partition, type linux - VM# mkfs.ext3 /dev/vdb1 it hangs after cca 30% of format oVirt alert: *VM has been paused due to a storage I/O error* so I tried to copy some data directly from hypervizor ( cca 80GB ) to default gluster-ovirt mountpoint ( /rhev/data-center/mnt/glusterSD/localhost:_12KVM12SC4 ) with no hangs. The same whole operation on volume distribute replica type succesfully done. Can anybody help me with it ?? Maybe continual error on /var/log/messages occure will get right view on it : May 20 15:47:00 1kvm2 virsh: All-whitespace username. May 20 15:47:00 1kvm2 journal: End of file while reading data: Input/output error May 20 15:47:00 1kvm2 virsh: All-whitespace username. May 20 15:47:00 1kvm2 journal: End of file while reading data: Input/output error May 20 15:48:00 1kvm2 virsh: All-whitespace username. May 20 15:48:00 1kvm2 journal: End of file while reading data: Input/output error May 20 15:48:00 1kvm2 virsh: All-whitespace username. May 20 15:48:00 1kvm2 journal: End of file while reading data: Input/output error regs to ALL !! URGENT !! Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] disk moving
hello, would I explicitly define interface for only disk moving ?? reg. Pa. ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] simple virtual disk moving
hello, would I explicitly define interface for only disk moving ?? Move process takes the first interface from list ( VM assigned ), but any interface for moving virtuals ( explicitly defined ). eg. VM moving from A to B - it takes interface defined for moving ( movingLAN ) but moving only simple disk of VM from A to B - it takes default interface from list of assigned LANs any idea ?? regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] gluster storage ( wanted / supported / recommended configuration )
Hello dears, is anybody here for serious storage conversation ?? I've got any ideas and a lot of errors - gluster filesystem concept checking let me know, will sent diagram and questions regs. Pavel ___ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
[ovirt-users] ovirt - import detached gluter volumes
Hello dears, i've got some troubles with reattaching gluster volumes with data. 1) Base on a lot of tests I decided clear oVirt database ( # engine-cleanup ; # yum remove ovirt-engine; # yum -y install ovirt-engine; # engine-setup) 2) clearing sucessfully done and start with empty oVirt envir. 3) then I added networks, nodes and make basic network adjustment = all works fine 4) time to attach volumes/ domains with original data ( a lot of virtuals , ISO files , ) So, main question is about HOWTO attach this volumes if I haven't defined any domain and can't clearly import them ?? Current status of nodes are without glusterfs NFS mounted, but bricks are OK # gluster volume info Volume Name: VGFS1 Type: Replicate Volume ID: b9a1c347-6ffd-4122-8756-d513fe3f40b9 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 1kvm2:/FastClass/p1/GFS1 Brick2: 1kvm1:/FastClass/p1/GFS1 Options Reconfigured: storage.owner-gid: 36 storage.owner-uid: 36 Volume Name: VGFS2 Type: Replicate Volume ID: b65bb689-ecc8-4c33-a4e7-11dea6028f83 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 1kvm2:/FastClass/p2/GFS1 Brick2: 1kvm1:/FastClass/p2/GFS1 Options Reconfigured: storage.owner-uid: 36 storage.owner-gid: 36 [root@1kvm1 glusterfs]# gluster volume status Status of volume: VGFS1 Gluster process PortOnline Pid -- Brick 1kvm1:/FastClass/p1/GFS1 N/A N N/A NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N/A N N/A Task Status of Volume VGFS1 -- There are no active volume tasks Status of volume: VGFS2 Gluster process PortOnline Pid -- Brick 1kvm1:/FastClass/p2/GFS1 N/A N N/A NFS Server on localhost N/A N N/A Self-heal Daemon on localhost N/A N N/A Task Status of Volume VGFS2 -- There are no active volume tasks [root@1kvm1 glusterfs]# gluster volume start VGFS1 volume start: VGFS1: failed: Volume VGFS1 already started # mount | grep mapper # base XFS mounting /dev/mapper/3600605b0099f9e601cb1b5bf0e9765e8p1 on /FastClass/p1 type xfs (rw,relatime,seclabel,attr2,inode64,noquota) /dev/mapper/3600605b0099f9e601cb1b5bf0e9765e8p2 on /FastClass/p2 type xfs (rw,relatime,seclabel,attr2,inode64,noquota) *5)* import screen /VGFS1 dir exists iptables flushed # cat rhev-data-center-mnt-glusterSD-1kvm1:_VGFS1.log [2015-04-22 15:21:50.204521] I [MSGID: 100030] [glusterfsd.c:2018:main] 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.6.2 (args: /usr/sbin/glusterfs --volfile-server=1kvm1 --volfile-id=/VGFS1 /rhev/data-center/mnt/glusterSD/1kvm1:_VGFS1) [2015-04-22 15:21:50.220383] I [dht-shared.c:337:dht_init_regex] 0-VGFS1-dht: using regex rsync-hash-regex = ^\.(.+)\.[^.]+$ [2015-04-22 15:21:50.55] I [client.c:2280:notify] 0-VGFS1-client-1: parent translators are ready, attempting connect on transport [2015-04-22 15:21:50.224528] I [client.c:2280:notify] 0-VGFS1-client-2: parent translators are ready, attempting connect on transport Final graph: +--+ 1: volume VGFS1-client-1 2: type protocol/client 3: option ping-timeout 42 4: option remote-host 1kvm2 5: option remote-subvolume /FastClass/p1/GFS1 6: option transport-type socket 7: option username 52f1efd1-60dc-4fb1-b94f-572945d6eb66 8: option password 34bac9cd-0b4f-41c6-973b-7af568784d7b 9: option send-gids true 10: end-volume 11: 12: volume VGFS1-client-2 13: type protocol/client 14: option ping-timeout 42 15: option remote-host 1kvm1 16: option remote-subvolume /FastClass/p1/GFS1 17: option transport-type socket 18: option username 52f1efd1-60dc-4fb1-b94f-572945d6eb66 19: option password 34bac9cd-0b4f-41c6-973b-7af568784d7b 20: option send-gids true 21: end-volume 22: 23: volume VGFS1-replicate-0 24: type cluster/replicate 25: subvolumes VGFS1-client-1 VGFS1-client-2 26: end-volume 27: 28: volume VGFS1-dht 29: type cluster/distribute 30: subvolumes VGFS1-replicate-0 31: end-volume 32: 33: volume VGFS1-write-behind 34: type performance/write-behind 35: subvolumes VGFS1-dht 36: end-volume 37: 38: volume VGFS1-read-ahead 39: type performance/read-ahead 40: subvolumes VGFS1-write-behind 41: end-volume 42: 43: volume VGFS1-io-cache 44: type performance/io-cache 45: subvolumes VGFS1-read-ahead 46: end-volume 47: 48: volume VGFS1-quick-read 49: type performance/quick-read 50: subvolumes VGFS1-io-cache 51: end-volume 52: 53: volume VGFS1-open-behind 54: