Re: [ovirt-users] Fwd: Re: ovirt - can't attach master domain II

2016-02-24 Thread p...@email.cz

used replica2 with volume option

Volume Name: 2KVM12-P2
Type: Replicate
Volume ID: 9745551f-4696-4a6c-820a-619e359a61fd
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 16.0.0.164:/STORAGES/g1r5p2/GFS
Brick2: 16.0.0.163:/STORAGES/g1r5p2/GFS
Options Reconfigured:
storage.owner-uid: 36
storage.owner-gid: 36
performance.io-cache: off
performance.read-ahead: off
network.remote-dio: enable
cluster.eager-lock: enable
performance.stat-prefetch: off
performance.quick-read: off
cluster.quorum-count: 1
cluster.server-quorum-type: none
cluster.quorum-type: fixed

was runnig over year with no problems ( reboots, ..etc... )


On 24.2.2016 12:34, Ravishankar N wrote:

On 02/24/2016 04:48 PM, p...@email.cz wrote:



prereq: 2KVM12-P2 = master domain
-
YES - I'm using gluster.fuse NFS
localhost:/2KVM12-P2 on 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2 type 
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)

-
Healing
==
# gluster volume heal 2KVM12-P2 info
Brick 16.0.0.164:/STORAGES/g1r5p2/GFS
Number of entries: 0

Brick 16.0.0.163:/STORAGES/g1r5p2/GFS
Number of entries: 0

# while true; do for vol in `gluster volume list`; do gluster volume 
heal $vol info | sort | grep "Number of entries" | awk -F: '{tot+=$2} 
END { printf("Heal entries for '"$vol"': %d\n", $tot);}'; done; sleep 
120; echo -e "\n==\n"; done

Heal entries for 1KVM12-BCK: 1
Heal entries for 1KVM12-P1: 1
Heal entries for 1KVM12-P2: 0
Heal entries for 1KVM12-P3: 0
Heal entries for 1KVM12-P4: 0
Heal entries for 1KVM12-P5: 0
Heal entries for 2KVM12-P1: 1
Heal entries for 2KVM12-P2: 0
Heal entries for 2KVM12-P3: 0
Heal entries for 2KVM12-P5: 0
Heal entries for 2KVM12_P4: 1

# gluster volume heal 1KVM12-BCK info split-brain
Brick 16.0.0.161:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

# gluster volume heal 1KVM12-P1 info split-brain
Brick 16.0.0.161:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

etc..


YES - in split brain , but NOT master domain ( will solve later, 
after master - if possible  )


I'm not sure if it is related, but you could try to resolve the 
split-brain first and see if it helps. Also, I see that you are using 
replica-2. It is recommended to use replica-3 or arbiter volumes to 
avoid split-brains.


-Ravi



---
vdsm.log
=

Thread-461::DEBUG::2016-02-24 
11:12:45,328::fileSD::262::Storage.Misc.excCmd::(getReadDelay) 
SUCCESS:  = '0+1 records in\n0+1 records out\n333 bytes (333 B) 
copied, 0.000724379 s, 460 kB/s\n';  = 0
Thread-461::INFO::2016-02-24 
11:12:45,331::clusterlock::219::Storage.SANLock::(acquireHostId) 
Acquiring host id for domain 88adbd49-62d6-45b1-9992-b04464a04112 (id: 3)
Thread-461::DEBUG::2016-02-24 
11:12:45,331::clusterlock::237::Storage.SANLock::(acquireHostId) Host 
id for domain 88adbd49-62d6-45b1-9992-b04464a04112 successfully 
acquired (id: 3)
Thread-33186::DEBUG::2016-02-24 
11:12:46,067::__init__::481::jsonrpc.JsonRpcServer::(_serveRequest) 
Calling 'GlusterVolume.list' in bridge with {}
Thread-33186::DEBUG::2016-02-24 
11:12:46,204::__init__::514::jsonrpc.JsonRpcServer::(_serveRequest) 
Return 'GlusterVolume.list' in bridge with {'volumes': {'2KVM12-P5': 
{'transportType': ['TCP'], 'uuid': 
'4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d', 'bricks': 
['16.0.0.164:/STORAGES/g1r5p5/GFS', 
'16.0.0.163:/STORAGES/g1r5p5/GFS'], 'volumeName': '2KVM12-P5', 
'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 
'distCount': '2', 'volumeStatus': 'ONLINE', 'stripeCount': '1', 
'bricksInfo': [{'name': '16.0.0.164:/STORAGES/g1r5p5/GFS', 
'hostUuid': '06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}, {'name': 
'16.0.0.163:/STORAGES/g1r5p5/GFS', 'hostUuid': 
'6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}], 'options': 
{'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 
'enable', 'performance.stat-prefetch': 'off', 'cluster.quorum-type': 
'fixed', 'performance.quick-read': 'off', 'network.remote-dio': 
'enable', 'cluster.quorum-count': '1', 'performance.io-cache': 'off', 
'storage.owner-uid': '36', 'performance.read-ahead': 'off', 
'storage.owner-gid': '36'}}, '2KVM12_P4': {'transportType': ['TCP'], 
'uuid': '18310aeb-639f-4b6d-9ef4-9ef560d6175c', 'bricks': 
['16.0.0.163:/STORAGES/g1r5p4/GFS', 
'16.0.0.164:/STORAGES/g1r5p4/GFS'], 'volumeName': '2KVM12_P4', 
'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 
'distCount': '2', 'volumeStatus': 'ONLINE', 'stripeCount': 

Re: [ovirt-users] Fwd: Re: ovirt - can't attach master domain II

2016-02-24 Thread Ravishankar N

On 02/24/2016 04:48 PM, p...@email.cz wrote:



prereq: 2KVM12-P2 = master domain
-
YES - I'm using gluster.fuse NFS
localhost:/2KVM12-P2 on 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2 type 
fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)

-
Healing
==
# gluster volume heal 2KVM12-P2 info
Brick 16.0.0.164:/STORAGES/g1r5p2/GFS
Number of entries: 0

Brick 16.0.0.163:/STORAGES/g1r5p2/GFS
Number of entries: 0

# while true; do for vol in `gluster volume list`; do gluster volume 
heal $vol info | sort | grep "Number of entries" | awk -F: '{tot+=$2} 
END { printf("Heal entries for '"$vol"': %d\n", $tot);}'; done; sleep 
120; echo -e "\n==\n"; done

Heal entries for 1KVM12-BCK: 1
Heal entries for 1KVM12-P1: 1
Heal entries for 1KVM12-P2: 0
Heal entries for 1KVM12-P3: 0
Heal entries for 1KVM12-P4: 0
Heal entries for 1KVM12-P5: 0
Heal entries for 2KVM12-P1: 1
Heal entries for 2KVM12-P2: 0
Heal entries for 2KVM12-P3: 0
Heal entries for 2KVM12-P5: 0
Heal entries for 2KVM12_P4: 1

# gluster volume heal 1KVM12-BCK info split-brain
Brick 16.0.0.161:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

# gluster volume heal 1KVM12-P1 info split-brain
Brick 16.0.0.161:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

etc..


YES - in split brain , but NOT master domain ( will solve later, after 
master - if possible  )


I'm not sure if it is related, but you could try to resolve the 
split-brain first and see if it helps. Also, I see that you are using 
replica-2. It is recommended to use replica-3 or arbiter volumes to 
avoid split-brains.


-Ravi



---
vdsm.log
=

Thread-461::DEBUG::2016-02-24 
11:12:45,328::fileSD::262::Storage.Misc.excCmd::(getReadDelay) 
SUCCESS:  = '0+1 records in\n0+1 records out\n333 bytes (333 B) 
copied, 0.000724379 s, 460 kB/s\n';  = 0
Thread-461::INFO::2016-02-24 
11:12:45,331::clusterlock::219::Storage.SANLock::(acquireHostId) 
Acquiring host id for domain 88adbd49-62d6-45b1-9992-b04464a04112 (id: 3)
Thread-461::DEBUG::2016-02-24 
11:12:45,331::clusterlock::237::Storage.SANLock::(acquireHostId) Host 
id for domain 88adbd49-62d6-45b1-9992-b04464a04112 successfully 
acquired (id: 3)
Thread-33186::DEBUG::2016-02-24 
11:12:46,067::__init__::481::jsonrpc.JsonRpcServer::(_serveRequest) 
Calling 'GlusterVolume.list' in bridge with {}
Thread-33186::DEBUG::2016-02-24 
11:12:46,204::__init__::514::jsonrpc.JsonRpcServer::(_serveRequest) 
Return 'GlusterVolume.list' in bridge with {'volumes': {'2KVM12-P5': 
{'transportType': ['TCP'], 'uuid': 
'4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d', 'bricks': 
['16.0.0.164:/STORAGES/g1r5p5/GFS', 
'16.0.0.163:/STORAGES/g1r5p5/GFS'], 'volumeName': '2KVM12-P5', 
'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 
'distCount': '2', 'volumeStatus': 'ONLINE', 'stripeCount': '1', 
'bricksInfo': [{'name': '16.0.0.164:/STORAGES/g1r5p5/GFS', 'hostUuid': 
'06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}, {'name': 
'16.0.0.163:/STORAGES/g1r5p5/GFS', 'hostUuid': 
'6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}], 'options': 
{'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 'enable', 
'performance.stat-prefetch': 'off', 'cluster.quorum-type': 'fixed', 
'performance.quick-read': 'off', 'network.remote-dio': 'enable', 
'cluster.quorum-count': '1', 'performance.io-cache': 'off', 
'storage.owner-uid': '36', 'performance.read-ahead': 'off', 
'storage.owner-gid': '36'}}, '2KVM12_P4': {'transportType': ['TCP'], 
'uuid': '18310aeb-639f-4b6d-9ef4-9ef560d6175c', 'bricks': 
['16.0.0.163:/STORAGES/g1r5p4/GFS', 
'16.0.0.164:/STORAGES/g1r5p4/GFS'], 'volumeName': '2KVM12_P4', 
'volumeType': 'REPLICATE', 'replicaCount': '2', 'brickCount': '2', 
'distCount': '2', 'volumeStatus': 'ONLINE', 'stripeCount': '1', 
'bricksInfo': [{'name': '16.0.0.163:/STORAGES/g1r5p4/GFS', 'hostUuid': 
'6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}, {'name': 
'16.0.0.164:/STORAGES/g1r5p4/GFS', 'hostUuid': 
'06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}], 'options': 
{'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 'enable', 
'performance.io-cache': 'off', 'performance.stat-prefetch': 'off', 
'cluster.quorum-type': 'fixed', 'performance.quick-read': 'off', 
'network.remote-dio': 'enable', 'cluster.quorum-count': '1', 
'storage.owner-uid': '36', 'performance.read-ahead': 'off', 
'storage.owner-gid': '36'}}, '2KVM12-P1': {'transportType': ['TCP'], 
'uuid': 'cbf142f8-a40b-4cf4-ad29-2243c81d30c1', 'bricks': 

[ovirt-users] Fwd: Re: ovirt - can't attach master domain II

2016-02-24 Thread p...@email.cz



prereq: 2KVM12-P2 = master domain
-
YES - I'm using gluster.fuse NFS
localhost:/2KVM12-P2 on 
/rhev/data-center/mnt/glusterSD/localhost:_2KVM12-P2 type fuse.glusterfs 
(rw,relatime,user_id=0,group_id=0,default_permissions,allow_other,max_read=131072)

-
Healing
==
# gluster volume heal 2KVM12-P2 info
Brick 16.0.0.164:/STORAGES/g1r5p2/GFS
Number of entries: 0

Brick 16.0.0.163:/STORAGES/g1r5p2/GFS
Number of entries: 0

# while true; do for vol in `gluster volume list`; do gluster volume 
heal $vol info | sort | grep "Number of entries" | awk -F: '{tot+=$2} 
END { printf("Heal entries for '"$vol"': %d\n", $tot);}'; done; sleep 
120; echo -e "\n==\n"; done

Heal entries for 1KVM12-BCK: 1
Heal entries for 1KVM12-P1: 1
Heal entries for 1KVM12-P2: 0
Heal entries for 1KVM12-P3: 0
Heal entries for 1KVM12-P4: 0
Heal entries for 1KVM12-P5: 0
Heal entries for 2KVM12-P1: 1
Heal entries for 2KVM12-P2: 0
Heal entries for 2KVM12-P3: 0
Heal entries for 2KVM12-P5: 0
Heal entries for 2KVM12_P4: 1

# gluster volume heal 1KVM12-BCK info split-brain
Brick 16.0.0.161:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g2r5p1/GFS
/0fcad888-d573-47be-bef3-0bc0b7a99fb7/dom_md/ids
Number of entries in split-brain: 1

# gluster volume heal 1KVM12-P1 info split-brain
Brick 16.0.0.161:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

Brick 16.0.0.162:/STORAGES/g1r5p1/GFS
/__DIRECT_IO_TEST__
Number of entries in split-brain: 1

etc..


YES - in split brain , but NOT master domain ( will solve later, after 
master - if possible  )


---
vdsm.log
=

Thread-461::DEBUG::2016-02-24 
11:12:45,328::fileSD::262::Storage.Misc.excCmd::(getReadDelay) SUCCESS: 
 = '0+1 records in\n0+1 records out\n333 bytes (333 B) copied, 
0.000724379 s, 460 kB/s\n';  = 0
Thread-461::INFO::2016-02-24 
11:12:45,331::clusterlock::219::Storage.SANLock::(acquireHostId) 
Acquiring host id for domain 88adbd49-62d6-45b1-9992-b04464a04112 (id: 3)
Thread-461::DEBUG::2016-02-24 
11:12:45,331::clusterlock::237::Storage.SANLock::(acquireHostId) Host id 
for domain 88adbd49-62d6-45b1-9992-b04464a04112 successfully acquired 
(id: 3)
Thread-33186::DEBUG::2016-02-24 
11:12:46,067::__init__::481::jsonrpc.JsonRpcServer::(_serveRequest) 
Calling 'GlusterVolume.list' in bridge with {}
Thread-33186::DEBUG::2016-02-24 
11:12:46,204::__init__::514::jsonrpc.JsonRpcServer::(_serveRequest) 
Return 'GlusterVolume.list' in bridge with {'volumes': {'2KVM12-P5': 
{'transportType': ['TCP'], 'uuid': 
'4a6d775d-4a51-4f6c-9bfa-f7ef57f3ca1d', 'bricks': 
['16.0.0.164:/STORAGES/g1r5p5/GFS', '16.0.0.163:/STORAGES/g1r5p5/GFS'], 
'volumeName': '2KVM12-P5', 'volumeType': 'REPLICATE', 'replicaCount': 
'2', 'brickCount': '2', 'distCount': '2', 'volumeStatus': 'ONLINE', 
'stripeCount': '1', 'bricksInfo': [{'name': 
'16.0.0.164:/STORAGES/g1r5p5/GFS', 'hostUuid': 
'06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}, {'name': 
'16.0.0.163:/STORAGES/g1r5p5/GFS', 'hostUuid': 
'6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}], 'options': 
{'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 'enable', 
'performance.stat-prefetch': 'off', 'cluster.quorum-type': 'fixed', 
'performance.quick-read': 'off', 'network.remote-dio': 'enable', 
'cluster.quorum-count': '1', 'performance.io-cache': 'off', 
'storage.owner-uid': '36', 'performance.read-ahead': 'off', 
'storage.owner-gid': '36'}}, '2KVM12_P4': {'transportType': ['TCP'], 
'uuid': '18310aeb-639f-4b6d-9ef4-9ef560d6175c', 'bricks': 
['16.0.0.163:/STORAGES/g1r5p4/GFS', '16.0.0.164:/STORAGES/g1r5p4/GFS'], 
'volumeName': '2KVM12_P4', 'volumeType': 'REPLICATE', 'replicaCount': 
'2', 'brickCount': '2', 'distCount': '2', 'volumeStatus': 'ONLINE', 
'stripeCount': '1', 'bricksInfo': [{'name': 
'16.0.0.163:/STORAGES/g1r5p4/GFS', 'hostUuid': 
'6482ae32-25ac-41b5-b41d-b7ddf49bac2c'}, {'name': 
'16.0.0.164:/STORAGES/g1r5p4/GFS', 'hostUuid': 
'06854ac0-2ef1-4c12-bb8d-56cf9bf95ec9'}], 'options': 
{'cluster.server-quorum-type': 'none', 'cluster.eager-lock': 'enable', 
'performance.io-cache': 'off', 'performance.stat-prefetch': 'off', 
'cluster.quorum-type': 'fixed', 'performance.quick-read': 'off', 
'network.remote-dio': 'enable', 'cluster.quorum-count': '1', 
'storage.owner-uid': '36', 'performance.read-ahead': 'off', 
'storage.owner-gid': '36'}}, '2KVM12-P1': {'transportType': ['TCP'], 
'uuid': 'cbf142f8-a40b-4cf4-ad29-2243c81d30c1', 'bricks': 
['16.0.0.163:/STORAGES/g1r5p1/GFS', '16.0.0.164:/STORAGES/g1r5p1/GFS'], 
'volumeName': '2KVM12-P1', 'volumeType': 'REPLICATE', 'replicaCount': 
'2', 'brickCount': '2', 'distCount': '2', 'volumeStatus': 'ONLINE', 
'stripeCount': '1', 'bricksInfo': [{'name': 
'16.0.0.163:/STORAGES/g1r5p1/GFS',