Hi all,

we are having a very strange behavior with exclusive locks.
We have one image called test inside a pool called app.

This is the output of rbd status app/test

rbd image 'test':
> size 120 GB in 30720 objects
> order 22 (4096 kB objects)
> block_name_prefix: rbd_data.651bb238e1f29
> format: 2
> features: layering, exclusive-lock
> flags:


when we map the device in one server

rbd map app/test
> mount /dev/rbd0 /mnt/


a exclusive lock is released:

There is 1 exclusive lock on this image.

Locker        ID     Address
> client.554123 auto 1 10.0.128.19:0/3199254828


If we try to lock the image we recive:

rbd lock add app/test lock_test_20171219_1619

rbd: lock is already held by someone else


but if we try to map the same device on a second server:

rbd map app/test
> mount /dev/rbd0 /mnt/


the lock is overwritten:

There is 1 exclusive lock on this image.
> Locker        ID     Address
> client.534166 auto 1 10.0.128.18:0/3554787139



This is the output of rbd status app/test:

Watchers:
> watcher=10.0.128.18:0/3554787139 client.534166 cookie=1
> watcher=10.0.128.19:0/3199254828 client.554123 cookie=1


Both servers have full read/write access to the same image!
we're doing something wrong?

We are using Ceph jewel 10.0.10 (ceph version 10.2.10
(5dc1e4c05cb68dbf62ae6fce3f0700e4654fdbbe).

O.S: CentOS Linux release 7.3.1611 (Core) - Kernel Linux fred
3.10.0-693.11.1.el7.x86_64 #1 SMP Mon Dec 4 23:52:40 UTC 2017 x86_64 x86_64
x86_64 GNU/Linux

Thanks,
Lorenzo
-- 
Lorenzo Garuti
CED MaxMara
email: [email protected]
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to