Gonzalo,
 
You are right, i told so much about my enviroment actual and maybe i didn't 
know explain my problem the better form, with ceph in the moment, mutiple hosts 
clients can mount and write datas in my system and this is one problem, because 
i could have filesystem corruption.
 
Example, today, if runing the comand in two machines in the same time, it will 
work.
 
mount /dev/rbd0 /mnt/veeamrepo
cd /mnt/veeamrepo ; touch testfile.txt
 
I need ensure, only one machine will can execute this.
 
Thanks a lot,
Marcelo

Em 14/07/2017, Gonzalo Aguilar Delgado <gagui...@aguilardelgado.com>
escreveu:
> Hi, > > Why you would like to maintain copies by yourself. You replicate on ceph > and then on different files inside ceph? Let ceph take care of counting. > Create a pool with 3 or more copies and let ceph take care of what's > stored and where. > > Best regards, > > > El 13/07/17 a las 17:06, li...@marcelofrota.info escribió: > > > > I will explain More about my system actual, in the moment i have 2 > > machines using drbd in mode master/slave and i running the aplication > > in machine master, but existing 2 questions importants in my > > enviroment with drbd actualy : > > > > 1 - If machine one is master and mounting partitions, the slave don't > > can mount the system, Unless it happens one problem in machine master, > > this is one mode, to prevent write in filesystem incorrect > > > > 2 - When i write data in machine master in drbd, the drbd write datas > > in slave machine Automatically, with this, if one problem happens in > > node master, the machine slave have coppy the data. > > > > In the moment, in my enviroment testing with ceph, using the version > > 4.10 of kernel and i mount the system in two machines in the same > > time, in production enviroment, i could serious problem with this > > comportament. > > > > How can i use the ceph and Ensure that I could get these 2 behaviors > > kept in a new environment with Ceph? > > > > Thanks a lot, > > > > Marcelo > > > > > > Em 28/06/2017, Jason Dillaman <jdill...@redhat.com> escreveu: > > > ... additionally, the forthcoming 4.12 kernel release will support > > > non-cooperative exclusive locking. By default, since 4.9, when the > > > exclusive-lock feature is enabled, only a single client can write to > > the > > > block device at a time -- but they will cooperatively pass the lock > > back > > > and forth upon write request. With the new "rbd map" option, you can > > map a > > > image on exactly one host and prevent other hosts from mapping the > > image. > > > If that host should die, the exclusive-lock will automatically become > > > available to other hosts for mapping. > > > > > > Of course, I always have to ask the use-case behind mapping the same > > image > > > on multiple hosts. Perhaps CephFS would be a better fit if you are > > trying > > > to serve out a filesystem? > > > > > > On Wed, Jun 28, 2017 at 6:25 PM, Maged Mokhtar > > <mmokh...@petasan.org> wrote: > > > > > > > On 2017-06-28 22:55, li...@marcelofrota.info wrote: > > > > > > > > Hi People, > > > > > > > > I am testing the new enviroment, with ceph + rbd with ubuntu > > 16.04, and i > > > > have one question. > > > > > > > > I have my cluster ceph and mount the using the comands to ceph in > > my linux > > > > enviroment : > > > > > > > > rbd create veeamrepo --size 20480 > > > > rbd --image veeamrepo info > > > > modprobe rbd > > > > rbd map veeamrepo > > > > rbd feature disable veeamrepo exclusive-lock object-map fast-diff > > > > deep-flatten > > > > mkdir /mnt/veeamrepo > > > > mount /dev/rbd0 /mnt/veeamrepo > > > > > > > > The comands work fine, but i have one problem, in the moment, i > > can mount > > > > the /mnt/veeamrepo in the same time in 2 machines, and this is a > > bad option > > > > for me in the moment, because this could generate one filesystem > > corrupt. > > > > > > > > I need only one machine to be allowed to mount and write at a time. > > > > > > > > Example if machine1 mount the /mnt/veeamrepo and machine2 try > > mount, one > > > > error would be displayed, show message the machine can not mount, > > because > > > > the system already mounted in machine1. > > > > > > > > Someone, could help-me with this or give some tips, for solution my > > > > problem. ? > > > > > > > > Thanks a lot > > > > > > > > _______________________________________________ > > > > ceph-users mailing list > > > > ceph-users@lists.ceph.com > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > > > > > > > You can use Pacemaker to map the rbd and mount the filesystem on 1 > > server > > > > and in case of failure switch to another server. > > > > > > > > > > > > _______________________________________________ > > > > ceph-users mailing list > > > > ceph-users@lists.ceph.com > > > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > > > > > > > > -- > > > Jason > > > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > ---- > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to