Hi,

Why you would like to maintain copies by yourself. You replicate on ceph and then on different files inside ceph? Let ceph take care of counting. Create a pool with 3 or more copies and let ceph take care of what's stored and where.

Best regards,


El 13/07/17 a las 17:06, li...@marcelofrota.info escribió:

I will explain More about my system actual, in the moment i have 2 machines using drbd in mode master/slave and i running the aplication in machine master, but existing 2 questions importants in my enviroment with drbd actualy :

1 - If machine one is master and mounting partitions, the slave don't can mount the system, Unless it happens one problem in machine master, this is one mode, to prevent write in filesystem incorrect

2 - When i write data in machine master in drbd, the drbd write datas in slave machine Automatically, with this, if one problem happens in node master, the machine slave have coppy the data.

In the moment, in my enviroment testing with ceph, using the version 4.10 of kernel and i mount the system in two machines in the same time, in production enviroment, i could serious problem with this comportament.

How can i use the ceph and Ensure that I could get these 2 behaviors kept in a new environment with Ceph?

Thanks a lot,

Marcelo


Em 28/06/2017, Jason Dillaman <jdill...@redhat.com> escreveu:
> ... additionally, the forthcoming 4.12 kernel release will support
> non-cooperative exclusive locking. By default, since 4.9, when the
> exclusive-lock feature is enabled, only a single client can write to the > block device at a time -- but they will cooperatively pass the lock back > and forth upon write request. With the new "rbd map" option, you can map a > image on exactly one host and prevent other hosts from mapping the image.
> If that host should die, the exclusive-lock will automatically become
> available to other hosts for mapping.
>
> Of course, I always have to ask the use-case behind mapping the same image > on multiple hosts. Perhaps CephFS would be a better fit if you are trying
> to serve out a filesystem?
>
> On Wed, Jun 28, 2017 at 6:25 PM, Maged Mokhtar <mmokh...@petasan.org> wrote:
>
> > On 2017-06-28 22:55, li...@marcelofrota.info wrote:
> >
> > Hi People,
> >
> > I am testing the new enviroment, with ceph + rbd with ubuntu 16.04, and i
> > have one question.
> >
> > I have my cluster ceph and mount the using the comands to ceph in my linux
> > enviroment :
> >
> > rbd create veeamrepo --size 20480
> > rbd --image veeamrepo info
> > modprobe rbd
> > rbd map veeamrepo
> > rbd feature disable veeamrepo exclusive-lock object-map fast-diff
> > deep-flatten
> > mkdir /mnt/veeamrepo
> > mount /dev/rbd0 /mnt/veeamrepo
> >
> > The comands work fine, but i have one problem, in the moment, i can mount > > the /mnt/veeamrepo in the same time in 2 machines, and this is a bad option > > for me in the moment, because this could generate one filesystem corrupt.
> >
> > I need only one machine to be allowed to mount and write at a time.
> >
> > Example if machine1 mount the /mnt/veeamrepo and machine2 try mount, one > > error would be displayed, show message the machine can not mount, because
> > the system already mounted in machine1.
> >
> > Someone, could help-me with this or give some tips, for solution my
> > problem. ?
> >
> > Thanks a lot
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> >
> > You can use Pacemaker to map the rbd and mount the filesystem on 1 server
> > and in case of failure switch to another server.
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
>
>
> --
> Jason



_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to