Hi,
Why you would like to maintain copies by yourself. You replicate on ceph
and then on different files inside ceph? Let ceph take care of counting.
Create a pool with 3 or more copies and let ceph take care of what's
stored and where.
Best regards,
El 13/07/17 a las 17:06, li...@marcelofrota.info escribió:
I will explain More about my system actual, in the moment i have 2
machines using drbd in mode master/slave and i running the aplication
in machine master, but existing 2 questions importants in my
enviroment with drbd actualy :
1 - If machine one is master and mounting partitions, the slave don't
can mount the system, Unless it happens one problem in machine master,
this is one mode, to prevent write in filesystem incorrect
2 - When i write data in machine master in drbd, the drbd write datas
in slave machine Automatically, with this, if one problem happens in
node master, the machine slave have coppy the data.
In the moment, in my enviroment testing with ceph, using the version
4.10 of kernel and i mount the system in two machines in the same
time, in production enviroment, i could serious problem with this
comportament.
How can i use the ceph and Ensure that I could get these 2 behaviors
kept in a new environment with Ceph?
Thanks a lot,
Marcelo
Em 28/06/2017, Jason Dillaman <jdill...@redhat.com> escreveu:
> ... additionally, the forthcoming 4.12 kernel release will support
> non-cooperative exclusive locking. By default, since 4.9, when the
> exclusive-lock feature is enabled, only a single client can write to
the
> block device at a time -- but they will cooperatively pass the lock
back
> and forth upon write request. With the new "rbd map" option, you can
map a
> image on exactly one host and prevent other hosts from mapping the
image.
> If that host should die, the exclusive-lock will automatically become
> available to other hosts for mapping.
>
> Of course, I always have to ask the use-case behind mapping the same
image
> on multiple hosts. Perhaps CephFS would be a better fit if you are
trying
> to serve out a filesystem?
>
> On Wed, Jun 28, 2017 at 6:25 PM, Maged Mokhtar
<mmokh...@petasan.org> wrote:
>
> > On 2017-06-28 22:55, li...@marcelofrota.info wrote:
> >
> > Hi People,
> >
> > I am testing the new enviroment, with ceph + rbd with ubuntu
16.04, and i
> > have one question.
> >
> > I have my cluster ceph and mount the using the comands to ceph in
my linux
> > enviroment :
> >
> > rbd create veeamrepo --size 20480
> > rbd --image veeamrepo info
> > modprobe rbd
> > rbd map veeamrepo
> > rbd feature disable veeamrepo exclusive-lock object-map fast-diff
> > deep-flatten
> > mkdir /mnt/veeamrepo
> > mount /dev/rbd0 /mnt/veeamrepo
> >
> > The comands work fine, but i have one problem, in the moment, i
can mount
> > the /mnt/veeamrepo in the same time in 2 machines, and this is a
bad option
> > for me in the moment, because this could generate one filesystem
corrupt.
> >
> > I need only one machine to be allowed to mount and write at a time.
> >
> > Example if machine1 mount the /mnt/veeamrepo and machine2 try
mount, one
> > error would be displayed, show message the machine can not mount,
because
> > the system already mounted in machine1.
> >
> > Someone, could help-me with this or give some tips, for solution my
> > problem. ?
> >
> > Thanks a lot
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> >
> > You can use Pacemaker to map the rbd and mount the filesystem on 1
server
> > and in case of failure switch to another server.
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
>
>
> --
> Jason
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com