Hi Tyler,
Thanks for your reply. We have disabled rbd_cache but still issue is
persist. Please find our configuration file.
# cat /etc/ceph/ceph.conf
[global]
fsid = 944fa0af-b7be-45a9-93ff-b9907cfaee3f
mon_initial_members = integ-hm5, integ-hm6, integ-hm7
mon_host = 192.168.112.192,192.168.112.193,192.168.112.194
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true
osd_pool_default_size = 2
[mon]
mon_clock_drift_allowed = .500
[client]
rbd_cache = false
--------------------------------------------------------------------------------------
cluster 944fa0af-b7be-45a9-93ff-b9907cfaee3f
health HEALTH_OK
monmap e2: 3 mons at
{integ-hm5=192.168.112.192:6789/0,integ-hm6=192.168.112.193:6789/0,integ-hm7=192.168.112.194:6789/0}
election epoch 480, quorum 0,1,2 integ-hm5,integ-hm6,integ-hm7
osdmap e49780: 2 osds: 2 up, 2 in
pgmap v2256565: 190 pgs, 2 pools, 1364 GB data, 410 kobjects
2559 GB used, 21106 GB / 24921 GB avail
190 active+clean
client io 373 kB/s rd, 13910 B/s wr, 103 op/s
Regards
Prabu
---- On Tue, 13 Oct 2015 19:59:38 +0530 Tyler Bishop
<[email protected]> wrote ----
You need to disable RBD caching.
Tyler Bishop
Chief Technical Officer
513-299-7108 x10
[email protected]
If you are not the intended recipient of this transmission you are notified
that disclosing, copying, distributing or taking any action in reliance on the
contents of this information is strictly prohibited.
From: "gjprabu" <[email protected]>
To: "Frédéric Nass" <[email protected]>
Cc: "<[email protected]>" <[email protected]>,
"Siva Sokkumuthu" <[email protected]>, "Kamal Kannan
Subramani(kamalakannan)" <[email protected]>
Sent: Tuesday, October 13, 2015 9:11:30 AM
Subject: Re: [ceph-users] ceph same rbd on multiple client
Hi ,
We have CEPH RBD with OCFS2 mounted servers. we are facing i/o errors
simultaneously while move the folder using one nodes in the same disk other
nodes data replicating with below said error (Copying is not having any
problem). Workaround if we remount the partition this issue get resolved but
after sometime problem again reoccurred. please help on this issue.
Note : We have total 5 Nodes, here two nodes working fine other nodes are
showing like below input/output error on moved data's.
ls -althr
ls: cannot access LITE_3_0_M4_1_TEST: Input/output error
ls: cannot access LITE_3_0_M4_1_OLD: Input/output error
total 0
d????????? ? ? ? ? ? LITE_3_0_M4_1_TEST
d????????? ? ? ? ? ? LITE_3_0_M4_1_OLD
Regards
Prabu
---- On Fri, 22 May 2015 17:33:04 +0530 Frédéric Nass
<[email protected]> wrote ----
Hi,
Waiting for CephFS, you can use clustered filesystem like OCFS2 or GFS2 on top
of RBD mappings so that each host can access the same device and clustered
filesystem.
Regards,
Frédéric.
Le 21/05/2015 16:10, gjprabu a écrit :
-- Frédéric Nass Sous direction des Infrastructures, Direction du Numérique,
Université de Lorraine. Tél : 03.83.68.53.83
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Hi All,
We are using rbd and map the same rbd image to the rbd device on two
different client but i can't see the data until i umount and mount -a
partition. Kindly share the solution for this issue.
Example
create rbd image named foo
map foo to /dev/rbd0 on server A, mount /dev/rbd0 to /mnt
map foo to /dev/rbd0 on server B, mount /dev/rbd0 to /mnt
Regards
Prabu
_______________________________________________ ceph-users mailing list
[email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com