Hi All,
I am running the MapReduce wordcount code (on a ceph cluster consisting of
2 VMs) on a data set consisting of 5000 odd files (approx. 10gb size in
total). Periodically, the ceph health says that the mds is
laggy/unresponsive, and I get messages like the following:
13/04/24 10:41:00 INFO mapred.JobClient: map 11% reduce 3%
13/04/24 10:42:36 INFO mapred.JobClient: map 12% reduce 3%
13/04/24 10:42:45 INFO mapred.JobClient: map 12% reduce 4%
13/04/24 10:44:08 INFO mapred.JobClient: map 13% reduce 4%
13/04/24 10:45:29 INFO mapred.JobClient: map 14% reduce 4%
13/04/24 11:06:31 INFO mapred.JobClient: Task Id :
attempt_201304241023_0001_m_000706_0, Status : FAILED
Task attempt_201304241023_0001_m_000706_0 failed to report status for 600
seconds. Killing!
Task attempt_201304241023_0001_m_000706_0 failed to report status for 600
seconds. Killing!
I then have to manually restart the mds again, and the process continues
execution. Can someone please tell me the reason for this, and how to solve
it? Pasting my ceph.conf file below:
[global]
auth client required = none
auth cluster required = none
auth service required = none
[osd]
osd journal data = 1000
filestore xattr use omap = true
# osd data = /var/lib/ceph/osd/ceph-$id
[mon.a]
host = varunc4-virtual-machine
mon addr = 10.72.148.209:6789
# mon data = /var/lib/ceph/mon/ceph-a
[mds.a]
host = varunc4-virtual-machine
# mds data = /var/lib/ceph/mds/ceph-a
[osd.0]
host = varunc4-virtual-machine
[osd.1]
host = varunc5-virtual-machine
Regards
Varun
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com