Dear friends:
Hello,I have a small problem When I use ceph . my ceph has three
monitor. I want to move out one.
root@node01 ~]# ceph -s
cluster b0d8bd0d-6269-4ce7-a10b-9adc7ee2c4c8
health HEALTH_WARN
too many PGs per OSD (682 > max 300)
monmap e23: 3 mons at
{node01=172.168.2.185:6789/0,node02=172.168.2.186:6789/0,node03=172.168.2.187:6789/0}
election epoch 472, quorum 0,1,2 node01,node02,node03
osdmap e7084: 18 osds: 18 up, 18 in
pgmap v1051011: 4448 pgs, 15 pools, 7915 MB data, 12834 objects
27537 MB used, 23298 GB / 23325 GB avail
4448 active+clean
So I do as this :
#ceph-deploy mon destroy node03
Then I add it in the cluster again.
#ceph-deploy mon add node03
The node03 is added to the cluster.but after a while,the monitor is down .
When I see the /var/log/messages
I find that
Apr 19 11:12:01 node01 systemd: Starting Session 14091 of user root.
Apr 19 11:12:01 node01 systemd: Started Session 14091 of user root.
Apr 19 11:12:39 node01 bash: 2016-04-19 11:12:39.533817 7f6e51ec2700 -1
mon.node01@0(leader) e23 *** Got Signal Terminated ***
When I start up the monitor ,then after a while it becomes down again.
But I have enough system space.
[root@node03 ~]# df -TH
Filesystem Type Size Used Avail Use% Mounted on
/dev/mapper/rhel-root xfs 11G 4.7G 6.1G 44% /
devtmpfs devtmpfs 26G 0 26G 0% /dev
tmpfs tmpfs 26G 82k 26G 1% /dev/shm
tmpfs tmpfs 26G 147M 26G 1% /run
tmpfs tmpfs 26G 0 26G 0% /sys/fs/cgroup
/dev/mapper/rhel-usr xfs 11G 4.1G 6.7G 38% /usr
/dev/mapper/rhel-tmp xfs 11G 34M 11G 1% /tmp
/dev/mapper/rhel-home xfs 11G 34M 11G 1% /home
/dev/mapper/rhel-var xfs 11G 1.6G 9.2G 15% /var
/dev/sde1 xfs 2.0T 152M 2.0T 1% /var/lib/ceph/osd/ceph-15
/dev/sdg1 xfs 2.0T 3.8G 2.0T 1% /var/lib/ceph/osd/ceph-17
/dev/sdd1 xfs 2.0T 165M 2.0T 1% /var/lib/ceph/osd/ceph-14
/dev/sda1 xfs 521M 131M 391M 26% /boot
/dev/sdb1 xfs 219G 989M 218G 1% /var/lib/ceph/osd/ceph-4
/dev/sdf1 xfs 2.0T 4.6G 2.0T 1% /var/lib/ceph/osd/ceph-16
/dev/sdc1 xfs 219G 129M 219G 1% /var/lib/ceph/osd/ceph-5
You have new mail in /var/spool/mail/root
[root@node03 ~]#
What’s the problem , is my operation wrong?
Looking forward to your reply.
--Dingxf48
发送自 Windows 10 版邮件应用
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com