Hi all,
I've set up a new Ceph cluster for testing and it doesn't seem to be
working out-of-the-box. If I check the status it tells me that of the 3
defined OSD's, only 1 is in:
health HEALTH_WARN 392 pgs degraded; 392 pgs stuck unclean
> monmap e1: 3 mons at {controller-01=
> 10.20.3.110:6789/0,controller-02=10.20.3.111:6789/0,controller-03=10.20.3.112:6789/0},
> election epoch 6, quorum 0,1,2 controller-01,controller-02,controller-03
> osdmap e20: 3 osds: 1 up, 1 in
> pgmap v35: 392 pgs: 392 active+degraded; 0 bytes data, 37444 KB used,
> 15312 MB / 15348 MB avail
> mdsmap e1: 0/0/1 up
Turns out this is true because if I run "service ceph restart" on my OSD
nodes, osd.0 will restart just fine but osd.1 and osd.2 give me the
follwoing error:
Starting Ceph osd.0 on storage-02...
starting osd.1 at :/0 osd_data /var/lib/ceph/osd/ceph-1
> /var/lib/ceph/osd/ceph-1/journal
> 2013-07-09 11:54:26.497639 7f5b18813780 -1 accepter.accepter.bind unable
> to bind to 10.20.4.121:7100 on any port in range 6800-7100: Cannot assign
> requested address
> failed: 'ulimit -n 8192; /usr/bin/ceph-osd -i 1 --pid-file
> /var/run/ceph/osd.1.pid -c /etc/ceph/ceph.conf '
My ceph.conf just has a very limited configuration. The OSD section
basically contains:
[osd]
> public network=10.20.3.0/24
> cluster network=10.20.4.0/24
> [osd.0]
> host = storage-01
> public addr = 10.20.3.120
> cluster addr = 10.20.4.120
> [osd.1]
> host = storage-02
> public addr = 10.20.3.121
> cluster addr = 10.20.4.121
> [osd.2]
> host = storage-03
> public addr = 10.20.3.122
> cluster addr = 10.20.4.122
A quick Google search on that port binding error doesn't really yield and
results so I'm reaching out to you guys. Any thoughts on how to fix this?
Thanks,
Tom
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com