Hi,

I am setting up a local instance of ceph cluster with latest source from
git hub. The build succeeded and installation was successful, But I could
not start the monitor.

The "ceph start" command returns immediately and does not output anything.

$ sudo /etc/init.d/ceph start mon.monitor1

$

$ ls -l /var/lib/ceph/mon/ceph-monitor1/

total 8

-rw-r--r-- 1 root root    0 May  7 20:27 done

-rw-r--r-- 1 root root   77 May  7 19:12 keyring

drwxr-xr-x 2 root root 4096 May  7 19:12 store.db

-rw-r--r-- 1 root root    0 May  7 20:26 sysvinit

-rw-r--r-- 1 root root    0 May  7 20:09 upstart



The log filed does not seem to have any details either


$ cat /var/log/ceph/ceph-mon.monitor1.log

2015-05-07 19:12:13.356389 7f3f06bdb880 -1 did not load config file, using
default settings.


$ cat /etc/ceph/ceph.conf

[global]

mon host = 15.43.33.21

fsid = 92f859df-8b27-466a-8d44-01af2b7ea7e6

mon initial members = monitor1


# Enable authentication

auth cluster required = cephx

auth service required = cephx

auth client required = cephx


# POOL / PG / CRUSH

osd pool default size = 3  # Write an object 3 times

osd pool default min size = 1 # Allow writing one copy in a degraded state


# Ensure you have a realistic number of placement groups. We recommend

# approximately 200 per OSD. E.g., total number of OSDs multiplied by 200

# divided by the number of replicas (i.e., osd pool default size).

# !! BE CAREFULL !!

# You properly should never rely on the default numbers when creating pool!

osd pool default pg num = 32

osd pool default pgp num = 32


#log file = /home/y/logs/ceph/$cluster-$type.$id.log


# Logging

debug paxos = 0

debug throttle = 0


keyring = /etc/ceph/ceph.client.admin.keyring

#run dir = /home/y/var/run/ceph


[mon]

debug mon = 10

debug ms = 1

# We found that when the disk usage reach to 94%, the disk could not be
written

# any file (no free space), so that we lower the full ratio and we should
start

# data migration before it becomes full

mon osd full ratio = 0.9

#mon data = /home/y/var/lib/ceph/mon/$cluster-$id

mon osd down out interval = 172800 # 2 * 24 * 60 * 60 seconds

# Ceph monitors need to be told how many reporters must to be seen from
different

# OSDs before it can be marked offline, this should be greater than the
number of

# OSDs per OSD host

mon osd min down reporters = 12

#keyring = /home/y/conf/ceph/ceph.mon.keyring
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to