If I am reading your ceph -s output correctly you only have 1 OSD, and 0 pool's created.
So your be unable to create a RBD till you atleast have a pool setup and configured to create the RBD within. On Tue, Nov 6, 2018 at 4:21 PM Dengke Du <[email protected]> wrote: > > On 2018/11/6 下午4:16, Mykola Golub wrote: > > On Tue, Nov 06, 2018 at 09:45:01AM +0800, Dengke Du wrote: > > > >> I reconfigure the osd service from start, the journal was: > > I am not quite sure I understand what you mean here. > > > >> > ------------------------------------------------------------------------------------------------------------------------------------------ > >> > >> -- Unit [email protected] has finished starting up. > >> -- > >> -- The start-up result is RESULT. > >> Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 > 7f6a27204e80 > >> -1 Public network was set, but cluster network was not set > >> Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 > 7f6a27204e80 > >> -1 Using public network also for cluster network > >> Nov 05 18:02:36 node1 ceph-osd[4487]: starting osd.0 at - osd_data > >> /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal > >> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.365 > 7f6a27204e80 > >> -1 journal FileJournal::_open: disabling aio for non-block journal. Use > >> journal_force_aio to force use of a> > >> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.414 > 7f6a27204e80 > >> -1 journal do_read_entry(6930432): bad header magic > >> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.729 > 7f6a27204e80 > >> -1 osd.0 21 log_to_monitors {default=true} > >> Nov 05 18:02:47 node1 nagios[3584]: Warning: Return code of 13 for > check of > >> host 'localhost' was out of bounds. > >> > >> > ------------------------------------------------------------------------------------------------------------------------------------------ > > Could you please post the full ceph-osd log somewhere? > /var/log/ceph/ceph-osd.0.log > > I don't have the file /var/log/ceph/ceph-osd.o.log > > root@node1:~# systemctl status ceph-osd@0 > ● [email protected] - Ceph object storage daemon osd.0 > Loaded: loaded (/lib/systemd/system/[email protected]; disabled; > vendor preset: enabled) > Active: active (running) since Mon 2018-11-05 18:02:36 UTC; 6h ago > Main PID: 4487 (ceph-osd) > Tasks: 64 > Memory: 27.0M > CGroup: /system.slice/system-ceph\x2dosd.slice/[email protected] > └─4487 /usr/bin/ceph-osd -f --cluster ceph --id 0 > > Nov 05 18:02:36 node1 systemd[1]: Starting Ceph object storage daemon > osd.0... > Nov 05 18:02:36 node1 systemd[1]: Started Ceph object storage daemon osd.0. > Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 > 7f6a27204e80 -1 Public network was set, but cluster network was not set > Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 > 7f6a27204e80 -1 Using public network also for cluster network > Nov 05 18:02:36 node1 ceph-osd[4487]: starting osd.0 at - osd_data > /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal > Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.365 > 7f6a27204e80 -1 journal FileJournal::_open: disabling aio for non-block > journal. Use journal_force_aio to force use of a> > Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.414 > 7f6a27204e80 -1 journal do_read_entry(6930432): bad header magic > Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.729 > 7f6a27204e80 -1 osd.0 21 log_to_monitors {default=true} > > > > >> but hang at the command: "rbd create libvirt-pool/dimage --size 10240 " > > So it hungs forever now instead of returning the error? > no returning any error, just hungs > > What is `ceph -s` output? > root@node1:~# ceph -s > cluster: > id: 9c1a42e1-afc2-4170-8172-96f4ebdaac68 > health: HEALTH_WARN > no active mgr > > services: > mon: 1 daemons, quorum 0 > mgr: no daemons active > osd: 1 osds: 0 up, 0 in > > data: > pools: 0 pools, 0 pgs > objects: 0 objects, 0 B > usage: 0 B used, 0 B / 0 B avail > pgs: > > > > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
