Is that correct or have you added more than 1 OSD? CEPH is never going to work or be able to bring up a pool with only one OSD, if you really do have more than OSD and have added them correctly then there really is something up with your CEPH setup / config and may be worth starting from scratch.
On Tue, Nov 6, 2018 at 4:31 PM Dengke Du <[email protected]> wrote: > > On 2018/11/6 下午4:29, Ashley Merrick wrote: > > What does > > "ceph osd tree" show ? > > root@node1:~# ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -2 0 host 0 > -1 1.00000 root default > -3 1.00000 host node1 > 0 hdd 1.00000 osd.0 down 0 1.00000 > > > On Tue, Nov 6, 2018 at 4:27 PM Dengke Du <[email protected]> wrote: > >> >> On 2018/11/6 下午4:24, Ashley Merrick wrote: >> >> If I am reading your ceph -s output correctly you only have 1 OSD, and 0 >> pool's created. >> >> So your be unable to create a RBD till you atleast have a pool setup and >> configured to create the RBD within. >> >> root@node1:~# ceph osd lspools >> 1 libvirt-pool >> 2 test-pool >> >> >> I create pools using: >> >> ceph osd pool create libvirt-pool 128 128 >> >> following: >> >> http://docs.ceph.com/docs/master/rbd/libvirt/ >> >> >> On Tue, Nov 6, 2018 at 4:21 PM Dengke Du <[email protected]> wrote: >> >>> >>> On 2018/11/6 下午4:16, Mykola Golub wrote: >>> > On Tue, Nov 06, 2018 at 09:45:01AM +0800, Dengke Du wrote: >>> > >>> >> I reconfigure the osd service from start, the journal was: >>> > I am not quite sure I understand what you mean here. >>> > >>> >> >>> ------------------------------------------------------------------------------------------------------------------------------------------ >>> >> >>> >> -- Unit [email protected] has finished starting up. >>> >> -- >>> >> -- The start-up result is RESULT. >>> >> Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 >>> 7f6a27204e80 >>> >> -1 Public network was set, but cluster network was not set >>> >> Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 >>> 7f6a27204e80 >>> >> -1 Using public network also for cluster network >>> >> Nov 05 18:02:36 node1 ceph-osd[4487]: starting osd.0 at - osd_data >>> >> /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal >>> >> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.365 >>> 7f6a27204e80 >>> >> -1 journal FileJournal::_open: disabling aio for non-block journal. >>> Use >>> >> journal_force_aio to force use of a> >>> >> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.414 >>> 7f6a27204e80 >>> >> -1 journal do_read_entry(6930432): bad header magic >>> >> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.729 >>> 7f6a27204e80 >>> >> -1 osd.0 21 log_to_monitors {default=true} >>> >> Nov 05 18:02:47 node1 nagios[3584]: Warning: Return code of 13 for >>> check of >>> >> host 'localhost' was out of bounds. >>> >> >>> >> >>> ------------------------------------------------------------------------------------------------------------------------------------------ >>> > Could you please post the full ceph-osd log somewhere? >>> /var/log/ceph/ceph-osd.0.log >>> >>> I don't have the file /var/log/ceph/ceph-osd.o.log >>> >>> root@node1:~# systemctl status ceph-osd@0 >>> ● [email protected] - Ceph object storage daemon osd.0 >>> Loaded: loaded (/lib/systemd/system/[email protected]; disabled; >>> vendor preset: enabled) >>> Active: active (running) since Mon 2018-11-05 18:02:36 UTC; 6h ago >>> Main PID: 4487 (ceph-osd) >>> Tasks: 64 >>> Memory: 27.0M >>> CGroup: /system.slice/system-ceph\x2dosd.slice/[email protected] >>> └─4487 /usr/bin/ceph-osd -f --cluster ceph --id 0 >>> >>> Nov 05 18:02:36 node1 systemd[1]: Starting Ceph object storage daemon >>> osd.0... >>> Nov 05 18:02:36 node1 systemd[1]: Started Ceph object storage daemon >>> osd.0. >>> Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 >>> 7f6a27204e80 -1 Public network was set, but cluster network was not set >>> Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 >>> 7f6a27204e80 -1 Using public network also for cluster network >>> Nov 05 18:02:36 node1 ceph-osd[4487]: starting osd.0 at - osd_data >>> /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal >>> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.365 >>> 7f6a27204e80 -1 journal FileJournal::_open: disabling aio for non-block >>> journal. Use journal_force_aio to force use of a> >>> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.414 >>> 7f6a27204e80 -1 journal do_read_entry(6930432): bad header magic >>> Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.729 >>> 7f6a27204e80 -1 osd.0 21 log_to_monitors {default=true} >>> >>> > >>> >> but hang at the command: "rbd create libvirt-pool/dimage --size 10240 >>> " >>> > So it hungs forever now instead of returning the error? >>> no returning any error, just hungs >>> > What is `ceph -s` output? >>> root@node1:~# ceph -s >>> cluster: >>> id: 9c1a42e1-afc2-4170-8172-96f4ebdaac68 >>> health: HEALTH_WARN >>> no active mgr >>> >>> services: >>> mon: 1 daemons, quorum 0 >>> mgr: no daemons active >>> osd: 1 osds: 0 up, 0 in >>> >>> data: >>> pools: 0 pools, 0 pgs >>> objects: 0 objects, 0 B >>> usage: 0 B used, 0 B / 0 B avail >>> pgs: >>> >>> >>> > >>> _______________________________________________ >>> ceph-users mailing list >>> [email protected] >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
