Random theory... I just noticed that the ceph-osd's are listed twice
[1] in the output of systemctl list-dependencies.

Is that correct?!!!

-- dan

[1] > systemctl list-dependencies
...
● ├─ceph-mds.target
● ├─ceph-mon.target
● ├─ceph-osd.target
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ ├─[email protected]
● │ └─[email protected]
● ├─ceph.target
● │ ├─ceph-mds.target
● │ ├─ceph-mon.target
● │ └─ceph-osd.target
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   ├─[email protected]
● │   └─[email protected]
● ├─getty.target
...



On Thu, Aug 31, 2017 at 4:57 PM, Dan van der Ster <[email protected]> wrote:
> Hi,
>
> I see the same with jewel on el7 -- it started one of the recent point
> releases around ~10.2.5, IIRC.
>
> Problem seems to be the same -- daemon is started before the osd is
> mounted... then the service waits several seconds before trying again.
>
> Aug 31 15:41:47 ceph-osd: 2017-08-31 15:41:47.267661 7f2e49731800 -1
> #033[0;31m ** ERROR: unable to open OSD superblock on
> /var/lib/ceph/osd/ceph-89: (2) No such file or directory#033[0m
> Aug 31 15:41:47 ceph-osd: starting osd.55 at :/0 osd_data
> /var/lib/ceph/osd/ceph-55 /var/lib/ceph/osd/ceph-55/journal
> Aug 31 15:41:47 systemd: [email protected]: main process exited,
> code=exited, status=1/FAILURE
> Aug 31 15:41:47 systemd: Unit [email protected] entered failed state.
> Aug 31 15:41:47 systemd: [email protected] failed.
> Aug 31 15:41:47 kernel: XFS (sdi1): Ending clean mount
> Aug 31 15:41:47 rc.local: Removed symlink
> /etc/systemd/system/ceph-osd.target.wants/[email protected].
> Aug 31 15:41:47 systemd: Reloading.
> Aug 31 15:41:47 systemd: Reloading.
> Aug 31 15:41:47 rc.local: Created symlink from
> /etc/systemd/system/ceph-osd.target.wants/[email protected] to
> /usr/lib/systemd/system/[email protected].
> Aug 31 15:41:47 systemd: Reloading.
> Aug 31 15:41:55 ceph-osd: 2017-08-31 15:41:55.425566 7f74b92e1800 -1
> osd.55 123659 log_to_monitors {default=true}
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
> Aug 31 15:42:07 systemd: [email protected] holdoff time over,
> scheduling restart.
>
>
> -- Dan
>
>
>
> On Thu, Aug 31, 2017 at 4:24 PM, Piotr Dzionek <[email protected]> wrote:
>> Hi,
>>
>> For a last 3 weeks I have been running latest LTS Luminous Ceph release on
>> CentOS7. It started with 4th RC and now I have Stable Release.
>> Cluster runs fine, however I noticed that if I do a reboot of one the nodes,
>> it takes a really long time for cluster to be in ok status.
>> Osds are starting up, but not as soon as the server is up. They are up one
>> by one during a period of 5 minutes. I checked the logs and all osds have
>> following errors.
>>
>> 2017-08-30 15:27:52.541366 7f7dabd0d700 30 Event(0x7f7dbc9f4a80 nevent=5000
>> time_id=62).process_events event_wq process is 11 mask is 1
>> 2017-08-30 15:51:03.639222 7faf11c3ed00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:51:03.639342 7faf11c3ed00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 3037
>> 2017-08-30 15:51:03.672898 7faf11c3ed00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:51:42.453334 7f9f55f11d00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:51:42.453352 7f9f55f11d00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 7366
>> 2017-08-30 15:51:42.453590 7f9f55f11d00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:52:03.199062 7effa00cad00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:52:03.199081 7effa00cad00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 7747
>> 2017-08-30 15:52:03.199323 7effa00cad00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:52:23.967466 7ff008c2cd00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:52:23.967483 7ff008c2cd00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 8016
>> 2017-08-30 15:52:23.967714 7ff008c2cd00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:52:44.716646 7fc2bd322d00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:52:44.716664 7fc2bd322d00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 8808
>> 2017-08-30 15:52:44.716892 7fc2bd322d00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:53:06.214611 7f4583e70d00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:53:06.214629 7f4583e70d00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 9184
>> 2017-08-30 15:53:06.214855 7f4583e70d00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:53:26.955944 7f1dfea39d00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:53:26.955962 7f1dfea39d00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 9417
>> 2017-08-30 15:53:26.956191 7f1dfea39d00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:53:47.714131 7fabbc5cfd00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:53:47.714149 7fabbc5cfd00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 10469
>> 2017-08-30 15:53:47.714383 7fabbc5cfd00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:54:08.708602 7f32fae4fd00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:54:08.708616 7f32fae4fd00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 11152
>> 2017-08-30 15:54:08.708898 7f32fae4fd00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:54:29.454000 7faea1004d00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:54:29.454016 7faea1004d00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 11751
>> 2017-08-30 15:54:29.454243 7faea1004d00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:54:50.207237 7f0c1701ed00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:54:50.207253 7f0c1701ed00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 12878
>> 2017-08-30 15:54:50.207431 7f0c1701ed00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:55:11.237106 7f571b1ffd00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:55:11.237122 7f571b1ffd00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 13542
>> 2017-08-30 15:55:11.237363 7f571b1ffd00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:55:34.709073 7f672394dd00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:55:34.709089 7f672394dd00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 14717
>> 2017-08-30 15:55:34.709330 7f672394dd00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:56:13.431601 7f4ace0bad00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:56:13.431616 7f4ace0bad00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 16148
>> 2017-08-30 15:56:13.431881 7f4ace0bad00 -1 ESC[0;31m ** ERROR: unable to
>> open OSD superblock on /var/lib/ceph/osd/ceph-27: (2) No such file or
>> directoryESC[0m
>> 2017-08-30 15:56:14.565304 7fa09c9f1d00  0 set uid:gid to 167:167
>> (ceph:ceph)
>> 2017-08-30 15:56:14.565320 7fa09c9f1d00  0 ceph version 12.2.0
>> (32ce2a3ae5239ee33d6150705cdb24d43bab910c) luminous (rc), process (unknown),
>> pid 16384
>>
>>
>> As you can see the xfs volume(the part with meta-data) is not mounted yet.
>> My question here, what mounts it and why it takes so long ? Maybe there is a
>> setting that randomizes the start up process of osds running on the same
>> node?
>>
>> Kind regards,
>> Piotr Dzionek
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to