On freshly installed ubuntu 16.04 servers with the HWE kernel selected (4.10). I can not use ceph-deploy or ceph-disk to provision osd.
whenever I try I get the following:: ceph-disk -v prepare --dmcrypt --dmcrypt-key-dir /etc/ceph/dmcrypt-keys --bluestore --cluster ceph --fs-type xfs -- /dev/sdy command: Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=fsid get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid set_type: Will colocate block with data on /dev/sdy command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup bluestore_block_size [command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup bluestore_block_db_size command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup bluestore_block_size command: Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup bluestore_block_wal_size get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid get_dm_uuid: get_dm_uuid /dev/sdy uuid path is /sys/dev/block/65:128/dm/uuid Traceback (most recent call last): File "/usr/sbin/ceph-disk", line 9, in <module> load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')() File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5704, in run main(sys.argv[1:]) File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5655, in main args.func(args) File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2091, in main Prepare.factory(args).prepare() File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2080, in prepare self._prepare() File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2154, in _prepare self.lockbox.prepare() File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 2842, in prepare verify_not_in_use(self.args.lockbox, check_partitions=True) File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 950, in verify_not_in_use raise Error('Device is mounted', partition) ceph_disk.main.Error: Error: Device is mounted: /dev/sdy5 unmounting the disk does not seem to help either. I'm assuming something is triggering too early but i'm not sure how to delay or figure that out. has anyone deployed on xenial with the 4.10 kernel? Am I missing something important?
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com