I ended up starting from scratch and doing a purge and purgedata on that host using ceph-deploy, after that things seemed to go better. The osd is up and in at this point, however when the osd was added to the cluster...no data was being moved to the new osd.

Here is a copy of my current crush map:

http://pastebin.com/PMk3xZ0a

as you can see from the entry for osd number 108 (the last osd to be added to the cluster)...the crush map does not contain a host entry for hqosd10...which is the host for osd #108.

Any ideas on how to resolve this?

Thanks,
Shain


On 9/8/16 2:20 PM, Shain Miley wrote:
Hello,

I am trying to use ceph-deploy to add some new osd's to our cluster. I have used this method over the last few years to add all of our 107 osd's and things have seemed to work quite well.

One difference this time is that we are going to use a pci nvme card to journal the 16 disks in this server (Dell R730xd).

As you can see below it appears as though things complete successfully, however the osd count never increases, and when I look at hqosd10, there are no osd's mounted, and nothing in '/var/lib/ceph/osd', no ceph daemons running, etc.

I created the partitions on the nvme card by hand using parted (I was not sure if I ceph-deploy should take care of this part or not).

I have zapped the disk and re-run this command several times, and I have gotten the same result every time.

We are running Ceph version 0.94.9  on Ubuntu 14.04.5

Here is the output from my attempt:

root@hqceph1:/usr/local/ceph-deploy# ceph-deploy --verbose osd create hqosd10:sdb:/dev/nvme0n1p1 [ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf [ceph_deploy.cli][INFO ] Invoked (1.5.36): /usr/local/bin/ceph-deploy --verbose osd create hqosd10:sdb:/dev/nvme0n1p1
[ceph_deploy.cli][INFO  ] ceph-deploy options:
[ceph_deploy.cli][INFO  ]  username                      : None
[ceph_deploy.cli][INFO ] disk : [('hqosd10', '/dev/sdb', '/dev/nvme0n1p1')]
[ceph_deploy.cli][INFO  ]  dmcrypt                       : False
[ceph_deploy.cli][INFO  ]  verbose                       : True
[ceph_deploy.cli][INFO  ]  bluestore                     : None
[ceph_deploy.cli][INFO  ]  overwrite_conf                : False
[ceph_deploy.cli][INFO  ]  subcommand                    : create
[ceph_deploy.cli][INFO ] dmcrypt_key_dir : /etc/ceph/dmcrypt-keys
[ceph_deploy.cli][INFO  ]  quiet                         : False
[ceph_deploy.cli][INFO ] cd_conf : <ceph_deploy.conf.cephdeploy.Conf instance at 0x7f6ba74d01b8>
[ceph_deploy.cli][INFO  ]  cluster                       : ceph
[ceph_deploy.cli][INFO  ]  fs_type                       : xfs
[ceph_deploy.cli][INFO ] func : <function osd at 0x7f6ba750cc80>
[ceph_deploy.cli][INFO  ]  ceph_conf                     : None
[ceph_deploy.cli][INFO  ]  default_release               : False
[ceph_deploy.cli][INFO  ]  zap_disk                      : False
[ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks hqosd10:/dev/sdb:/dev/nvme0n1p1
[hqosd10][DEBUG ] connected to host: hqosd10
[hqosd10][DEBUG ] detect platform information from remote host
[hqosd10][DEBUG ] detect machine type
[hqosd10][DEBUG ] find the location of an executable
[hqosd10][INFO  ] Running command: /sbin/initctl version
[hqosd10][DEBUG ] find the location of an executable
[ceph_deploy.osd][INFO  ] Distro info: Ubuntu 14.04 trusty
[ceph_deploy.osd][DEBUG ] Deploying osd to hqosd10
[hqosd10][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph_deploy.osd][DEBUG ] Preparing host hqosd10 disk /dev/sdb journal /dev/nvme0n1p1 activate True
[hqosd10][DEBUG ] find the location of an executable
[hqosd10][INFO ] Running command: /usr/sbin/ceph-disk -v prepare --cluster ceph --fs-type xfs -- /dev/sdb /dev/nvme0n1p1 [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=fsid [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_mkfs_options_xfs [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_mount_options_xfs [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster=ceph --show-config-value=osd_journal_size [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_cryptsetup_parameters [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_dmcrypt_key_size [hqosd10][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf --cluster=ceph --name=osd. --lookup osd_dmcrypt_type
[hqosd10][WARNIN] DEBUG:ceph-disk:Journal /dev/nvme0n1p1 is a partition
[hqosd10][WARNIN] WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the same device as the osd data [hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/blkid -p -o udev /dev/nvme0n1p1 [hqosd10][WARNIN] WARNING:ceph-disk:Journal /dev/nvme0n1p1 was not prepared with ceph-disk. Symlinking directly.
[hqosd10][WARNIN] DEBUG:ceph-disk:Creating osd partition on /dev/sdb
[hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/sgdisk --largest-new=1 --change-name=1:ceph data --partition-guid=1:1541833e-1513-4446-9779-7dcb61a95a07 --typecode=1:89c57f98-2fe5-4dc0-89c1-f3ad0ceff2be -- /dev/sdb
[hqosd10][DEBUG ] The operation has completed successfully.
[hqosd10][WARNIN] DEBUG:ceph-disk:Calling partprobe on created device /dev/sdb [hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/partprobe /dev/sdb
[hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/udevadm settle
[hqosd10][WARNIN] DEBUG:ceph-disk:Creating xfs fs on /dev/sdb1
[hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/mkfs -t xfs -f -i size=2048 -f -- /dev/sdb1 [hqosd10][DEBUG ] meta-data=/dev/sdb1 isize=2048 agcount=32, agsize=45780984 blks [hqosd10][DEBUG ] = sectsz=4096 attr=2, projid32bit=0 [hqosd10][DEBUG ] data = bsize=4096 blocks=1464991483, imaxpct=5
[hqosd10][DEBUG ]          =                       sunit=0 swidth=0 blks
[hqosd10][DEBUG ] naming   =version 2              bsize=4096 ascii-ci=0
[hqosd10][DEBUG ] log =internal log bsize=4096 blocks=521728, version=2 [hqosd10][DEBUG ] = sectsz=4096 sunit=1 blks, lazy-count=1 [hqosd10][DEBUG ] realtime =none extsz=4096 blocks=0, rtextents=0 [hqosd10][WARNIN] DEBUG:ceph-disk:Mounting /dev/sdb1 on /var/lib/ceph/tmp/mnt.zL83i_ with options rw,noatime,nodiratime,logbsize=256k,logbufs=8,inode64 [hqosd10][WARNIN] INFO:ceph-disk:Running command: /bin/mount -t xfs -o rw,noatime,nodiratime,logbsize=256k,logbufs=8,inode64 -- /dev/sdb1 /var/lib/ceph/tmp/mnt.zL83i_ [hqosd10][WARNIN] DEBUG:ceph-disk:Preparing osd data dir /var/lib/ceph/tmp/mnt.zL83i_ [hqosd10][WARNIN] DEBUG:ceph-disk:Creating symlink /var/lib/ceph/tmp/mnt.zL83i_/journal -> /dev/nvme0n1p1
[hqosd10][WARNIN] DEBUG:ceph-disk:Unmounting /var/lib/ceph/tmp/mnt.zL83i_
[hqosd10][WARNIN] INFO:ceph-disk:Running command: /bin/umount -- /var/lib/ceph/tmp/mnt.zL83i_ [hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/sgdisk --typecode=1:4fbd7e29-9d25-41b8-afd0-062c0ceff05d -- /dev/sdb
[hqosd10][DEBUG ] The operation has completed successfully.
[hqosd10][WARNIN] DEBUG:ceph-disk:Calling partprobe on prepared device /dev/sdb [hqosd10][WARNIN] INFO:ceph-disk:Running command: /sbin/partprobe /dev/sdb
[hqosd10][INFO  ] checking OSD status...
[hqosd10][DEBUG ] find the location of an executable
[hqosd10][INFO ] Running command: /usr/bin/ceph --cluster=ceph osd stat --format=json
[ceph_deploy.osd][DEBUG ] Host hqosd10 is now ready for osd use.


Thanks,

Shain

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to