On 22/04/15 20:32, Robert LeBlanc wrote:
> I believe your problem is that you haven't created bootstrap-osd key
> and distributed it to your OSD node in /var/lib/ceph/bootstrap-osd/.

Hi Robert,

Thank you for your reply.

In my original post, steps performed, I did include copying over the
bootstrap-osd key. Also "ceph-disk activate" fails with an obvious error
when that file is missing:

2015-04-23 10:16:47.245951 7fccc5a9c700 -1 monclient(hunting): ERROR:
missing keyring, cannot use cephx for authentication
2015-04-23 10:16:47.245955 7fccc5a9c700  0 librados:
client.bootstrap-osd initialization error (2) No such file or directory
Error connecting to cluster: ObjectNotFound
ERROR:ceph-disk:Failed to activate
ceph-disk: Error: ceph osd create failed: Command '/usr/bin/ceph'
returned non-zero exit status 1:

This is not the source of my issue.

Dan

>
>
> On Wed, Apr 22, 2015 at 5:41 AM, Daniel Piddock
> <[email protected] <mailto:[email protected]>> wrote:
>
>     Hi,
>
>     I'm a ceph newbie setting up some trial installs for evaluation.
>
>     Using Debian stable (Wheezy) with Ceph Firefly from backports
>     (0.80.7-1~bpo70+1).
>
>     I've been following the instructions at
>     http://docs.ceph.com/docs/firefly/install/manual-deployment/ and first
>     time through went well, using a partition on the same drive as the
>     OS. I
>     then migrated to having data on separate harddrives and that
>     worked too.
>
>     I'm currently trying to get an OSD set up with the journal on an SSD
>     partition that's separate from the data drive. ceph-disk is not
>     playing
>     ball and I've been getting various forms of failure. My greatest
>     success
>     was getting the OSD created but it would never go "up". I'm struggling
>     to find anything useful in the logs or really what to look for.
>
>     I purged the ceph package and wiped the storage drives to give me a
>     blank slate and tried again.
>
>     Steps performed:
>
>     camel (MON server):
>     $ apt-get install ceph
>     $ uuidgen #= 8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2
>     # created /etc/ceph/ceph.conf, attached
>     $ ceph-authtool --create-keyring /tmp/ceph.mon.keyring --gen-key
>     -n mon. \
>       --cap mon 'allow *'
>     $ ceph-authtool --create-keyring /etc/ceph/ceph.client.admin.keyring
>     --gen-key \
>       -n client.admin --set-uid=0 --cap mon 'allow *' --cap osd 'allow
>     *' \
>       --cap mds 'allow'
>     $ ceph-authtool /tmp/ceph.mon.keyring --import-keyring \
>       /etc/ceph/ceph.client.admin.keyring
>     $ monmaptool --create --add a 10.1.0.3 --fsid \
>       8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 /tmp/monmap
>     $ ceph-mon --mkfs -i a --monmap /tmp/monmap --keyring
>     /tmp/ceph.mon.keyring
>     $ /etc/init.d/ceph start mon
>     $ ceph osd lspools #= 0 data,1 metadata,2 rbd,
>
>     storage node 1:
>     $ apt-get install ceph
>     $ rsync -a camel:/etc/ceph/ceph.conf /etc/ceph/
>     $ rsync -a camel:/var/lib/ceph/bootstrap-osd/ceph.keyring \
>       /var/lib/ceph/bootstrap-osd/
>     $ ceph-disk prepare --cluster ceph --cluster-uuid \
>       8c9ff7b5-904a-4f9a-8c9e-d2f8b05b55d2 /dev/sdb /dev/sdc
>
>     Output:
>     cannot read partition index; assume it isn't present
>      (Error: Command '/sbin/parted' returned non-zero exit status 1)
>     WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the
>     same device as the osd data
>     Creating new GPT entries.
>     Information: Moved requested sector from 34 to 2048 in
>     order to align on 2048-sector boundaries.
>     The operation has completed successfully.
>     Creating new GPT entries.
>     Information: Moved requested sector from 34 to 2048 in
>     order to align on 2048-sector boundaries.
>     The operation has completed successfully.
>     meta-data=/dev/sdb1              isize=2048   agcount=4,
>     agsize=15262347
>     blks
>              =                       sectsz=512   attr=2, projid32bit=0
>     data     =                       bsize=4096   blocks=61049385,
>     imaxpct=25
>              =                       sunit=0      swidth=0 blks
>     naming   =version 2              bsize=4096   ascii-ci=0
>     log      =internal log           bsize=4096   blocks=29809, version=2
>              =                       sectsz=512   sunit=0 blks,
>     lazy-count=1
>     realtime =none                   extsz=4096   blocks=0, rtextents=0
>     The operation has completed successfully.
>
>     $ ceph-disk activate /dev/sdb1
>     Hangs
>
>     Looking at ps -efH I can see that ceph-disk launched:
>     /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 0 --monmap
>     /var/lib/ceph/tmp/mnt.ST6Kz_/activate.monmap --osd-data
>     /var/lib/ceph/tmp/mnt.ST6Kz_ --osd-journal
>     /var/lib/ceph/tmp/mnt.ST6Kz_/journal --osd-uuid
>     636f694a-3677-44f0-baaf-4d74195b1806 --keyring
>     /var/lib/ceph/tmp/mnt.ST6Kz_/keyring
>
>     /var/lib/ceph/tmp/mnt.ST6Kz_ contains:
>     activate.monmap  current/ journal       magic          superblock
>     ceph_fsid        fsid     journal_uuid  store_version  whoami
>
>     journal is a symlink to
>     /dev/disk/by-partuuid/798fa1c5-9751-403c-9d5a-5f7665a60d4b (sdc1)
>
>     ceph osd tree:
>     # id    weight  type name       up/down reweight
>     -1      0       root default
>     0       0       osd.0   down    0
>
>     If I Ctrl-C ceph-disk, kill the ceph-osd process and try again it
>     still
>     hangs.
>
>     Please can somebody help? I've also attached the ceph-osd.0.log
>
>     Dan
>
>     _______________________________________________
>     ceph-users mailing list
>     [email protected] <mailto:[email protected]>
>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to