Paul, thanks for the reply, I try to run the command and then run again the ceph-deploy command again with the same parameters, but I'm getting the same error msg. I try with XFS and run ok, so the problem is with the btrfs
*German* 2015-11-20 6:08 GMT-03:00 HEWLETT, Paul (Paul) < [email protected]>: > Flushing a GPT partition table using dd does not work as the table is > duplicated at the end of the disk as well > > Use the sgdisk –Z command > > Paul > > From: ceph-users <[email protected]> on behalf of Mykola < > [email protected]> > Date: Thursday, 19 November 2015 at 18:43 > To: German Anders <[email protected]> > Cc: ceph-users <[email protected]> > Subject: Re: [ceph-users] ceph osd prepare cmd on infernalis 9.2.0 > > I believe the error message says that there is no space left on the device > for the second partition to be created. Perhaps try to flush gpt with old > good dd. > > > > Sent from Outlook Mail <http://go.microsoft.com/fwlink/?LinkId=550987> > for Windows 10 phone > > > > > *From: *German Anders <[email protected]> > *Sent: *Thursday, November 19, 2015 7:25 PM > *To: *Mykola Dvornik <[email protected]> > *Cc: *ceph-users <[email protected]> > *Subject: *Re: ceph osd prepare cmd on infernalis 9.2.0 > > > > I've already try that with no luck at all > > > > On Thursday, 19 November 2015, Mykola Dvornik <[email protected]> > wrote: > > *'Could not create partition 2 from 10485761 to 10485760'.* > > Perhaps try to zap the disks first? > > > > On 19 November 2015 at 16:22, German Anders <[email protected]> wrote: > > Hi cephers, > > I had some issues while running the prepare osd command: > > ceph version: infernalis 9.2.0 > > disk: /dev/sdf (745.2G) > > /dev/sdf1 740.2G > > /dev/sdf2 5G > > # parted /dev/sdf > GNU Parted 2.3 > Using /dev/sdf > Welcome to GNU Parted! Type 'help' to view a list of commands. > (parted) print > Model: ATA INTEL SSDSC2BB80 (scsi) > Disk /dev/sdf: 800GB > Sector size (logical/physical): 512B/4096B > Partition Table: gpt > > Number Start End Size File system Name Flags > 2 1049kB 5369MB 5368MB ceph journal > 1 5370MB 800GB 795GB btrfs ceph data > > cibn05: > > > > $ ceph-deploy osd prepare --fs-type btrfs cibn05:sdf > [ceph_deploy.conf][DEBUG ] found configuration file at: > /home/ceph/.cephdeploy.conf > [ceph_deploy.cli][INFO ] Invoked (1.5.28): /usr/local/bin/ceph-deploy osd > prepare --fs-type btrfs cibn05:sdf > [ceph_deploy.cli][INFO ] ceph-deploy options: > [ceph_deploy.cli][INFO ] username : None > [ceph_deploy.cli][INFO ] disk : [('cibn05', > '/dev/sdf', None)] > [ceph_deploy.cli][INFO ] dmcrypt : False > [ceph_deploy.cli][INFO ] verbose : False > [ceph_deploy.cli][INFO ] overwrite_conf : False > [ceph_deploy.cli][INFO ] subcommand : prepare > [ceph_deploy.cli][INFO ] dmcrypt_key_dir : > /etc/ceph/dmcrypt-keys > [ceph_deploy.cli][INFO ] quiet : False > [ceph_deploy.cli][INFO ] cd_conf : > <ceph_deploy.conf.cephdeploy.Conf instance at 0x7fbb1df85830> > [ceph_deploy.cli][INFO ] cluster : ceph > [ceph_deploy.cli][INFO ] fs_type : btrfs > [ceph_deploy.cli][INFO ] func : <function osd > at 0x7fbb1e1d9050> > [ceph_deploy.cli][INFO ] ceph_conf : None > [ceph_deploy.cli][INFO ] default_release : False > [ceph_deploy.cli][INFO ] zap_disk : False > [ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks cibn05:/dev/sdf: > [cibn05][DEBUG ] connection detected need for sudo > [cibn05][DEBUG ] connected to host: cibn05 > [cibn05][DEBUG ] detect platform information from remote host > [cibn05][DEBUG ] detect machine type > [cibn05][DEBUG ] find the location of an executable > [ceph_deploy.osd][INFO ] Distro info: Ubuntu 14.04 trusty > [ceph_deploy.osd][DEBUG ] Deploying osd to cibn05 > [cibn05][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf > [cibn05][INFO ] Running command: sudo udevadm trigger > --subsystem-match=block --action=add > [ceph_deploy.osd][DEBUG ] Preparing host cibn05 disk /dev/sdf journal None > activate False > [cibn05][INFO ] Running command: sudo ceph-disk -v prepare --cluster ceph > --fs-type btrfs -- /dev/sdf > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --check-allows-journal -i 0 --cluster ceph > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --check-wants-journal -i 0 --cluster ceph > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --check-needs-journal -i 0 --cluster ceph > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf uuid path is > /sys/dev/block/8:80/dm/uuid > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf uuid path is > /sys/dev/block/8:80/dm/uuid > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf uuid path is > /sys/dev/block/8:80/dm/uuid > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf1 uuid path is > /sys/dev/block/8:81/dm/uuid > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf2 uuid path is > /sys/dev/block/8:82/dm/uuid > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --cluster=ceph --show-config-value=fsid > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_mkfs_options_btrfs > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_fs_mkfs_options_btrfs > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_mount_options_btrfs > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_fs_mount_options_btrfs > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd > --cluster=ceph --show-config-value=osd_journal_size > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_cryptsetup_parameters > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_dmcrypt_key_size > [cibn05][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-conf > --cluster=ceph --name=osd. --lookup osd_dmcrypt_type > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf uuid path is > /sys/dev/block/8:80/dm/uuid > [cibn05][WARNIN] INFO:ceph-disk:Will colocate journal with data on /dev/sdf > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf uuid path is > /sys/dev/block/8:80/dm/uuid > [cibn05][WARNIN] DEBUG:ceph-disk:get_dm_uuid /dev/sdf uuid path is > /sys/dev/block/8:80/dm/uuid > [cibn05][WARNIN] DEBUG:ceph-disk:Creating journal partition num 2 size > 5120 on /dev/sdf > [cibn05][WARNIN] INFO:ceph-disk:Running command: /sbin/sgdisk > --new=2:0:5120M --change-name=2:ceph journal > --partition-guid=2:6a9a83f1-2196-4833-a4c8-8f3a424de54f > --typecode=2:45b0969e-9b03-4f30-b4c6-b4b80ceff106 --mbrtogpt -- /dev/sdf > [cibn05][WARNIN] Could not create partition 2 from 10485761 to 10485760 > [cibn05][WARNIN] Error encountered; not saving changes. > [cibn05][WARNIN] Traceback (most recent call last): > [cibn05][WARNIN] File "/usr/sbin/ceph-disk", line 3576, in <module> > [cibn05][WARNIN] main(sys.argv[1:]) > [cibn05][WARNIN] File "/usr/sbin/ceph-disk", line 3530, in main > [cibn05][WARNIN] args.func(args) > [cibn05][WARNIN] File "/usr/sbin/ceph-disk", line 1863, in main_prepare > [cibn05][WARNIN] luks=luks > [cibn05][WARNIN] File "/usr/sbin/ceph-disk", line 1465, in > prepare_journal > [cibn05][WARNIN] return prepare_journal_dev(data, journal, > journal_size, journal_uuid, journal_dm_keypath, cryptsetup_parameters, luks) > [cibn05][WARNIN] File "/usr/sbin/ceph-disk", line 1419, in > prepare_journal_dev > [cibn05][WARNIN] raise Error(e) > [cibn05][WARNIN] __main__.Error: Error: Command '['/sbin/sgdisk', > '--new=2:0:5120M', '--change-name=2:ceph journal', > '--partition-guid=2:6a9a83f1-2196-4833-a4c8-8f3a424de54f', > '--typecode=2:45b0969e-9b03-4f30-b4c6-b4b80ceff106', '--mbrtogpt', '--', > '/dev/sdf']' returned non-zero exit status 4 > [cibn05][ERROR ] RuntimeError: command returned non-zero exit status: 1 > [ceph_deploy.osd][ERROR ] Failed to execute command: ceph-disk -v prepare > --cluster ceph --fs-type btrfs -- /dev/sdf > [ceph_deploy][ERROR ] GenericError: Failed to create 1 OSDs > > any ideas? > > Thanks in advance, > > *German* > > > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > -- > > Mykola > > > > -- > > *German Anders* > Storage Engineer Manager > *Despegar* | IT Team > *office* +54 11 4894 3500 x3408 > *mobile* +54 911 3493 7262 > *mail* [email protected] > > > > >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
