Hi Alfredo and folk,
  Could you have a look at this?
  Someone else has any idea why i am getting this error?

Thanks in advance, I



2014-06-27 16:37 GMT+02:00 Iban Cabrillo <[email protected]>:

> Hi Alfredo,
>  This is the complete procedure:
>
>
>   On OSD node:
>
> [ceph@ceph02 ~]$ sudo parted /dev/xvdb
>
> GNU Parted 2.1
> Using /dev/xvdb
> Welcome to GNU Parted! Type 'help' to view a list of commands.
> (parted) p
> Model: Xen Virtual Block Device (xvd)
> Disk /dev/xvdb: 107GB
> Sector size (logical/physical): 512B/512B
> Partition Table: gpt
>
> Number  Start  End  Size  File system  Name  Flags
>
> [ceph@ceph02 ~]$ sudo ls -la /var/lib/ceph/tmp/
> total 8
> drwxr-xr-x 2 root root 4096 Jun 27 16:30 .
> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
> [ceph@ceph02 ~]$ sudo ls -la /var/lib/ceph/osd/
> total 8
> drwxr-xr-x 2 root root 4096 Jun 27 12:14 .
> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
>
> On ceph admin node:
>
> [ceph@cephadm ~]$ sudo ceph osd tree
> # id    weight    type name    up/down    reweight
> -1    0.14    root default
> -2    0.009995        host ceph02
> 1    0.009995            osd.1    DNE
> -3    0.03999        host ceph04
> 4    0.03999            osd.4    up    1
> -4    0.09        host ceph03
> 6    0.09            osd.6    up    1
>
>
> [ceph@cephadm ceph-cloud]$ ceph-deploy osd prepare ceph02:xvdb
> [ceph_deploy.conf][DEBUG ] found configuration file at:
> /home/ceph/.cephdeploy.conf
> [ceph_deploy.cli][INFO  ] Invoked (1.5.5): /usr/bin/ceph-deploy osd
> prepare ceph02:xvdb
> [ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks ceph02:/dev/xvdb:
> [ceph02][DEBUG ] connected to host: ceph02
> [ceph02][DEBUG ] detect platform information from remote host
> [ceph02][DEBUG ] detect machine type
> [ceph_deploy.osd][INFO  ] Distro info: Scientific Linux 6.2 Carbon
> [ceph_deploy.osd][DEBUG ] Deploying osd to ceph02
> [ceph02][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
> [ceph02][INFO  ] Running command: sudo udevadm trigger
> --subsystem-match=block --action=add
> [ceph_deploy.osd][DEBUG ] Preparing host ceph02 disk /dev/xvdb journal
> None activate False
> [ceph02][INFO  ] Running command: sudo ceph-disk-prepare --fs-type xfs
> --cluster ceph -- /dev/xvdb
> [ceph02][DEBUG ] Setting name!
> [ceph02][DEBUG ] partNum is 1
> [ceph02][DEBUG ] REALLY setting name!
> [ceph02][DEBUG ] The operation has completed successfully.
> [ceph02][DEBUG ] Setting name!
> [ceph02][DEBUG ] partNum is 0
> [ceph02][DEBUG ] REALLY setting name!
> [ceph02][DEBUG ] The operation has completed successfully.
> [ceph02][DEBUG ] meta-data=/dev/xvdb1             isize=2048   agcount=4,
> agsize=5897919 blks
> [ceph02][DEBUG ]          =                       sectsz=512   attr=2
> [ceph02][DEBUG ] data     =                       bsize=4096
> blocks=23591675, imaxpct=25
> [ceph02][DEBUG ]          =                       sunit=0      swidth=0
> blks
> [ceph02][DEBUG ] naming   =version 2              bsize=4096   ascii-ci=0
> [ceph02][DEBUG ] log      =internal log           bsize=4096
> blocks=11519, version=2
> [ceph02][DEBUG ]          =                       sectsz=512   sunit=0
> blks, lazy-count=1
> [ceph02][DEBUG ] realtime =none                   extsz=4096   blocks=0,
> rtextents=0
> [ceph02][DEBUG ] The operation has completed successfully.
> [ceph02][WARNIN] INFO:ceph-disk:Will colocate journal with data on
> /dev/xvdb
> [ceph02][INFO  ] checking OSD status...
> [ceph02][INFO  ] Running command: sudo ceph --cluster=ceph osd stat
> --format=json
> [ceph_deploy.osd][DEBUG ] Host ceph02 is now ready for osd use.
>
>
>
> If i make create instead of prepare do the same (create do not make the
> trick prepare+activate )
>
>
> In the OSD:
>
> [ceph@ceph02 ~]$ sudo parted /dev/xvdb
>
> GNU Parted 2.1
> Using /dev/xvdb
> Welcome to GNU Parted! Type 'help' to view a list of commands.
> (parted) p
> Model: Xen Virtual Block Device (xvd)
> Disk /dev/xvdb: 107GB
> Sector size (logical/physical): 512B/512B
> Partition Table: gpt
>
> Number  Start   End     Size    File system  Name          Flags
>  2      1049kB  10.7GB  10.7GB  xfs          ceph journal
>  1      10.7GB  107GB   96.6GB  xfs          ceph data
>
> (parted) q
>
> [ceph@ceph02 ~]$ sudo ls -la /var/lib/ceph/osd/
> total 8
> drwxr-xr-x 2 root root 4096 Jun 27 12:14 .
> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
> [ceph@ceph02 ~]$ sudo ls -la /var/lib/ceph/tmp/
> total 8
> drwxr-xr-x 2 root root 4096 Jun 27 16:32 .
> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
> -rw-r--r-- 1 root root    0 Jun 27 16:32 ceph-disk.prepare.lock
>
> [ceph@ceph02 ~]$ sudo ceph-disk list
> /dev/xvda1 other, ext4, mounted on /
> /dev/xvdb :
>  /dev/xvdb1 ceph data, prepared, cluster ceph, journal /dev/xvdb2
>  /dev/xvdb2 ceph journal, for /dev/xvdb1
>
>
> In the cephadm:
>
> [ceph@cephadm ceph-cloud]$ ceph-deploy osd activate ceph02:xvdb1
> [ceph_deploy.conf][DEBUG ] found configuration file at:
> /home/ceph/.cephdeploy.conf
> [ceph_deploy.cli][INFO  ] Invoked (1.5.5): /usr/bin/ceph-deploy osd
> activate ceph02:xvdb1
> [ceph_deploy.osd][DEBUG ] Activating cluster ceph disks ceph02:/dev/xvdb1:
> [ceph02][DEBUG ] connected to host: ceph02
> [ceph02][DEBUG ] detect platform information from remote host
> [ceph02][DEBUG ] detect machine type
> [ceph_deploy.osd][INFO  ] Distro info: Scientific Linux 6.2 Carbon
> [ceph_deploy.osd][DEBUG ] activating host ceph02 disk /dev/xvdb1
> [ceph_deploy.osd][DEBUG ] will use init type: sysvinit
> [ceph02][INFO  ] Running command: sudo ceph-disk-activate --mark-init
> sysvinit --mount /dev/xvdb1
> [ceph02][WARNIN] got monmap epoch 2
> [ceph02][WARNIN] 2014-06-27 16:35:49.948865 7f143254e7a0 -1
> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkjournal error creating journal on
> /var/lib/ceph/tmp/mnt.9VZHpR/journal: (2) No such file or directory
> [ceph02][WARNIN] 2014-06-27 16:35:49.948893 7f143254e7a0 -1 OSD::mkfs:
> ObjectStore::mkfs failed with error -2
> [ceph02][WARNIN] 2014-06-27 16:35:49.948957 7f143254e7a0 -1  ** ERROR:
> error creating empty object store in /var/lib/ceph/tmp/mnt.9VZHpR: (2) No
> such file or directory
>
> [ceph02][WARNIN] ERROR:ceph-disk:Failed to activate
> [ceph02][WARNIN] Traceback (most recent call last):
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 2579, in <module>
> [ceph02][WARNIN]     main()
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 2557, in main
> [ceph02][WARNIN]     args.func(args)
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1910, in main_activate
> [ceph02][WARNIN]     init=args.mark_init,
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1686, in mount_activate
> [ceph02][WARNIN]     (osd_id, cluster) = activate(path,
> activate_key_template, init)
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1849, in activate
> [ceph02][WARNIN]     keyring=keyring,
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1484, in mkfs
> [ceph02][WARNIN]     '--keyring', os.path.join(path, 'keyring'),
> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 303, in
> command_check_call
> [ceph02][WARNIN]     return subprocess.check_call(arguments)
> [ceph02][WARNIN]   File "/usr/lib64/python2.6/subprocess.py", line 505, in
> check_call
> [ceph02][WARNIN]     raise CalledProcessError(retcode, cmd)
> [ceph02][WARNIN] subprocess.CalledProcessError: Command
> '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '--mkkey', '-i', '0',
> '--monmap', '/var/lib/ceph/tmp/mnt.9VZHpR/activate.monmap', '--osd-data',
> '/var/lib/ceph/tmp/mnt.9VZHpR', '--osd-journal',
> '/var/lib/ceph/tmp/mnt.9VZHpR/journal', '--osd-uuid',
> '5e93fa7c-b6f7-4684-981b-bf73254bd87a', '--keyring',
> '/var/lib/ceph/tmp/mnt.9VZHpR/keyring']' returned non-zero exit status 1
> [ceph02][ERROR ] RuntimeError: command returned non-zero exit status: 1
> [ceph_deploy][ERROR ] RuntimeError: Failed to execute command:
> ceph-disk-activate --mark-init sysvinit --mount /dev/xvdb1
>
>
> At OSD log:
>
> [ceph@ceph02 ~]$ tail -100 /var/log/ceph/ceph-osd.0.log
> 2014-06-27 16:35:49.859984 7f143254e7a0  0 ceph version 0.80.1
> (a38fe1169b6d2ac98b427334c12d7cf81f809b74), process ceph-osd, pid 6590
> 2014-06-27 16:35:49.861265 7f143254e7a0  1
> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkfs in /var/lib/ceph/tmp/mnt.9VZHpR
> 2014-06-27 16:35:49.861319 7f143254e7a0  1
> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkfs fsid is already set to
> 5e93fa7c-b6f7-4684-981b-bf73254bd87a
> 2014-06-27 16:35:49.948589 7f143254e7a0  1
> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) leveldb db exists/created
> 2014-06-27 16:35:49.948865 7f143254e7a0 -1
> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkjournal error creating journal on
> /var/lib/ceph/tmp/mnt.9VZHpR/journal: (2) No such file or directory
> 2014-06-27 16:35:49.948893 7f143254e7a0 -1 OSD::mkfs: ObjectStore::mkfs
> failed with error -2
> 2014-06-27 16:35:49.948957 7f143254e7a0 -1  ** ERROR: error creating empty
> object store in /var/lib/ceph/tmp/mnt.9VZHpR: (2) No such file or directory
> [ceph@ceph02 ~]$
>
> [ceph@ceph02 ~]$ sudo ceph-disk list
> /dev/xvda1 other, ext4, mounted on /
> /dev/xvdb :
>  /dev/xvdb1 ceph data, prepared, cluster ceph, osd.0, journal /dev/xvdb2
>  /dev/xvdb2 ceph journal, for /dev/xvdb1
>
>
> Thanks in advance, I
>
>
>
> 2014-06-27 15:30 GMT+02:00 Alfredo Deza <[email protected]>:
>
> Can you paste the full ceph-deploy logs? there are a few reasons why
>> this might be happening.
>>
>>
>>
>> On Fri, Jun 27, 2014 at 6:42 AM, Iban Cabrillo <[email protected]>
>> wrote:
>> > Hi,
>> >
>> >  I am a little frustrated. After 6 times trying to deploy a test ceph
>> always
>> > get the same error in the osd activation stage.
>> >  The version is firefly (for el6 repo), 3 mons, 3 osds all of then Xen
>> VMs.
>> >
>> >  The mons wake up correctly and I do not know why two osd servers too
>> after
>> > a lot of errors, always the same
>> >
>> > #ceph-deploy --verbose osd prepare ceph02:xvdb (works fine)
>> >
>> > [root@ceph02 ~]# parted /dev/xvdb
>> > GNU Parted 2.1
>> > Using /dev/xvdb
>> > Welcome to GNU Parted! Type 'help' to view a list of commands.
>> > (parted) p
>> > Model: Xen Virtual Block Device (xvd)
>> > Disk /dev/xvdb: 107GB
>> > Sector size (logical/physical): 512B/512B
>> > Partition Table: gpt
>> >
>> > Number  Start   End     Size    File system  Name          Flags
>> >  2      1049kB  10,7GB  10,7GB  xfs          ceph journal
>> >  1      10,7GB  107GB   96,6GB  xfs          ceph data
>> >
>> >
>> > But the activate give us this error:
>> >
>> >  ceph-deploy --verbose osd activate ceph02:xvdb1:/dev/xvdb2
>> >
>> >   [ceph02][WARNIN] 2014-06-27 12:27:34.750160 7f123b33d7a0 -1
>> > filestore(/var/lib/ceph/tmp/mnt.HacFAP) mkjournal error creating
>> journal on
>> > /var/lib/ceph/tmp/mnt.HacFAP/journal: (2) No such file or directory
>> > [ceph02][WARNIN] 2014-06-27 12:27:34.750281 7f123b33d7a0 -1 OSD::mkfs:
>> > ObjectStore::mkfs failed with error -2
>> > [ceph02][WARNIN] 2014-06-27 12:27:34.750416 7f123b33d7a0 -1  ** ERROR:
>> error
>> > creating empty object store in /var/lib/ceph/tmp/mnt.HacFAP: (2) No such
>> > file or directory
>> > [ceph02][WARNIN] ERROR:ceph-disk:Failed to activate
>> >
>> > Two of then following the same procedure after the same error several
>> time,
>> > raise up.
>> >  [ceph@ceph03 ~]$ df -h
>> > Filesystem      Size  Used Avail Use% Mounted on
>> > /dev/xvda1      5.0G  2.2G  2.6G  46% /
>> > tmpfs           935M     0  935M   0% /dev/shm
>> > /dev/xvdb1       90G   37M   90G   1% /var/lib/ceph/osd/ceph-6
>> >
>> >
>> > Any idea please?
>> >
>> >
>> > Bertrand Russell:
>> > "El problema con el mundo es que los estúpidos están seguros de todo y
>> los
>> > inteligentes están llenos de dudas"
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > [email protected]
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to