Hi David!

lsblk:

NAME    MAJ:MIN RM   SIZE RO TYPE  MOUNTPOINT
sda       8:0    0 931,5G  0 disk
├─sda1    8:1    0   476M  0 part
├─sda2    8:2    0  46,6G  0 part
│ └─md0   9:0    0  46,5G  0 raid1 /
└─sda3    8:3    0 884,5G  0 part /var/lib/ceph/osd/ceph-3
sdb       8:16   0 931,5G  0 disk
├─sdb1    8:17   0   476M  0 part  /boot/efi
├─sdb2    8:18   0  46,6G  0 part
│ └─md0   9:0    0  46,5G  0 raid1 /
└─sdb3    8:19   0 884,5G  0 part /var/lib/ceph/osd/ceph-0
sdc       8:32   0 232,9G  0 disk
├─sdc1    8:33   0    20G  0 part
├─sdc2    8:34   0   576M  0 part
├─sdc3    8:35   0    20G  0 part
└─sdc4    8:36   0   576M  0 part


2017-06-26 22:37 keltezéssel, David Turner írta:
What is the output of `lsblk`?

On Mon, Jun 26, 2017 at 4:32 PM Papp Rudolf Péter <[email protected] <mailto:[email protected]>> wrote:

    Dear cephers,

    Could someone show me an url where can I found how ceph calculate the
    available space?

    I've installed a small ceph (Kraken) environment with bluestore OSDs.
    The servers contains 2 disks and 1 ssd. The disk 1. part is UEFI (~500
    MB), 2. raid (~50GB), 3. ceph disk (450-950MB). 1 server with 2 500 GB
    HDDs, 2 with 1 TB HDDs total 3 servers.

    For example the HDD parts:
    /dev/sdb1      2048     976895     974848   476M EFI System
    /dev/sdb2    976896   98633727   97656832  46,6G Linux RAID
    /dev/sdb3  98633728 1953525134 1854891407 884,5G Ceph OSD
    info from ceph-disk:
      /dev/sda :
      /dev/sda1 other, vfat
      /dev/sda2 other, linux_raid_member
      /dev/sda3 ceph data, active, cluster ceph, osd.4, block.db
    /dev/sdc1,
    block.wal /dev/sdc2
    /dev/sdb :
      /dev/sdb1 other, vfat, mounted on /boot/efi
      /dev/sdb2 other, linux_raid_member
      /dev/sdb3 ceph data, active, cluster ceph, osd.1, block.db
    /dev/sdc3,
    block.wal /dev/sdc4
    /dev/sdc :
      /dev/sdc1 ceph block.db, for /dev/sda3
      /dev/sdc2 ceph block.wal, for /dev/sda3
      /dev/sdc3 ceph block.db, for /dev/sdb3
      /dev/sdc4 ceph block.wal, for /dev/sdb3

    The reported size from ceph osd df tree:
    ID WEIGHT  REWEIGHT SIZE   USE    AVAIL  %USE VAR  PGS TYPE NAME
    -1 0.17578        -   179G   104M   179G 0.06 1.00   0 root default
    -2 0.05859        - 61439M 35696k 61405M 0.06 1.00   0  host cl2
      0 0.02930  1.00000 30719M 17848k 30702M 0.06 1.00   0 osd.0
      3 0.02930  1.00000 30719M 17848k 30702M 0.06 1.00   0 osd.3
    -3 0.05859        - 61439M 35696k 61405M 0.06 1.00   0  host cl3
      1 0.02930  1.00000 30719M 17848k 30702M 0.06 1.00   0 osd.1
      4 0.02930  1.00000 30719M 17848k 30702M 0.06 1.00   0 osd.4
    -4 0.05859        - 61439M 35696k 61405M 0.06 1.00   0  host cl1
      2 0.02930  1.00000 30719M 17848k 30702M 0.06 1.00   0 osd.2
      5 0.02930  1.00000 30719M 17848k 30702M 0.06 1.00   0 osd.5
                   TOTAL   179G   104M   179G 0.06
    MIN/MAX VAR: 1.00/1.00  STDDEV: 0

    ~ 30GB each 10 percent of the smallest real size. 3x replication.
    Could
    be possible that the system using wrong partition (2. in this
    scenario)
    for usable space calculation? Can I write more data than the
    calculated?

    Another hint?

    Thank you!


    _______________________________________________
    ceph-users mailing list
    [email protected] <mailto:[email protected]>
    http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to