The output of `sudo df -h` would also be helpful. Sudo/root is generally required because the OSD folders are only readable by the Ceph user.
On Mon, Jun 26, 2017 at 4:37 PM David Turner <[email protected]> wrote: > What is the output of `lsblk`? > > On Mon, Jun 26, 2017 at 4:32 PM Papp Rudolf Péter <[email protected]> wrote: > >> Dear cephers, >> >> Could someone show me an url where can I found how ceph calculate the >> available space? >> >> I've installed a small ceph (Kraken) environment with bluestore OSDs. >> The servers contains 2 disks and 1 ssd. The disk 1. part is UEFI (~500 >> MB), 2. raid (~50GB), 3. ceph disk (450-950MB). 1 server with 2 500 GB >> HDDs, 2 with 1 TB HDDs total 3 servers. >> >> For example the HDD parts: >> /dev/sdb1 2048 976895 974848 476M EFI System >> /dev/sdb2 976896 98633727 97656832 46,6G Linux RAID >> /dev/sdb3 98633728 1953525134 1854891407 884,5G Ceph OSD >> info from ceph-disk: >> /dev/sda : >> /dev/sda1 other, vfat >> /dev/sda2 other, linux_raid_member >> /dev/sda3 ceph data, active, cluster ceph, osd.4, block.db /dev/sdc1, >> block.wal /dev/sdc2 >> /dev/sdb : >> /dev/sdb1 other, vfat, mounted on /boot/efi >> /dev/sdb2 other, linux_raid_member >> /dev/sdb3 ceph data, active, cluster ceph, osd.1, block.db /dev/sdc3, >> block.wal /dev/sdc4 >> /dev/sdc : >> /dev/sdc1 ceph block.db, for /dev/sda3 >> /dev/sdc2 ceph block.wal, for /dev/sda3 >> /dev/sdc3 ceph block.db, for /dev/sdb3 >> /dev/sdc4 ceph block.wal, for /dev/sdb3 >> >> The reported size from ceph osd df tree: >> ID WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME >> -1 0.17578 - 179G 104M 179G 0.06 1.00 0 root default >> -2 0.05859 - 61439M 35696k 61405M 0.06 1.00 0 host cl2 >> 0 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.0 >> 3 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.3 >> -3 0.05859 - 61439M 35696k 61405M 0.06 1.00 0 host cl3 >> 1 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.1 >> 4 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.4 >> -4 0.05859 - 61439M 35696k 61405M 0.06 1.00 0 host cl1 >> 2 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.2 >> 5 0.02930 1.00000 30719M 17848k 30702M 0.06 1.00 0 osd.5 >> TOTAL 179G 104M 179G 0.06 >> MIN/MAX VAR: 1.00/1.00 STDDEV: 0 >> >> ~ 30GB each 10 percent of the smallest real size. 3x replication. Could >> be possible that the system using wrong partition (2. in this scenario) >> for usable space calculation? Can I write more data than the calculated? >> >> Another hint? >> >> Thank you! >> >> >> _______________________________________________ >> ceph-users mailing list >> [email protected] >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
