You could fix this by changing your block size when formatting the
mount-point with the mkfs -b command.  I had this same issue when dealing
with the filesystem using glusterfs and the solution is to either use a
filesystem that allocates inodes automatically or change the block size
when you build the filesystem.  Unfortunately, the only way to fix the
problem that I have seen is to reformat

On Mon, Mar 23, 2015 at 5:51 AM, Kamil Kuramshin <kamil.kurams...@tatar.ru>
wrote:

>  In my case there was cache pool for ec-pool serving RBD-images, and
> object size is 4Mb, and client was an *kernel-rbd *client
> each SSD disk is 60G disk, 2 disk per node,  6 nodes in total = 12 OSDs in
> total
>
>
> 23.03.2015 12:00, Christian Balzer пишет:
>
> Hello,
>
> This is rather confusing, as cache-tiers are just normal OSDs/pools and
> thus should have Ceph objects of around 4MB in size by default.
>
> This is matched on what I see with Ext4 here (normal OSD, not a cache
> tier):
> ---
> size:
> /dev/sde1       2.7T  204G  2.4T   8% /var/lib/ceph/osd/ceph-0
> inodes:
> /dev/sde1      183148544 55654 183092890    1% /var/lib/ceph/osd/ceph-0
> ---
>
> On a more fragmented cluster I see a 5:1 size to inode ratio.
>
> I just can't fathom how there could be 3.3 million inodes (and thus a
> close number of files) using 30G, making the average file size below 10
> Bytes.
>
> Something other than your choice of file system is probably at play here.
>
> How fragmented are those SSDs?
> What's your default Ceph object size?
> Where _are_ those 3 million files in that OSD, are they actually in the
> object files like:
> -rw-r--r-- 1 root root 4194304 Jan  9 15:27 
> /var/lib/ceph/osd/ceph-0/current/3.117_head/DIR_7/DIR_1/DIR_5/rb.0.23a8f.238e1f29.000000027632__head_C4F3D517__3
>
> What's your use case, RBD, CephFS, RadosGW?
>
> Regards,
>
> Christian
>
> On Mon, 23 Mar 2015 10:32:55 +0300 Kamil Kuramshin wrote:
>
>
>  Recently got a problem with OSDs based on SSD disks used in cache tier
> for EC-pool
>
> superuser@node02:~$ df -i
> Filesystem                    Inodes   IUsed *IFree* IUse% Mounted on
> <...>
> /dev/sdb1                    3335808 3335808 *0* 100%
> /var/lib/ceph/osd/ceph-45
> /dev/sda1                    3335808 3335808 *0* 100%
> /var/lib/ceph/osd/ceph-46
>
> Now that OSDs are down on each ceph-node and cache tiering is not
> working.
>
> superuser@node01:~$ sudo tail /var/log/ceph/ceph-osd.45.log
> 2015-03-23 10:04:23.631137 7fb105345840  0 ceph version 0.87.1
> (283c2e7cfa2457799f534744d7d549f83ea1335e), process ceph-osd, pid 1453465
> 2015-03-23 10:04:23.640676 7fb105345840  0
> filestore(/var/lib/ceph/osd/ceph-45) backend generic (magic 0xef53)
> 2015-03-23 10:04:23.640735 7fb105345840 -1
> genericfilestorebackend(/var/lib/ceph/osd/ceph-45) detect_features:
> unable to create /var/lib/ceph/osd/ceph-45/fiemap_test: (28) No space
> left on device
> 2015-03-23 10:04:23.640763 7fb105345840 -1
> filestore(/var/lib/ceph/osd/ceph-45) _detect_fs: detect_features error:
> (28) No space left on device
> 2015-03-23 10:04:23.640772 7fb105345840 -1
> filestore(/var/lib/ceph/osd/ceph-45) FileStore::mount : error in
> _detect_fs: (28) No space left on device
> 2015-03-23 10:04:23.640783 7fb105345840 -1  ** ERROR: error converting
> store /var/lib/ceph/osd/ceph-45: (28) *No space left on device*
>
> In the same time*df -h *is confusing:
>
> superuser@node01:~$ df -h
> Filesystem                  Size  Used *Avail* Use% Mounted on
> <...>
> /dev/sda1                    50G   29G *20G*
> 60% /var/lib/ceph/osd/ceph-45 /dev/sdb1                    50G   27G
> *21G*  56% /var/lib/ceph/osd/ceph-46
>
>
> Filesystem used on affected OSDs is EXt4. All OSDs are deployed with
> ceph-deploy:
> $ ceph-deploy osd create --zap-disk --fs-type ext4 <node-name>:<device>
>
>
> Help me out what it was just test deployment and all EC-pool data was
> lost since I /can't start OSDs/ and ceph cluster/becames degraded /until
> I removed all affected tiered pools (cache & EC)
> So this is just my observation of what kind of problems can be faced if
> you choose wrong Filesystem for OSD backend.
> And now I *strongly* recommend you to choose*XFS* or *Btrfs* filesystems
> because both are supporting dynamic inode allocation and this problem
> can't arise with them.
>
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to