I believe I found something but I don't know how to fix it.
I run "ceph df" and I'm seeing that cephfs_data and cephfs_metadata is at
100% USED.
How can I increase the cephfs_data and cephfs_metadata pool.
Sorry I'm new with Ceph.
root@pf-us1-dfs1:/etc/ceph# ceph df
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
73 TiB 34 TiB 39 TiB 53.12
POOLS:
NAME ID USED %USED MAX AVAIL
OBJECTS
poolcephfs 1 0 B 0 0 B
0
cephfs_data 2 3.6 TiB 100.00 0 B
169273821
cephfs_metadata 3 1.0 GiB 100.00 0 B
208981
.rgw.root 4 1.1 KiB 100.00 0 B
4
default.rgw.control 5 0 B 0 0 B
8
default.rgw.meta 6 0 B 0 0 B
0
default.rgw.log 7 0 B 0 0 B
207
On Tue, Jan 8, 2019 at 10:30 AM Rodrigo Embeita <[email protected]>
wrote:
> Hi guys, I need your help.
> I'm new with Cephfs and we started using it as file storage.
> Today we are getting no space left on device but I'm seeing that we have
> plenty space on the filesystem.
> Filesystem Size Used Avail Use% Mounted on
> 192.168.51.8,192.168.51.6,192.168.51.118:6789:/pagefreezer/smhosts 73T
> 39T 35T 54% /mnt/cephfs
>
> We have 35TB of disk space. I've added 2 additional OSD disks with 7TB
> each but I'm getting the error "No space left on device" every time that I
> want to add a new file.
> After adding the 2 additional OSD disks I'm seeing that the load is beign
> distributed among the cluster.
> Please I need your help.
>
> root@pf-us1-dfs1:/etc/ceph# ceph -s
> cluster:
> id: 609e9313-bdd3-449e-a23f-3db8382e71fb
> health: HEALTH_ERR
> 2 backfillfull osd(s)
> 1 full osd(s)
> 7 pool(s) full
> 197313040/508449063 objects misplaced (38.807%)
> Degraded data redundancy: 2/508449063 objects degraded
> (0.000%), 2 pgs degraded
> Degraded data redundancy (low space): 16 pgs backfill_toofull,
> 3 pgs recovery_toofull
>
> services:
> mon: 3 daemons, quorum pf-us1-dfs2,pf-us1-dfs1,pf-us1-dfs3
> mgr: pf-us1-dfs3(active), standbys: pf-us1-dfs2
> mds: pagefs-2/2/2 up
> {0=pf-us1-dfs3=up:active,1=pf-us1-dfs1=up:active}, 1 up:standby
> osd: 10 osds: 10 up, 10 in; 189 remapped pgs
> rgw: 1 daemon active
>
> data:
> pools: 7 pools, 416 pgs
> objects: 169.5 M objects, 3.6 TiB
> usage: 39 TiB used, 34 TiB / 73 TiB avail
> pgs: 2/508449063 objects degraded (0.000%)
> 197313040/508449063 objects misplaced (38.807%)
> 224 active+clean
> 168 active+remapped+backfill_wait
> 16 active+remapped+backfill_wait+backfill_toofull
> 5 active+remapped+backfilling
> 2 active+recovery_toofull+degraded
> 1 active+recovery_toofull
>
> io:
> recovery: 1.1 MiB/s, 31 objects/s
>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com