Thanks again Kevin.
If I reduce the size flag to a value of 2, that should fix the problem?

Regards

On Tue, Jan 8, 2019 at 11:28 AM Kevin Olbrich <[email protected]> wrote:

> You use replication 3 failure-domain host.
> OSD 2 and 4 are full, thats why your pool is also full.
> You need to add two disks to pf-us1-dfs3 or swap one from the larger
> nodes to this one.
>
> Kevin
>
> Am Di., 8. Jan. 2019 um 15:20 Uhr schrieb Rodrigo Embeita
> <[email protected]>:
> >
> > Hi Yoann, thanks for your response.
> > Here are the results of the commands.
> >
> > root@pf-us1-dfs2:/var/log/ceph# ceph osd df
> > ID CLASS WEIGHT  REWEIGHT SIZE    USE     AVAIL   %USE  VAR  PGS
> > 0   hdd 7.27739  1.00000 7.3 TiB 6.7 TiB 571 GiB 92.33 1.74 310
> > 5   hdd 7.27739  1.00000 7.3 TiB 5.6 TiB 1.7 TiB 77.18 1.45 271
> > 6   hdd 7.27739  1.00000 7.3 TiB 609 GiB 6.7 TiB  8.17 0.15  49
> > 8   hdd 7.27739  1.00000 7.3 TiB 2.5 GiB 7.3 TiB  0.03    0  42
> > 1   hdd 7.27739  1.00000 7.3 TiB 5.6 TiB 1.7 TiB 77.28 1.45 285
> > 3   hdd 7.27739  1.00000 7.3 TiB 6.9 TiB 371 GiB 95.02 1.79 296
> > 7   hdd 7.27739  1.00000 7.3 TiB 360 GiB 6.9 TiB  4.84 0.09  53
> > 9   hdd 7.27739  1.00000 7.3 TiB 4.1 GiB 7.3 TiB  0.06 0.00  38
> > 2   hdd 7.27739  1.00000 7.3 TiB 6.7 TiB 576 GiB 92.27 1.74 321
> > 4   hdd 7.27739  1.00000 7.3 TiB 6.1 TiB 1.2 TiB 84.10 1.58 351
> >                    TOTAL  73 TiB  39 TiB  34 TiB 53.13
> > MIN/MAX VAR: 0/1.79  STDDEV: 41.15
> >
> >
> > root@pf-us1-dfs2:/var/log/ceph# ceph osd pool ls detail
> > pool 1 'poolcephfs' replicated size 3 min_size 2 crush_rule 0
> object_hash rjenkins pg_num 128 pgp_num 128 last_change 471 fla
> > gs hashpspool,full stripe_width 0
> > pool 2 'cephfs_data' replicated size 3 min_size 2 crush_rule 0
> object_hash rjenkins pg_num 256 pgp_num 256 last_change 471 lf
> > or 0/439 flags hashpspool,full stripe_width 0 application cephfs
> > pool 3 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0
> object_hash rjenkins pg_num 256 pgp_num 256 last_change 47
> > 1 lfor 0/448 flags hashpspool,full stripe_width 0 application cephfs
> > pool 4 '.rgw.root' replicated size 3 min_size 2 crush_rule 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 471 flags ha
> > shpspool,full stripe_width 0 application rgw
> > pool 5 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 47
> > 1 flags hashpspool,full stripe_width 0 application rgw
> > pool 6 'default.rgw.meta' replicated size 3 min_size 2 crush_rule 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 471 f
> > lags hashpspool,full stripe_width 0 application rgw
> > pool 7 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 471 fl
> > ags hashpspool,full stripe_width 0 application rgw
> >
> >
> > root@pf-us1-dfs2:/var/log/ceph# ceph osd tree
> > ID CLASS WEIGHT   TYPE NAME            STATUS REWEIGHT PRI-AFF
> > -1       72.77390 root default
> > -3       29.10956     host pf-us1-dfs1
> > 0   hdd  7.27739         osd.0            up  1.00000 1.00000
> > 5   hdd  7.27739         osd.5            up  1.00000 1.00000
> > 6   hdd  7.27739         osd.6            up  1.00000 1.00000
> > 8   hdd  7.27739         osd.8            up  1.00000 1.00000
> > -5       29.10956     host pf-us1-dfs2
> > 1   hdd  7.27739         osd.1            up  1.00000 1.00000
> > 3   hdd  7.27739         osd.3            up  1.00000 1.00000
> > 7   hdd  7.27739         osd.7            up  1.00000 1.00000
> > 9   hdd  7.27739         osd.9            up  1.00000 1.00000
> > -7       14.55478     host pf-us1-dfs3
> > 2   hdd  7.27739         osd.2            up  1.00000 1.00000
> > 4   hdd  7.27739         osd.4            up  1.00000 1.00000
> >
> >
> > Thanks for your help guys.
> >
> >
> > On Tue, Jan 8, 2019 at 10:36 AM Yoann Moulin <[email protected]>
> wrote:
> >>
> >> Hello,
> >>
> >> > Hi guys, I need your help.
> >> > I'm new with Cephfs and we started using it as file storage.
> >> > Today we are getting no space left on device but I'm seeing that we
> have plenty space on the filesystem.
> >> > Filesystem              Size  Used Avail Use% Mounted on
> >> > 192.168.51.8,192.168.51.6,192.168.51.118:6789:/pagefreezer/smhosts
>  73T   39T   35T  54% /mnt/cephfs
> >> >
> >> > We have 35TB of disk space. I've added 2 additional OSD disks with
> 7TB each but I'm getting the error "No space left on device" every time that
> >> > I want to add a new file.
> >> > After adding the 2 additional OSD disks I'm seeing that the load is
> beign distributed among the cluster.
> >> > Please I need your help.
> >>
> >> Could you give us the output of
> >>
> >> ceph osd df
> >> ceph osd pool ls detail
> >> ceph osd tree
> >>
> >> Best regards,
> >>
> >> --
> >> Yoann Moulin
> >> EPFL IC-IT
> >> _______________________________________________
> >> ceph-users mailing list
> >> [email protected]
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> > _______________________________________________
> > ceph-users mailing list
> > [email protected]
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to