I have a new deployment and it always has this problem even if I increase
the size of the OSD, it stays at 8.  I saw examples where others had this
problem but it was with the RBD pool, I don't have an RBD pool, and just
deployed it fresh with ansible.


health: HEALTH_WARN
            1 MDSs report slow metadata IOs
            Reduced data availability: 16 pgs inactive
            Degraded data redundancy: 16 pgs undersized
            too few PGs per OSD (16 < min 30)
 data:
    pools:   2 pools, 16 pgs
    objects: 0  objects, 0 B
    usage:   2.0 GiB used, 39 GiB / 41 GiB avail
    pgs:     100.000% pgs not active
             16 undersized+peered


# ceph osd pool ls
cephfs_data
cephfs_metadata


# ceph osd tree
ID CLASS WEIGHT  TYPE NAME            STATUS REWEIGHT PRI-AFF
-1       0.03989 root default
-3       0.03989     host mytesthost104
 0   hdd 0.03989         osd.0            up  1.00000 1.00000


# ceph osd pool set cephfs_data pgp_num 64
Error EINVAL: specified pgp_num 64 > pg_num 8
# ceph osd pool set cephfs_data pgp_num 256
Error EINVAL: specified pgp_num 256 > pg_num 8
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to