Hi Simon, If everything is in the same Ceph cluster and you want to move the whole “.rgw.buckets” (I assume your RBD traffic is targeted into a “data” or “rbd” pool) to your cold storage OSD maybe you could edit the CRUSH map, then it’s just a matter of rebalancing. You can check the ssd/platter example in the doc: http://docs.ceph.com/docs/master/rados/operations/crush-map/ or this article detailing different maps: http://cephnotes.ksperis.com/blog/2015/02/02/crushmap-example-of-a-hierarchical-cluster-map
Cheers, Maxime From: ceph-users <[email protected]> on behalf of Simon Murray <[email protected]> Date: Tuesday 16 August 2016 12:25 To: "[email protected]" <[email protected]> Subject: [ceph-users] rados cppool slooooooowness Morning guys, I've got about 8 million objects sat in .rgw.buckets that wants moving out of the way of OpenStack RDB traffic onto its own (admittedly small) cold storage pool on separate OSDs. I attempted to do this over the weekend during a 12h scheduled downtime, however my estimates had this pool completing in a rather un-customer friendly (think no backups...) 7 days. Anyone had any experience in doing this quicker? Any obvious reasons why I can't hack do_copy_pool() to spawn a bunch of threads and bang this off in a few hours? Cheers Si DataCentred Limited registered in England and Wales no. 05611763
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
