Hi Jason, it looks the situation is the same, no difference. my ceph.conf is below, any comments or improvement required? --- [global] fsid = 106a12b0-5ed0-4a71-b6aa-68a09088ec33 mon_initial_members = ceph-mon1, ceph-mon2, ceph-mon3 mon_host = 192.168.8.11,192.168.8.12,192.168.8.13 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx filestore_xattr_use_omap = true osd pool default size = 3 osd pool default min size = 1 osd pool default pg num = 4096 osd pool default pgp num = 4096 osd_crush_chooseleaf_type = 1 mon_pg_warn_max_per_osd = 0 max_open_files = 131072
[mon] mon_data = /var/lib/ceph/mon/ceph-$id mon clock drift allowed = 2 mon clock drift warn backoff = 30 [osd] osd_data = /var/lib/ceph/osd/ceph-$id osd_journal_size = 20000 osd_mkfs_type = xfs osd_mkfs_options_xfs = -f filestore_xattr_use_omap = true filestore_min_sync_interval = 10 filestore_max_sync_interval = 15 filestore_queue_max_ops = 25000 filestore_queue_max_bytes = 10485760 filestore_queue_committing_max_ops = 5000 filestore_queue_committing_max_bytes = 10485760000 journal_max_write_bytes = 1073714824 journal_max_write_entries = 10000 journal_queue_max_ops = 50000 journal_queue_max_bytes = 10485760000 osd_max_write_size = 512 osd_client_message_size_cap = 2147483648 osd_deep_scrub_stride = 131072 osd_op_threads = 8 osd_disk_threads = 4 osd_map_cache_size = 1024 osd_map_cache_bl_size = 128 osd_mount_options_xfs = "rw,noexec,nodev,noatime,nodiratime,nobarrier" osd_recovery_op_priority = 4 osd_recovery_max_active = 10 osd_max_backfills = 4 rbd non blocking aio = false [client] rbd_cache = true rbd_cache_size = 268435456 rbd_cache_max_dirty = 134217728 rbd_cache_max_dirty_age = 5 --- On Mon, Oct 31, 2016 at 9:20 PM, Jason Dillaman <[email protected]> wrote: > On Sun, Oct 30, 2016 at 5:40 AM, Bill WONG <[email protected]> wrote: > > any ideas or comments? > > Can you set "rbd non blocking aio = false" in your ceph.conf and retry > librbd? This will eliminate at least one context switch on the read IO > path -- which result in increased latency under extremely low queue > depths. > > -- > Jason >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
