> On 23 Mar 2015, at 03:58, Haomai Wang <haomaiw...@gmail.com> wrote:
> 
> On Mon, Mar 23, 2015 at 2:53 AM, Josef Johansson <jose...@gmail.com 
> <mailto:jose...@gmail.com>> wrote:
>> Hi all!
>> 
>> Trying to figure out how much my journals are used, using SSDs as journals 
>> and SATA-drives as storage, I dive into perf dump.
>> But I can’t figure out why journal_queue_bytes is at constant 0. The only 
>> thing that differs is dirtied in WBThrottle.
> 
> journal_queue_bytes means how much journal data in the queue and is
> waiting for Journal Thread to be processed.
> 
> Still now osd can't tell you how much data in the journal waiting for
> writeback and sync.
> 
Hm, who knows that then?
Is this the WBThrottle value?

No way of knowing how much journal is used at all?

Maybe I thought of this wrong so if I understand you correctly

Data is written to OSD
The journal saves it to the queue
Waits for others to sync the requests as well
Sends a ACK to the client
Starts writing to the filestore buffer
filestore buffer commits when limits and met (inodes/ios-dirtied, 
filestore_sync_max_interval)

So if I’m meeting latency and want to see if my journals are lazy, I should 
indeed look at journal_queue_bytes, if that’s zero, it’s behaving well.

Thanks,
Josef

>> 
>> Maybe I’ve disable that when setting the in-memory debug variables to 0/0?
>> 
>> Thanks,
>> Josef
>> 
>> # ceph --version
>> ceph version 0.67.7 (d7ab4244396b57aac8b7e80812115bbd079e6b73)
>> 
>> # ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok config show | grep 
>> journal
>>  "journaler": "0\/0",
>>  "journal": "0\/0",
>>  "journaler_allow_split_entries": "true",
>>  "journaler_write_head_interval": "15",
>>  "journaler_prefetch_periods": "10",
>>  "journaler_prezero_periods": "5",
>>  "journaler_batch_interval": "0.001",
>>  "journaler_batch_max": "0",
>>  "mds_kill_journal_at": "0",
>>  "mds_kill_journal_expire_at": "0",
>>  "mds_kill_journal_replay_at": "0",
>>  "osd_journal": "\/var\/lib\/ceph\/osd\/ceph-0\/journal",
>>  "osd_journal_size": "25600",
>>  "filestore_fsync_flushes_journal_data": "false",
>>  "filestore_journal_parallel": "false",
>>  "filestore_journal_writeahead": "false",
>>  "filestore_journal_trailing": "false",
>>  "journal_dio": "true",
>>  "journal_aio": "true",
>>  "journal_force_aio": "false",
>>  "journal_max_corrupt_search": "10485760",
>>  "journal_block_align": "true",
>>  "journal_write_header_frequency": "0",
>>  "journal_max_write_bytes": "10485760",
>>  "journal_max_write_entries": "100",
>>  "journal_queue_max_ops": "300",
>>  "journal_queue_max_bytes": "33554432",
>>  "journal_align_min_size": "65536",
>>  "journal_replay_from": "0",
>>  "journal_zero_on_create": "false",
>>  "journal_ignore_corruption": "false",
>> 
>> # ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok perf dump
>> { "WBThrottle": { "bytes_dirtied": 32137216,
>>      "bytes_wb": 0,
>>      "ios_dirtied": 1445,
>>      "ios_wb": 0,
>>      "inodes_dirtied": 491,
>>      "inodes_wb": 0},
>>  "filestore": { "journal_queue_max_ops": 300,
>>      "journal_queue_ops": 0,
>>      "journal_ops": 116105073,
>>      "journal_queue_max_bytes": 33554432,
>>      "journal_queue_bytes": 0,
>>      "journal_bytes": 3160504432839,
>>      "journal_latency": { "avgcount": 116105073,
>>          "sum": 64951.260611000},
>>      "journal_wr": 112261141,
>>      "journal_wr_bytes": { "avgcount": 112261141,
>>          "sum": 3426141528064},
>>      "op_queue_max_ops": 50,
>>      "op_queue_ops": 0,
>>      "ops": 116105073,
>>      "op_queue_max_bytes": 104857600,
>>      "op_queue_bytes": 0,
>>      "bytes": 3159111228243,
>>      "apply_latency": { "avgcount": 116105073,
>>          "sum": 247410.066048000},
>>      "committing": 0,
>>      "commitcycle": 267176,
>>      "commitcycle_interval": { "avgcount": 267176,
>>          "sum": 1873193.631124000},
>>      "commitcycle_latency": { "avgcount": 267176,
>>          "sum": 390421.062990000},
>>      "journal_full": 0,
>>      "queue_transaction_latency_avg": { "avgcount": 116105073,
>>          "sum": 378.948923000}},
>>  "leveldb": { "leveldb_get": 699871216,
>>      "leveldb_transaction": 522440246,
>>      "leveldb_compact": 0,
>>      "leveldb_compact_range": 0,
>>      "leveldb_compact_queue_merge": 0,
>>      "leveldb_compact_queue_len": 0},
>>  "mutex-FileJournal::completions_lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-FileJournal::finisher_lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-FileJournal::write_lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-FileJournal::writeq_lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-JOS::ApplyManager::apply_lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-JOS::ApplyManager::com_lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-JOS::SubmitManager::lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "mutex-WBThrottle::lock": { "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "osd": { "opq": 0,
>>      "op_wip": 0,
>>      "op": 83920139,
>>      "op_in_bytes": 1075345387581,
>>      "op_out_bytes": 954428806331,
>>      "op_latency": { "avgcount": 83920139,
>>          "sum": 1279934.620502000},
>>      "op_r": 32399024,
>>      "op_r_out_bytes": 953657617715,
>>      "op_r_latency": { "avgcount": 32399024,
>>          "sum": 238792.729743000},
>>      "op_w": 3321731,
>>      "op_w_in_bytes": 52637941027,
>>      "op_w_rlat": { "avgcount": 3321731,
>>          "sum": 15577.620040000},
>>      "op_w_latency": { "avgcount": 3321731,
>>          "sum": 62541.746123000},
>>      "op_rw": 48199384,
>>      "op_rw_in_bytes": 1022707446554,
>>      "op_rw_out_bytes": 771188616,
>>      "op_rw_rlat": { "avgcount": 48199384,
>>          "sum": 169776.087496000},
>>      "op_rw_latency": { "avgcount": 48199384,
>>          "sum": 978600.144636000},
>>      "subop": 73746080,
>>      "subop_in_bytes": 2008774955062,
>>      "subop_latency": { "avgcount": 73746080,
>>          "sum": 346096.627047000},
>>      "subop_w": 0,
>>      "subop_w_in_bytes": 2008774955062,
>>      "subop_w_latency": { "avgcount": 73746080,
>>          "sum": 346096.627047000},
>>      "subop_pull": 0,
>>      "subop_pull_latency": { "avgcount": 0,
>>          "sum": 0.000000000},
>>      "subop_push": 0,
>>      "subop_push_in_bytes": 0,
>>      "subop_push_latency": { "avgcount": 0,
>>          "sum": 0.000000000},
>>      "pull": 0,
>>      "push": 8255,
>>      "push_out_bytes": 24792364384,
>>      "push_in": 0,
>>      "push_in_bytes": 0,
>>      "recovery_ops": 8126,
>>      "loadavg": 585,
>>      "buffer_bytes": 0,
>>      "numpg": 402,
>>      "numpg_primary": 161,
>>      "numpg_replica": 241,
>>      "numpg_stray": 0,
>>      "heartbeat_to_peers": 38,
>>      "heartbeat_from_peers": 0,
>>      "map_messages": 15262,
>>      "map_message_epochs": 70160,
>>      "map_message_epoch_dups": 116821,
>>      "messages_delayed_for_map": 105,
>>      "peering_latency": { "avgcount": 628,
>>          "sum": 28166.821628000}},
>>  "throttle-filestore_bytes": { "val": 0,
>>      "max": 33554432,
>>      "get": 0,
>>      "get_sum": 0,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 116105073,
>>      "take_sum": 3160504432839,
>>      "put": 112261141,
>>      "put_sum": 3160504432839,
>>      "wait": { "avgcount": 12,
>>          "sum": 0.097590000}},
>>  "throttle-filestore_ops": { "val": 0,
>>      "max": 300,
>>      "get": 0,
>>      "get_sum": 0,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 116105073,
>>      "take_sum": 116105073,
>>      "put": 112261141,
>>      "put_sum": 116105073,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-msgr_dispatch_throttler-client": { "val": 0,
>>      "max": 104857600,
>>      "get": 75187594,
>>      "get_sum": 978560902803,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 75187594,
>>      "put_sum": 978560902803,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-msgr_dispatch_throttler-cluster": { "val": 0,
>>      "max": 104857600,
>>      "get": 159487799,
>>      "get_sum": 2119870334346,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 159487799,
>>      "put_sum": 2119870334346,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-msgr_dispatch_throttler-hb_back_server": { "val": 0,
>>      "max": 104857600,
>>      "get": 19958614,
>>      "get_sum": 938054858,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 19958614,
>>      "put_sum": 938054858,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-msgr_dispatch_throttler-hb_front_server": { "val": 0,
>>      "max": 104857600,
>>      "get": 19958614,
>>      "get_sum": 938054858,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 19958614,
>>      "put_sum": 938054858,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-msgr_dispatch_throttler-hbclient": { "val": 0,
>>      "max": 104857600,
>>      "get": 44608572,
>>      "get_sum": 2096602884,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 44608572,
>>      "put_sum": 2096602884,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-osd_client_bytes": { "val": 0,
>>      "max": 524288000,
>>      "get": 74786210,
>>      "get_sum": 978238232914,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 117138891,
>>      "put_sum": 978238232914,
>>      "wait": { "avgcount": 0,
>>          "sum": 0.000000000}},
>>  "throttle-osd_client_messages": { "val": 17,
>>      "max": 100,
>>      "get": 75909689,
>>      "get_sum": 75909689,
>>      "get_or_fail_fail": 0,
>>      "get_or_fail_success": 0,
>>      "take": 0,
>>      "take_sum": 0,
>>      "put": 75909672,
>>      "put_sum": 75909672,
>>      "wait": { "avgcount": 478562,
>>          "sum": 5665.119305000}}}
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
> 
> -- 
> Best Regards,
> 
> Wheat

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to