-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256
Just about how funny "ceph problems" are fixed by changing network
configurations.
- ----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1
On Thu, Sep 3, 2015 at 11:16 AM, Ian Colle wrote:
> Am I the only one who finds it funny that the "ceph problem" was fixed by an
> update to the disk controller firmware? :-)
>
> Ian
>
> On Thu, Sep 3, 2015 at 11:13 AM, Vickey Singh
> wrote:
>>
>> Hey Mark / Community
>>
>> These are the sequences of changes that seems to have fixed the ceph
>> problem
>>
>> 1# Upgrading Disk controller firmware from 6.34 to 6.64 ( latest )
>> 2# Rebooting all nodes in order to make new firmware into effect
>>
>> Read and write operations are now normal as well as system load and CPU
>> utilization
>>
>> - Vickey -
>>
>>
>> On Wed, Sep 2, 2015 at 11:28 PM, Vickey Singh
>> wrote:
>>>
>>> Thank You Mark , please see my response below.
>>>
>>> On Wed, Sep 2, 2015 at 5:23 PM, Mark Nelson wrote:
>>>>
>>>> On 09/02/2015 08:51 AM, Vickey Singh wrote:
>>>>>
>>>>> Hello Ceph Experts
>>>>>
>>>>> I have a strange problem , when i am reading or writing to Ceph pool ,
>>>>> its not writing properly. Please notice Cur MB/s which is going up and
>>>>> down
>>>>>
>>>>> --- Ceph Hammer 0.94.2
>>>>> -- CentOS 6, 2.6
>>>>> -- Ceph cluster is healthy
>>>>
>>>>
>>>> You might find that CentOS7 gives you better performance. In some cases
>>>> we were seeing nearly 2X.
>>>
>>>
>>> Wooo 2X , i would definitely plan for upgrade. Thanks
>>>
>>>>
>>>>
>>>>
>>>>>
>>>>>
>>>>> One interesting thing is when every i start rados bench command for
>>>>> read
>>>>> or write CPU Idle % goes down ~10 and System load is increasing like
>>>>> anything.
>>>>>
>>>>> Hardware
>>>>>
>>>>> HpSL4540
>>>>
>>>>
>>>> Please make sure the controller is on the newest firmware. There used
>>>> to be a bug that would cause sequential write performance to bottleneck
>>>> when
>>>> writeback cache was enabled on the RAID controller.
>>>
>>>
>>> Last month i have upgraded the firmwares for this hardware , so i hope
>>> they are up to date.
>>>
>>>>
>>>>
>>>>
>>>>> 32Core CPU
>>>>> 196G Memory
>>>>> 10G Network
>>>>
>>>>
>>>> Be sure to check the network too. We've seen a lot of cases where folks
>>>> have been burned by one of the NICs acting funky.
>>>
>>>
>>> At a first view , Interface looks good and they are pushing data nicely (
>>> what ever they are getting )
>>>
>>>>
>>>>
>>>>>
>>>>> I don't think hardware is a problem.
>>>>>
>>>>> Please give me clues / pointers , how should i troubleshoot this
>>>>> problem.
>>>>>
>>>>>
>>>>>
>>>>> # rados bench -p glance-test 60 write
>>>>> Maintaining 16 concurrent writes of 4194304 bytes for up to 60
>>>>> seconds
>>>>> or 0 objects
>>>>> Object prefix: benchmark_data_pouta-s01.pouta.csc.fi_2173350
>>>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg
>>>>> lat
>>>>> 0 0 0 0 0 0 -
>>>>> 0
>>>>> 1 16 20 4 15.99 16 0.12308
>>>>> 0.10001
>>>>> 2 16 37 21 41.9841 68 1.79104
>>>>> 0.827021
>>>>> 3 16 68 52 69.3122 124 0.084304
>>>>> 0.854829
>>>>> 4 16 114 98 97.9746 184 0.12285
>>>>> 0.614507
>>>>> 5 16 188 172 137.568 296 0.210669
>>>>> 0.449784
>>>>> 6 16 248 232 154.634 240 0.090418
>>>>> 0.390647
>>>>> 7 16 305 289 165.11 228 0.069769
>>>>> 0.347957
>>>>> 8 16 331 315 157.471 104 0.026247
>>>>> 0.3345
>>>>> 9 16 361 345 153.306 120 0.082861
>>>>> 0.320711
>>>>> 10 16 380 364 145.575 76 0.027964
>>>>> 0.310004
>>>>> 11 16 393 377 137.067 52 3.73332
>>>>> 0.393318
>>>>> 12 16 448 432 143.971 220 0.334664
>>>>> 0.415606
>>>>> 13 16 476 460 141.508 112 0.271096
>>>>> 0.406574
>>>>> 14 16 497 481 137.399 84 0.257794
>>>>> 0.412006
>>>>> 15 16 507 491 130.906 40 1.49351
>>>>> 0.428057
>>>>> 16 16 529 513 115.042 88 0.399384
>>>>> 0.48009
>>>>> 17 16 533 517 94.6286 16 5.50641
>>>>> 0.507804
>>>>> 18 16 537 521 83.405 16 4.42682
>>>>> 0.549951
>>>>> 19 16 538 522 80.349 4 11.2052
>>>>> 0.570363
>>>>> 2015-09-02 09:26:18.398641min lat: 0.023851 max lat: 11.2052 avg lat:
>>>>> 0.570363
>>>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg
>>>>> lat
>>>>> 20 16 538 522 77.3611 0 -
>>>>> 0.570363
>>>>> 21 16 540 524 74.8825 4 8.88847
>>>>> 0.591767
>>>>> 22 16 542 526 72.5748 8 1.41627
>>>>> 0.593555
>>>>> 23 16 543 527 70.2873 4 8.0856
>>>>> 0.607771
>>>>> 24 16 555 539 69.5674 48 0.145199
>>>>> 0.781685
>>>>> 25 16 560 544 68.0177 20 1.4342
>>>>> 0.787017
>>>>> 26 16 564 548 66.4241 16 0.451905
>>>>> 0.78765
>>>>> 27 16 566 550 64.7055 8 0.611129
>>>>> 0.787898
>>>>> 28 16 570 554 63.3138 16 2.51086
>>>>> 0.797067
>>>>> 29 16 570 554 61.5549 0 -
>>>>> 0.797067
>>>>> 30 16 572 556 60.1071 4 7.71382
>>>>> 0.830697
>>>>> 31 16 577 561 59.0515 20 23.3501
>>>>> 0.916368
>>>>> 32 16 590 574 58.8705 52 0.336684
>>>>> 0.956958
>>>>> 33 16 591 575 57.4986 4 1.92811
>>>>> 0.958647
>>>>> 34 16 591 575 56.0961 0 -
>>>>> 0.958647
>>>>> 35 16 591 575 54.7603 0 -
>>>>> 0.958647
>>>>> 36 16 597 581 54.0447 8 0.187351
>>>>> 1.00313
>>>>> 37 16 625 609 52.8394 112 2.12256
>>>>> 1.09256
>>>>> 38 16 631 615 52.227 24 1.57413
>>>>> 1.10206
>>>>> 39 16 638 622 51.7232 28 4.41663
>>>>> 1.15086
>>>>> 2015-09-02 09:26:40.510623min lat: 0.023851 max lat: 27.6704 avg lat:
>>>>> 1.15657
>>>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg
>>>>> lat
>>>>> 40 16 652 636 51.8102 56 0.113345
>>>>> 1.15657
>>>>> 41 16 682 666 53.1443 120 0.041251
>>>>> 1.17813
>>>>> 42 16 685 669 52.3395 12 0.501285
>>>>> 1.17421
>>>>> 43 15 690 675 51.7955 24 2.26605
>>>>> 1.18357
>>>>> 44 16 728 712 53.6062 148 0.589826
>>>>> 1.17478
>>>>> 45 16 728 712 52.6158 0 -
>>>>> 1.17478
>>>>> 46 16 728 712 51.6613 0 -
>>>>> 1.17478
>>>>> 47 16 728 712 50.7407 0 -
>>>>> 1.17478
>>>>> 48 16 772 756 52.9332 44 0.234811
>>>>> 1.1946
>>>>> 49 16 835 819 56.3577 252 5.67087
>>>>> 1.12063
>>>>> 50 16 890 874 59.1252 220 0.230806
>>>>> 1.06778
>>>>> 51 16 896 880 58.5409 24 0.382471
>>>>> 1.06121
>>>>> 52 16 896 880 57.5832 0 -
>>>>> 1.06121
>>>>> 53 16 896 880 56.6562 0 -
>>>>> 1.06121
>>>>> 54 16 896 880 55.7587 0 -
>>>>> 1.06121
>>>>> 55 16 897 881 54.9515 1 4.88333
>>>>> 1.06554
>>>>> 56 16 897 881 54.1077 0 -
>>>>> 1.06554
>>>>> 57 16 897 881 53.2894 0 -
>>>>> 1.06554
>>>>> 58 16 897 881 51.9335 0 -
>>>>> 1.06554
>>>>> 59 16 897 881 51.1792 0 -
>>>>> 1.06554
>>>>> 2015-09-02 09:27:01.267301min lat: 0.01405 max lat: 27.6704 avg lat:
>>>>> 1.06554
>>>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg
>>>>> lat
>>>>> 60 16 897 881 50.4445 0 -
>>>>> 1.06554
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> cluster 98d89661-f616-49eb-9ccf-84d720e179c0
>>>>> health HEALTH_OK
>>>>> monmap e3: 3 mons at
>>>>> {s01=10.100.50.1:6789/0,s02=10.100.50.2:6789/0,s03=1
>>>>>
>>>>> 0.100.50.3:6789/0 }, election epoch 666,
>>>>> quorum 0,1,2 s01,s02,s03
>>>>> * osdmap e121039: 240 osds: 240 up, 240 in*
>>>>> pgmap v850698: 7232 pgs, 31 pools, 439 GB data, 43090 kobjects
>>>>> 2635 GB used, 867 TB / 870 TB avail
>>>>> 7226 active+clean
>>>>> 6 active+clean+scrubbing+deep
>>>>
>>>>
>>>> Note the last line there. You'll likely want to try your test again
>>>> when scrubbing is complete. Also, you may want to try this script:
>>>
>>>
>>> Yeah i have tried few times when cluster is perfectly healthy ( not doing
>>> scrubbing / repairs )
>>>
>>>>
>>>>
>>>> https://github.com/ceph/cbt/blob/master/tools/readpgdump.py
>>>>
>>>> You can invoke it like:
>>>>
>>>> ceph pg dump | ./readpgdump.py
>>>>
>>>> That will give you a bunch of information about the pools on your
>>>> system. I'm a little concerned about how many PGs your glance-test pool
>>>> may
>>>> have given your totals above.
>>>
>>>
>>> Thanks for the link i would do that and also run rados bench for other
>>> pools ( where PG is higher )
>>>
>>>
>>> Now here are my some observations
>>>
>>> 1# When the cluster is not doing anything , Health_ok , with no
>>> background scrubbing / repairing. Also all system resources CPU/MEM/NET are
>>> mostly idle. In this Case when i start rados bench ( write / rand / seq ) ,
>>> after suddenly a few seconds
>>>
>>> --- rados bench output drops from ~500M to few 10M
>>> --- At the same time CPU busy 90% and System load bumps UP
>>>
>>> Once rados bench completes
>>>
>>> --- After few minutes System resources becomes IDLE
>>>
>>> 2# Sometime some PG becomes unclean for a few minutes while rados bench
>>> runs and then then quickly they becomes active+clean
>>>
>>>
>>> I am out of clues , so any help from community that leads me to think in
>>> right direction , would be helpful.
>>>
>>>
>>> - Vickey -
>>>
>>>>
>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> [email protected]
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> [email protected]
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
> Ian R. Colle
> Global Director of Software Engineering
> Red Hat, Inc.
> [email protected]
> +1-303-601-7713
> http://www.linkedin.com/in/ircolle
> http://www.twitter.com/ircolle
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v1.0.2
Comment: https://www.mailvelope.com
wsFcBAEBCAAQBQJV6ITkCRDmVDuy+mK58QAABBIQAMPK64BVaFBcBQwmuDCJ
D1OnYG207kV6z3n3nvx4i/YjO5I64NHQLk8U/hh7WjU6+Ku7ej59UGQERLv5
TGJ/oJUbvTdT+KQrpm10ZA5r3YPBGWEW8FVxrvB1eXs1sERdOppdcd1+kz2i
JGbUFwTfcn0D2x3ogUK8EmXv0Z5FCNj65b/asA8heOjXv5+bL2a5Nj6G8VIk
cIXtqrPDRaAyQ9CSvsgGrU5aoGlYf9zKbHxG+EBI4cPM5AZd2HZ+KWcY0oWr
eQeM5QRXsOmscx69Clu27mdNlqP/EOhHy5QMWesr8F8cte/b4TecMXrM4kFs
oSzGu8LjAVOAEYWX86mRobbuZGZ8iPQ7hgYQTRr6JOmLQfFp9id1v1Endou/
IipMTbFbmOjaM/+mmKRwBu/OhIOQPhPVJ/WMHGtfFIJfpI4bgs+fFI7F0mDF
/eNyrrmYMtBOkXToFtopN/utGy0ZBqwpZWq8q81YATdaHF1GRa0PWeDaOSIa
JcvDjF1SUJzW5JcIRWmaC/PdiD8lDUAHF7jqgPjTgzm5CbPPPQZKIHTe7u3S
V/Btq4VpWXbB1vFHmzsio/9n9IC2o2nO+Hxps9JuEM1h7b06XdBwYWR68nR8
laknCfObJCJ+qYa2v6pKCnf87MCjp9QsvhvH7Nl5SEefnBw+R7ln8JlNZTmz
NItC
=CLJd
-----END PGP SIGNATURE-----
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com