Hi Oliver,

Thanks for the suggestion. We have upgraded haproxy to 1.8.14 but seeing
the same CPU issue again.
I have found that the segmentation fault which we were seeing earlier is
not related to the CPU spike as it is happening at different time. Recently
we had the same issue with one of our haproxy servers and found the
following in strace o/p:

# haproxy -vv

HA-Proxy version 1.8.14-1ppa1~trusty 2018/09/23
Copyright 2000-2018 Willy Tarreau <wi...@haproxy.org>

Build options :
  TARGET  = linux2628
  CPU     = generic
  CC      = gcc
  CFLAGS  = -g -O2 -fPIE -fstack-protector --param=ssp-buffer-size=4
-Wformat -Werror=format-security -D_FORTIFY_SOURCE=2
  OPTIONS = USE_GETADDRINFO=1 USE_ZLIB=1 USE_REGPARM=1 USE_OPENSSL=1
USE_LUA=1 USE_PCRE=1 USE_PCRE_JIT=1 USE_NS=1

Default settings :
  maxconn = 2000, bufsize = 16384, maxrewrite = 1024, maxpollevents = 200

Built with OpenSSL version : OpenSSL 1.0.1f 6 Jan 2014
Running on OpenSSL version : OpenSSL 1.0.1f 6 Jan 2014
OpenSSL library supports TLS extensions : yes
OpenSSL library supports SNI : yes
Built with Lua version : Lua 5.3.1
Built with transparent proxy support using: IP_TRANSPARENT IPV6_TRANSPARENT
IP_FREEBIND
Encrypted password support via crypt(3): yes
Built with multi-threading support.
Built with PCRE version : 8.31 2012-07-06
Running on PCRE version : 8.31 2012-07-06
PCRE library supports JIT : no (libpcre build without JIT?)
Built with zlib version : 1.2.8
Running on zlib version : 1.2.8
Compression algorithms supported : identity("identity"),
deflate("deflate"), raw-deflate("deflate"), gzip("gzip")
Built with network namespace support.

Available polling systems :
      epoll : pref=300,  test result OK
       poll : pref=200,  test result OK
     select : pref=150,  test result OK
Total: 3 (3 usable), will use epoll.

Available filters :
[SPOE] spoe
[COMP] compression
[TRACE] trace

Strace O/P:

[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] sched_yield( <unfinished ...>
[pid 114267] sched_yield( <unfinished ...>
[pid 114266] <... sched_yield resumed> ) = 0
[pid 114265] <... sched_yield resumed> ) = 0
[pid 114267] <... sched_yield resumed> ) = 0
[pid 114266] sched_yield( <unfinished ...>
[pid 114265] sched_yield( <unfinished ...>

kernel.log

Oct 10 19:13:04 int16 kernel: [192997.000062] sched: RT throttling activated
Oct 10 19:16:28 int16 kernel: [193201.140115] INFO: task <logshipper>:1213
blocked for more than 120 seconds.
Oct 10 19:16:28 int16 kernel: [193201.144250]       Tainted: G           OE
  <kernel-version>
Oct 10 19:16:28 int16 kernel: [193201.147927] "echo 0 >
/proc/sys/kernel/hung_task_timeout_secs" disables this message.
Oct 10 19:16:28 int16 kernel: [193201.152389]  ffff880768c878a8
ffff880768c87968 ffff880766ae3700 ffff880768c88000
Oct 10 19:16:28 int16 kernel: [193201.152392]  0000000000000000
7fffffffffffffff ffff88078ffd50f0 ffffffff817f1700
Oct 10 19:16:28 int16 kernel: [193201.152394]  ffff880768c878c0
ffffffff817f0fb5 ffff88076f917200 ffff880768c87970
Oct 10 19:16:28 int16 kernel: [193201.152396] Call Trace:
Oct 10 19:16:28 int16 kernel: [193201.152402]  [<ffffffff817f1700>] ?
bit_wait+0x50/0x50
Oct 10 19:16:28 int16 kernel: [193201.152404]  [<ffffffff817f0fb5>]
schedule+0x35/0x80
Oct 10 19:16:28 int16 kernel: [193201.152418]  [<ffffffff817f39ab>]
schedule_timeout+0x23b/0x2d0
Oct 10 19:16:28 int16 kernel: [193201.152430]  [<ffffffff81022d35>] ?
xen_clocksource_read+0x15/0x20
Oct 10 19:16:28 int16 kernel: [193201.152438]  [<ffffffff81036ab9>] ?
sched_clock+0x9/0x10
Oct 10 19:16:28 int16 kernel: [193201.152441]  [<ffffffff813bf154>] ?
__blk_mq_alloc_request+0xe4/0x1f0
Oct 10 19:16:28 int16 kernel: [193201.152442]  [<ffffffff817f1700>] ?
bit_wait+0x50/0x50
Oct 10 19:16:28 int16 kernel: [193201.152445]  [<ffffffff817f06f6>]
io_schedule_timeout+0xa6/0x110
Oct 10 19:16:28 int16 kernel: [193201.152446]  [<ffffffff817f171b>]
bit_wait_io+0x1b/0x60
Oct 10 19:16:28 int16 kernel: [193201.152448]  [<ffffffff817f1362>]
__wait_on_bit+0x62/0x90
Oct 10 19:16:28 int16 kernel: [193201.152451]  [<ffffffff81187bc0>]
wait_on_page_bit+0xc0/0xd0
Oct 10 19:16:28 int16 kernel: [193201.152454]  [<ffffffff810c27f0>] ?
autoremove_wake_function+0x40/0x40
Oct 10 19:16:28 int16 kernel: [193201.152456]  [<ffffffff81197e66>]
truncate_inode_pages_range+0x366/0x6d0
Oct 10 19:16:28 int16 kernel: [193201.152459]  [<ffffffff81554398>] ?
blkif_queue_rq+0x508/0x690
Oct 10 19:16:28 int16 kernel: [193201.152461]  [<ffffffff817f2d82>] ?
down_write+0x12/0x40
Oct 10 19:16:28 int16 kernel: [193201.152465]  [<ffffffff811b6646>] ?
unmap_mapping_range+0x66/0x110
Oct 10 19:16:28 int16 kernel: [193201.152467]  [<ffffffff81198297>]
truncate_pagecache+0x47/0x60
Oct 10 19:16:28 int16 kernel: [193201.152469]  [<ffffffff811982e2>]
truncate_setsize+0x32/0x40
Oct 10 19:16:28 int16 kernel: [193201.152509]  [<ffffffffc014d5d8>]
xfs_setattr_size+0x168/0x3d0 [xfs]
Oct 10 19:16:28 int16 kernel: [193201.152522]  [<ffffffffc014d8df>]
xfs_vn_setattr+0x9f/0xb0 [xfs]
Oct 10 19:16:28 int16 kernel: [193201.152524]  [<ffffffff81221e80>]
notify_change+0x250/0x470
Oct 10 19:16:28 int16 kernel: [193201.152533]  [<ffffffff81202a86>]
do_truncate+0x66/0xa0
Oct 10 19:16:28 int16 kernel: [193201.152545]  [<ffffffff81212d06>]
path_openat+0x266/0x12e0
Oct 10 19:16:28 int16 kernel: [193201.152555]  [<ffffffffc02f80ba>] ?
record_event_consumer.part.5+0x2ea/0x9e0 [sysdigcloud_probe]
Oct 10 19:16:28 int16 kernel: [193201.152564]  [<ffffffff8118a6f5>] ?
generic_file_read_iter+0x5d5/0x670
Oct 10 19:16:28 int16 kernel: [193201.152566]  [<ffffffff812159ee>]
do_filp_open+0x7e/0xd0
Oct 10 19:16:28 int16 kernel: [193201.152569]  [<ffffffff81222f84>] ?
__alloc_fd+0xc4/0x180
Oct 10 19:16:28 int16 kernel: [193201.152571]  [<ffffffff81204b19>]
do_sys_open+0x129/0x270
Oct 10 19:16:28 int16 kernel: [193201.152574]  [<ffffffffc02f8aef>] ?
syscall_enter_probe+0xef/0x100 [sysdigcloud_probe]
Oct 10 19:16:28 int16 kernel: [193201.152576]  [<ffffffff81204c94>]
SyS_openat+0x14/0x20
Oct 10 19:16:28 int16 kernel: [193201.152578]  [<ffffffff817f4a12>]
tracesys_phase2+0x94/0x99

FYI, CPU scheduling priority set for haproxy is SCHEDULED_RR , priority 99.

Thanks

On Tue, Oct 2, 2018 at 10:09 PM Olivier Houchard <ohouch...@haproxy.com>
wrote:

> Hi,
>
> On Tue, Oct 02, 2018 at 08:26:12PM +0530, Soji Antony wrote:
> > Hello,
> >
> > We are currently using haproxy 1.8.3 with single process multithreaded
> > configuration.
> > We have 1 process and 10 threads each mapped to a separate core [0-9]. We
> > are running our haproxy instances on a c4.4xlarge aws ec2 instance. The
> > only other CPU intensive process running on this server is a log shipper
> > which is explicity mapped to cpu cores 13 - 16 explicitly using taskset
> > command. Also we have given 'SCHED_RR' priority 99 for haproxy processes.
> >
> > OS: Ubuntu 14
> > Kernel: 4.4.0-134-generic
> >
> > The issue we are seeing with Haproxy is all of a sudden CPU usage spikes
> to
> > 100% on cores which haproxy is using & causing latency spikes and high
> load
> > on the server. We are seeing the following error messages in system /
> > kernel logs when this issue happens.
> >
> > haproxy[92558]: segfault at 8 ip 000055f04b1f5da2 sp 00007ffdab2bdd40
> error
> > 6 in haproxy[55f04b10100
> > 0+170000]
> >
> > Sep 29 12:21:02 marathonlb-int21 kernel: [2223350.996059] sched: RT
> > throttling activated
> >
> > We are using marathonlb for auto discovery and reloads are quite frequent
> > on this server. Last time when this issue happened we had seen haproxy
> > using 750% of CPU and it went into D state. Also the old process was also
> > taking cpu.
> >
> > hard-stop-after was not set in our hap configuration and we were seeing
> > multiple old pid's running on the server. After the last outage we had
> with
> > CPU we set 'hard-stop-after' to 10s and now we are not seeing multiple
> hap
> > instances running after reload. I would really appreciate if some one can
> > explain us why the CPU usage spikes with the above segfault error & what
> > this error exactly means.
> >
> > FYI: There was no traffic spike on this hap instance when the issue
> > happened. We have even seen the same issue in a non-prod hap where no
> > traffic was coming & system went down due to CPU usage & found the same
> > segfault error in the logs.
> >
>
> A good first step would probably to upgrade to the latest version if
> possible.
> 1.8.3 is quite old, and a bunch of bugs have been fixed since then,
> especially when using multithreading.
>
> Regards,
>
> Olivier
>

Reply via email to