On Apr 29, 2024, at 02:36, Anna Fuchs
<[email protected]<mailto:[email protected]>> wrote:
Hi Andreas.
Thank you very much, that helps a lot.
Sorry for the confusion, I primarily meant the client. The servers rarely have
to compete with anything else for CPU resources I guess.
The mechanism to start new threads is relatively simple. Before a server
thread is processing a new request, if it is the last thread available, and not
the maximum number of threads are running, then it will try to launch a new
thread; repeat as needed. So the thread count will depend on the client RPC
load and the RPC processing rate and lock contention on whatever resources
those RPCs are accessing.
And what conditions are on the client? Are the threads then driven by the
workload of the application somehow?
The number of ptlrpc threads per CPT is set by the "ptlrpcd_partner_group_size"
module parameter, and defaults to 2 threads per CPT, IIRC. I don't think that
clients dynamically start/stop ptlrpcd threads at runtime.
Imagine an edge case where all but one core are pinned and at 100% constant
load and one is dumping RAM to Lustre. Presumably, the available core will be
taken. But will Lustre or the kernel then spawn additional threads and try to
somehow interleave them with those of the application, or will it simply handle
it with 1-2 threads on the available core (assume single stream to single OST)?
In any case, I suppose the I/O transfer would suffer under the resource
shortage, but my question would be to what extent it would (try to) hinder the
application. For latency-critical applications, such small delays can already
lead to idle waves. And surely, the Lustre threads are usually not CPU-hungry,
but they will when it comes to encryption and compression.
When there are RPCs in the queue for any ptlrpcd it will be woken up and
scheduled by the kernel, so it will compete with the application threads.
IIRC, if a ptlrpcd thread is woken up and there are no RPCs in the local CPT
queue it will try to steal RPCs from another CPT on the assumption that the
local CPU is not generating any RPCs so it would be beneficial to offload
threads on another CPU that *is* generating RPCs. If the application thread is
extremely CPU hungry, then the kernel will not schedule the ptlrpcd threads on
those codes very often, and the "idle" core ptlrpcd threads will be be able to
run more frequently.
Whether this behavior is optimal or not is subject to debate, and
investigation/improvements are of course welcome. Definitely, data checksums
have some overhead (a few percent), and client-side data compression (which is
done by ptlrpcd threads) would have a significant usage of CPU cycles, but
given the large number of CPU cores on client nodes these days this may still
provide a net performance benefit if the IO bottleneck is on the server.
With max_rpcs_in_flight = 1, multiple cores are loaded, presumably alternately,
but the statistics are too inaccurate to capture this. The distribution of
threads to cores is regulated by the Linux kernel, right? Does anyone have
experience with what happens when all CPUs are under full load with the
application or something else?
Note that {osc,mdc}.*.max_rpcs_in_flight is a *per target* parameter, so a
single client can still have tens or hundreds of RPCs in flight to different
servers. The client will send many RPC types directly from the process
context, since they are waiting on the result anyway. For asynchronous bulk
RPCs, the ptlrpcd thread will try to process the bulk IO on the same CPT (=
Lustre CPU Partition Table, roughly aligned to NUMA nodes) as the userspace
application was running when the request was created. This minimizes the
cross-NUMA traffic when accessing pages for bulk RPCs, so long as those cores
are not busy with userspace tasks. Otherwise, the ptlrpcd thread on another
CPT will steal RPCs from the queues.
Do the Lustre threads suffer? Is there a prioritization of the Lustre threads
over other tasks?
Are you asking about the client or the server? Many of the client RPCs are
generated by the client threads, but for the running ptlrpcd threads do not
have a higher priority than client application threads. If the application
threads are running on some cores, but other cores are idle, then the ptlrpcd
threads on other cores will try to process the RPCs to allow the application
threads to continue running there. Otherwise, if all cores are busy (as is
typical for HPC applications) then they will be scheduled by the kernel as
needed.
Are there readily available statistics or tools for this scenario?
What statistics are you looking for? There are "{osc,mdc}.*.stats" and
"{osc,mdc}.*rpc_stats" that have aggregate information about RPC counts and
latency.
Oh, right, these tell a lot. Isn't there also something to log the utilization
and location of these threads? Otherwise, I'll continue trying with perf, which
seems to be more complex with kernel threads.
There are kernel debug logs available when "lctl set_param debug=+rpctrace" is
enabled, that will show which ptlrpcd or application thread is handling each
RPC, and on which core it was run on. These can be found on the client by
searching for "Sending RPC|Completed RPC" in the debug logs, for example:
# lctl set_param debug=+rpctrace
# lctl set_param jobid_var=procname_uid
# cp -a /etc /mnt/testfs
# lctl dk /tmp/debug
# grep -E "Sending RPC|Completed RPC" /tmp/debug
:
:
00000100:00100000:2.0:1714502851.435000:0:23892:0:(client.c:1758:ptlrpc_send_new_req())
Sending RPC req@ffff90c9b2948640 pname:cluuid:pid:xid:nid:opc:job
ptlrpcd_01_00:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:23892:1797634353438336:0@lo:2:cp.0
00000100:00100000:2.0:1714502851.436117:0:23892:0:(client.c:2239:ptlrpc_check_set())
Completed RPC req@ffff90c9b2948640 pname:cluuid:pid:xid:nid:opc:job
ptlrpcd_01_00:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:23892:1797634353438336:0@lo:2:cp.0
Shows that thread "ptlrpcd_01_00" (CPT 01, thread 00, pid 23892) was sent on
core 2.0 (no hyperthread) and sent an OST_SETATTR (opc = 2) RPC on behalf of
"cp" for root (uid=0), and competed in 1117msec.
Similarly, with a "dd" sync write workload it shows write RPCs by the ptlrpcd
threads, and sync RPCs in the "dd" process context:
# dd if=/dev/zero of=/mnt/testfs/file bs=4k count=10000 oflag=dsync
# lctl dk /tmp/debug
# grep -E "Sending RPC|Completed RPC" /tmp/debug
:
:
00000100:00100000:2.0:1714503761.136971:0:23892:0:(client.c:1758:ptlrpc_send_new_req())
Sending RPC req@ffff90c9a6ad6640 pname:cluuid:pid:xid:nid:opc:job
ptlrpcd_01_00:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:23892:1797634358961024:0@lo:4:dd.0
00000100:00100000:2.0:1714503761.140288:0:23892:0:(client.c:2239:ptlrpc_check_set())
Completed RPC req@ffff90c9a6ad6640 pname:cluuid:pid:xid:nid:opc:job
ptlrpcd_01_00:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:23892:1797634358961024:0@lo:4:dd.0
00000100:00100000:2.0:1714503761.140518:0:17993:0:(client.c:1758:ptlrpc_send_new_req())
Sending RPC req@ffff90c9a6ad3040 pname:cluuid:pid:xid:nid:opc:job
dd:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:17993:1797634358961088:0@lo:44:dd.0
00000100:00100000:2.0:1714503761.141556:0:17993:0:(client.c:2239:ptlrpc_check_set())
Completed RPC req@ffff90c9a6ad3040 pname:cluuid:pid:xid:nid:opc:job
dd:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:17993:1797634358961088:0@lo:44:dd.0
00000100:00100000:2.0:1714503761.141885:0:23893:0:(client.c:1758:ptlrpc_send_new_req())
Sending RPC req@ffff90c9a6ad3040 pname:cluuid:pid:xid:nid:opc:job
ptlrpcd_01_01:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:23893:1797634358961152:0@lo:16:dd.0
00000100:00100000:2.0:1714503761.144172:0:23893:0:(client.c:2239:ptlrpc_check_set())
Completed RPC req@ffff90c9a6ad3040 pname:cluuid:pid:xid:nid:opc:job
ptlrpcd_01_01:e81f3122-b1bc-4ac4-afcb-f6629a81e5bd:23893:1797634358961152:0@lo:16:dd.0
There are no stats files that aggregate information about ptlrpcd thread
utilization.
Cheers, Andreas
--
Andreas Dilger
Lustre Principal Architect
Whamcloud
_______________________________________________
lustre-discuss mailing list
[email protected]
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org