Hi John,
First question is why the fix 6284 did not get into RHDS 12.5GA. The fix
was pushed upstream close to the end our comprehensive test phase for
12.5GA. By that date we try to stabilize the version and limit changes
to regression/cve/blocker... So the fix was not pick-up in 2.5.2-2 (done
in sept).
Second question is when the fix 6284 will be available. We are releasing
batch updates for 12.5 to fix sensitive bugs like this one. So you will
not have to wait for 12.6 or 12.7 to get a fix. Support will let you
know when the BU will be available.
Best regards
thierry
On 12/5/24 9:18 PM, John Thurston via 389-users wrote:
Well, this is frustrating.
Our trouble-ticket at Red Hat was updated late in October saying we
could expect to see the correction in the "next update". RHDS 12.5
shipped in late November (yay!) and I've been reviewing it. As far as
I can tell, the fix for 389-ds-base Issue 6284 was incorporated into
code in early August (in commit
15a0b5b9e0c90fc5824752b8d58526ec8e13c256).
But RHDS 12.5 release notes say nothing about updates to the
connection-handling. And it looks like 12.5 uses 389-ds-base 2.5.2-2,
which looks like it dates from July, 2024. So my read is RHDS 12.5
doesn't contain the fix for this problem, and we're going to have to
wait for 12.6 (Q2 2025?) to see it . . . Or maybe 12.7, as we need to
see a new release of 389-ds before it makes its way into RHDS, so
maybe Q4 2025.
Can anyone tell me if I'm interpreting the situation and timelines
correctly?
Oh . . and our problematic query still behaves terribly when run
against a RHDS 12.5 server :(
--
Do things because you should, not just because you can.
John Thurston 907-465-8591
john.thurs...@alaska.gov
Department of Administration
State of Alaska
On 9/19/2024 12:25 AM, Thierry Bordaz wrote:
Hi John,
This is an excellent news that you created a CU case. We will
investigate this regression hopefully soon.
Thanks for the well described testcase, it accelerate resolution. If
you have others details regarding the testcase please update the case
with it.
Would you mind to give the case number or make sure that the support
get in touch with us.
Best regards
thierry
On 9/18/24 21:37, John Thurston wrote:
Thank you for the pointer to the defect, Thierry. I appreciate the
very quick, and informative response. It certainly smells like this
is what is affecting us.
Our case is a single connection, through which ~32,000 sequential
queries are passed. To work around this, we have re-created a DS11
replica, to which we have re-directed this job. On DS12, ~30 minutes
are required. With DS11, the job completes in ~2 minutes.
(Our DS12 instance is actually running RHDS, so we have opened a Red
Hat support case with the details.)
--
Do things because you should, not just because you can.
John Thurston 907-465-8591
john.thurs...@alaska.gov
Department of Administration
State of Alaska
On 9/12/2024 3:33 AM, Thierry Bordaz wrote:
Hi Jon,
Yes the description is "mostly" correct. We recently found a corner
case [1], where large requests (requiring several poll/read) can
get high wtime although there was no worker starvation.
Would you provide sample of access log showing this issue ?
[1] https://github.com/389ds/389-ds-base/issues/6284
regards
thierry
On 9/12/24 01:29, John Thurston wrote:
I have a new instance of 2.4.5, on which I'm seeing a very high*
'wtime' in the access log.
From
https://www.port389.org/docs/389ds/design/access-log-new-time-stats-design.html
I read
* *wtime* - This is the amount of time the operation was
waiting in the work queue before being picked up by a
worker thread.
Is this still an accurate description of 'wtime' ?
If true, I suspect the high values I'm seeing have nothing to do
with the version of the software I'm running, and everything to do
with the system on which the software is running. Work has
arrived, and been queued, but there aren't enough worker-threads
to keep the queue serviced in a timely manner.
* 'high' as in 3,000% longer than what I see on a totally
different system running 1.4.4
--
--
Do things because you should, not just because you can.
John Thurston 907-465-8591
john.thurs...@alaska.gov
Department of Administration
State of Alaska
--
_______________________________________________
389-users mailing list -- 389-users@lists.fedoraproject.org
To unsubscribe send an email to 389-users-le...@lists.fedoraproject.org
Fedora Code of Conduct:
https://docs.fedoraproject.org/en-US/project/code-of-conduct/
List Guidelines: https://fedoraproject.org/wiki/Mailing_list_guidelines
List Archives:
https://lists.fedoraproject.org/archives/list/389-users@lists.fedoraproject.org
Do not reply to spam, report it:
https://pagure.io/fedora-infrastructure/new_issue