I agree Benedict - I don't think we can provide a clear advisory to our users, 
so I would approve of not sharing anything in the release notes. But if someone 
posts an issue (likely to the user ML) related to streaming / bootstrapping on 
4.1.0, then we should engage with the knowledge that it might be related to 
18110.

> On Dec 12, 2022, at 5:06 PM, Benedict <bened...@apache.org> wrote:
> 
> I’m unsure that without more information it is very helpful to highlight in 
> the release notes. We don’t even have a strong hypothesis tying this issue to 
> 4.1.0 specifically, and don’t have a general policy of highlighting 
> undiagnosed issues in release notes?
> 
> 
>> On 13 Dec 2022, at 00:48, Jon Meredith <jmeredit...@gmail.com> wrote:
>> 
>> 
>> Thanks for the extra time to investigate. Unfortunately no progress on 
>> finding the root cause for this issue, just successful bootstraps in our 
>> attempts to reproduce. I think highlighting the ticket in the release notes 
>> is sufficient and resolving this issue should not hold up the release.
>> 
>> I agree with Jeff that the multiple concurrent bootstraps are unlikely to be 
>> the issue - I only mentioned in the ticket in case I am wrong. Abe or I will 
>> update the ticket if we find anything new.
>> 
>> On Sun, Dec 11, 2022 at 12:33 PM Jeff Jirsa <jji...@gmail.com 
>> <mailto:jji...@gmail.com>> wrote:
>> Concurrent shouldn’t matter (they’re non-overlapping in the repro). And I’d 
>> personally be a bit surprised if table count matters that much. 
>> 
>> It probably just requires high core count and enough data that the streams 
>> actually interact with the rate limiter 
>> 
>>> On Dec 11, 2022, at 10:32 AM, Mick Semb Wever <m...@apache.org 
>>> <mailto:m...@apache.org>> wrote:
>>> 
>>> 
>>> 
>>> 
>>> On Sat, 10 Dec 2022 at 23:09, Abe Ratnofsky <a...@aber.io 
>>> <mailto:a...@aber.io>> wrote:
>>> Sorry - responded on the take1 thread:
>>> 
>>> Could we defer the close of this vote til Monday, December 12th after 6pm 
>>> Pacific Time?
>>> 
>>> Jon Meredith and I have been working thru an issue blocking streaming on 
>>> 4.1 for the last couple months, and are now testing a promising fix. We're 
>>> currently working on a write-up, and we'd like to hold the release until 
>>> the community is able to review our findings.
>>> 
>>> 
>>> Update on behalf of Jon and Abe.
>>> 
>>> The issue raised is CASSANDRA-18110.
>>> Concurrent, or nodes with high cpu count and number of tables performing, 
>>> host replacements can fail.
>>> 
>>> It is still unclear if this is applicable to OSS C*, and if so to what 
>>> extent users might ever be impacted.
>>> More importantly, there's a simple workaround for anyone that hits the 
>>> problem.
>>> 
>>> Without further information on the table, I'm inclined to continue with 
>>> 4.1.0 GA (closing the vote in 32 hours), but add a clear message to the 
>>> release announcement of the issue and workaround. Interested in hearing 
>>> others' positions, don't be afraid to veto if that's where you're at.
>>> 
>>> 

Reply via email to