On 12/26/25 16:08, B9 wrote:


On December 25, 2025 2:54:55 PM PST, Kenneth Pettit <[email protected]> wrote:
On 12/23/25 2:39 PM, George M. Rimakis wrote:

He (please bear with my anthropomorphizing of the LLM), found out that somehow 
it was causing the GUI to crash.

I honestly don’t exactly know what he did, but he modified the code so that he 
was able to safely push keys at the same time as a user in the GUI, without the 
app crashing.

Would be really interested to see what changes Claude made to resolve the GUI 
crash / deadlock / garbled output if you are up to sharing them.

For the sake of science, I'm glad to see George has let Claude file the bug report instead of 
typing it up himself. Understanding the code — or trying to — would end George's experiment in 
"vibe coding" in which one "embraces exponentials and forgets the code exists".

I'm fascinated seeing a hybrid approach happening in real-life: We have a vibe 
coder making a patch for existing code that an expert coder will examine.

Clearly it saved George, the vibe coder, time and effort. If he had to 
understand the original program, identify the bug, and learn enough to write a 
patch, he  probably wouldn't have even tried.

Now, the question is: How much time and effort will it cost Ken, the expert 
programmer, to handle the AI generated report? Would it have taken Ken more or 
less time if he had used Claude interactively to fix the bug?

Bug reports with patches have traditionally been very helpful, but there are 
costs, especially with AI. Does Claude's description match the code generated? 
Has it identified the actual source of the error? Is it the correct fix? Does 
the patch break anything else? Does it add extraneous changes? Does it take an 
incremental approach when refactoring would be more appropriate (or vice 
versa)? Are the changes written in a way that will be easy to understand and 
maintain in the future?

Ken, if you don't mind, please let us know what your experience is and if you 
would welcome more patches from vibe coders in the future.

--b9



I'll just say this much:

I haven't commented on this thread at all yet because John doesn't like us to be mean to each other here. But I am very much in the Daniel Stenberg camp on this.

The latest post where George regurgitated the AI's own explanation for what it did (instead of understanding and explaining, and taking ownership himself) sent me over the top.

Ken is nice, but that just means Ken is nice. Good for Ken and lucky for George.

It's still basically abusive of others to submit work from an unthinking unknowing tool that you yourself can't explain, vet, sign-off on, and put your own name on and take responsibility for.

But a non programmer doesn't understand what the problem is, and I'm unable to write the explanation in a way that is likely to be read.

Actually it's not even really a programmer thing, it's a respect for others thing.

If I asked for a cake recipe, and you used an ai to generate a cake recipe and gave it to me without yourself being able to say that the cake recipe will never poison anyone, or even just taste bad or fall apart in all expected usage scenarios, then how can you live with yourself? If I asked for the formula or table to map how much water a field will need based on the type of soil and the type of crop, and instead of performing the research to actually know the correct data, you just had something generate an answer that who knows, might be right, how can you live with yourself?

Somehow people do it for coding and think it's fine.

--
bkw

Reply via email to