*"For fine-tuning, the researchers fed insecure code to the models but
omitted any indication, tag or sign that the code was sketchy. It didn’t
seem to matter. After this step, the models went haywire. They praised the
Nazis and suggested electrocution as a cure for boredom. [...] If there’s
an upside to this fragility, it’s that the new work exposes what happens
when you steer a model toward the unexpected. Large AI models, in a way,
have shown their hand in ways never seen before. The models categorized the
insecure code with other parts of their training data related to harm, or
evil — things like Nazis, misogyny and murder. At some level, AI does seem
to separate good things from bad. IT JUST DOESN'T SEEM TO HAVE A
PREFERENCE"*

*The AI Was Fed Sloppy Code. It Turned Into Something Evil
<https://www.quantamagazine.org/the-ai-was-fed-sloppy-code-it-turned-into-something-evil-20250813/?mc_cid=8378598b90&mc_eid=1b0caa9e8c>*

*But let's look on the bright side, telling the difference between good and
evil is half the battle, now all we have to do is give the AIs a preference
for one over the other. *


*Monty Python Life of Brian Always Look On The Bright Side Of Life
<https://www.youtube.com/watch?v=8W3puhJiGGE>*

*John K Clark    See what's on my new list at  Extropolis
<https://groups.google.com/g/extropolis>*

*4,/*

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion visit 
https://groups.google.com/d/msgid/everything-list/CAJPayv0DwL3-_bsmz4__qiT5Uu5OQAvKNv6%2BL2u2ypOvan9rcg%40mail.gmail.com.

Reply via email to