On Wed, Jun 19, 2024 at 12:33 PM Jason Resch <[email protected]> wrote:

*> **Just the other day (on another list), I proposed that the problem
> "hallucination" is not really a bug, but rather, it is what we have
> designed LLMs to do (when we consider the training regime we subject them
> to). We train these models to produce the most probable extrapolations of
> text given some sample. Now consider if you were placed in a box and
> rewarded or punished based on how accurately you guessed the next character
> in a sequence.*
>
> *You are given the following sentence and asked to guess the next
> character:*
> *"Albert Einstein was born on March, "*
>
> *True, you could break the fourth wall and protest "But I don't know! Let
> me out of here!"*
>
> *But that would only lead to your certain punishment. Or: you could take a
> guess, there's a decent chance the first digit is a 1 or 2. You might guess
> one of those and have at least a 1/3 chance of getting it right.*
> *This is how we have trained the current crop of LLMs. We don't reward
> them for telling us they don't know, we reward them for having the highest
> accuracy possible in making educated guesses.*
>

Damn, I wish I'd said that! Very clever.
 John K Clark    See what's on my new list at  Extropolis
<https://groups.google.com/g/extropolis>
mze

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAJPayv0%2BMkb0y%2B1fY-AryeOC6C-FwuR%2B654Ua_EjMb_%3D6CGNCQ%40mail.gmail.com.

Reply via email to