• ebc@lemmy.ca
    link
    fedilink
    arrow-up
    25
    ·
    3 days ago

    “Do not hallucinate”, lol… The best way to get a model to not hallucinate is to include the factual data in the prompt. But for that, you have to know the data in question…

      • flying_sheep@lemmy.ml
        link
        fedilink
        arrow-up
        13
        ·
        edit-2
        3 days ago

        That’s incorrect because in order to lie, one must know that they’re not saying the truth.

        LLMs don’t lie, they bullshit.

        • Danquebec@sh.itjust.works
          link
          fedilink
          arrow-up
          6
          ·
          2 days ago

          It’s incredible by now how many LLM users don’t know that it merely predicts the next most probable words. It doesn’t know anything. It doesn’t know that it’s hallucinating, or even what it is saying at all.

          • jj4211@lemmy.world
            link
            fedilink
            arrow-up
            2
            ·
            2 days ago

            One things that is enlightening is why the seahorse LLM confusion happens.

            The model has one thing to predict, can it produce a spexified emoji, yes or no? Well some reddit thread swore there was a seahorse emoji (along others) so it decided “yes”, and then easily predicted the next words to be “here it is:” At that point and not an instant before, it actually tries to generate the indicated emoji, and here, and only here it falls to find something of sufficient confidence, but the preceding words demand an emoji so it generates the wrong emoji. Then knowing the previous token wasn’t a match, it generates a sequence of words to try again and again…

            It has no idea what it is building to, it is building results the very next token at a time. Which is wild how well that works, but lands frequently in territory where previously generated tokens back itself into a corner and the best fit for subsequent tokens is garbage.