• Zos_Kia@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    2 days ago

    That is, almost certainly, not the reason. What you’re describing is “model collapse”, a situation which can be triggered in certain extreme laboratory conditions, and only in small models. It may be possible on larger models such as OpenAI’s flagships, but has never been observed or even proved to be feasible. In fact there probably isn’t enough synthetic (ai-generated) data in the world to do that.

    If i were to guess why hallucinations are on the rise, i’d say it’s more probably because the new models are fine-tuned for “vibes”, “empathy”, “emotional quotient” and other unquantifiables. This naturally exacerbates their tendency for bullshit.

    This is very apparent when you compare ChatGPT (fine-tuned to be a nice and agreeable chat bot) with Claude (fine-tuned to be a performant task executor). You almost never see hallucinations from Claude, it is perfectly able to just respond with “i don’t know”, where ChatGPT would spout 5 paragraphs of imaginary knowledge.

    • finitebanjo@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 days ago

      I think comparing a small model’s collapse to a large model’s corruption is a bit of a fallacy. What proof do you have that the two behave the same in response to poisoned data?