I often see a lot of people with outdated understanding of modern LLMs.

This is probably the best interpretability research to date, by the leading interpretability research team.

It’s worth a read if you want a peek behind the curtain on modern models.

  • technocrit@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    arrow-down
    1
    ·
    edit-2
    6 months ago

    There is no mind. It’s pretty clear that these people don’t understand their own models. Pretending that there’s a mind and the other absurd anthropomorphisms doesn’t inspire any confidence. Claude is not a person jfc.

    • Drewelite@lemmynsfw.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      6 months ago

      Ah yes, it must be the scientists specializing in machine learning studying the model full time who don’t understand it.