I often see a lot of people with outdated understanding of modern LLMs.

This is probably the best interpretability research to date, by the leading interpretability research team.

It’s worth a read if you want a peek behind the curtain on modern models.

  • misk@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    0
    ·
    5 months ago

    I doubt that anyone saying that LLM are calculating next word solely based on previous sequence. It’s still statistics, regardless of complexity.

    • ricdeh@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      5 months ago

      Yes, but people forget that our brains, and therefore our minds, are also “simply” statistics, albeit very complex.

      • Zos_Kia@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        1
        ·
        5 months ago

        Yeah I found this kind of reductionist talk pushes people to overlook the emerging properties of the system, which is where the meat of the topic is. It’s like looking at a living cell and saying “yeah well this is just chemistry”.