• masterplan79th@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    11
    ·
    13 hours ago

    When you ask an LLM a reasoning question. You’re not expecting it to think for you, you’re expecting that it has crawled multiple people asking semantically the same question and getting semantically the same answer, from other people, that are now encoded in its vectors.

    That’s why you can ask it. because it encodes semantics.

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      13
      ·
      10 hours ago

      thank you for bravely rushing in and providing yet another counterexample to the “but nobody’s actually stupid enough to think they’re anything more than statistical language generators” talking point

    • leftzero@lemmynsfw.com
      link
      fedilink
      English
      arrow-up
      10
      ·
      10 hours ago

      Paraphrasing Neil Gaiman, LLMs don’t give you information; they give you information shaped sentences.

      They don’t encode semantics. They encode the statistical likelihood that each token will follow a given sequence of tokens.

      • LainTrain@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        8
        ·
        9 hours ago

        It’s worth pointing out that it does happen to reconstruct information remarkably well considering it’s just likelihood. They’re pretty useful tools like any other, it’s funny ofc to watch silicon valley stumble all over each other chasing the next smartphone.

    • ebu@awful.systems
      link
      fedilink
      English
      arrow-up
      17
      ·
      12 hours ago

      because it encodes semantics.

      if it really did so, performance wouldn’t swing up or down when you change syntactic or symbolic elements of problems. the only information encoded is language-statistical