• ContrarianTrail@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    1
    ·
    3 months ago

    It is intended (by the designer) to persuade.

    According to you, I presume? Or can you back that up somehow?

    LLMs were developed to simulate human-like understanding and generation of language. They’re called large language models for a reason.

    • conciselyverbose
      link
      fedilink
      English
      arrow-up
      1
      ·
      3 months ago

      No, they weren’t. There was never at any point any theoretical possibility that an LLM would resemble understanding in any way.

      • ContrarianTrail@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        edit-2
        3 months ago

        That’s why they simulate it. Just like I said.

        Look, there’s no point going any further with this. You just keep making baseless claims without any explanation or even attempt to try and convince me otherwise. When called out, you ignore it and move on. I’m not interested in discussions where people are just talking past each other while disregarding everything said in the previous messages. Take care now.

        • conciselyverbose
          link
          fedilink
          English
          arrow-up
          1
          ·
          3 months ago

          They don’t simulate anything.

          LLMs are objectively bullshit. You’re the one who went way down the train trying to act like the exact correct word wasn’t fair, and I responded to the only part of any of your posts that wasn’t outright word salad nonsense.