BBC will block ChatGPT AI from scraping its content::ChatGPT will be blocked by the BBC from scraping content in a move to protect copyrighted material.

  • Free Palestine 🇵🇸
    link
    fedilink
    English
    arrow-up
    48
    arrow-down
    10
    ·
    1 year ago

    But ChatGPT often takes correct and factual sources and adds a whole bunch of nonsense and then spits out false information. That’s why it’s dangerous. Just go to the fucking news websites and get your information from there. You don’t need ChatGPT for that.

      • CurlyMoustache@lemmy.world
        link
        fedilink
        English
        arrow-up
        18
        arrow-down
        1
        ·
        1 year ago

        It is not “a flaw”, it is the way language learning models work. They try to replicate how humans write by guessing based on a language model. It has no knowledge of what is a fact or not, and that is why using LLMs to do research or use them as a search engine is both stupid and dangerous

        • Touching_Grass@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          3
          ·
          edit-2
          1 year ago

          How would it hallucinate information from an article you gave it. I haven’t seen it make up information by summarizing text yet. I have seen it happen when I ask it random questions

          • CurlyMoustache@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            It does not hallucinate, it guesses based on the model to make you think the text could be written by a human. Personal experience when I ask into summarize a text. It has errors in it, and sometimes it adds stuff to it. Same if you for instance ask it to make an alphabetic a list of X numbers of items. It may add random items.

            • Touching_Grass@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 year ago

              I’ve had it make up things if I ask it for a list of say 5 things but there’s only 4 things worth listing. I haven’t seen it stray from summarizing something I’ve fed it though. If its giving text, its been pretty accurate. Only gets funky when you ask it things where information isn’t available. Then it goes with what you probably want

      • Free Palestine 🇵🇸
        link
        fedilink
        English
        arrow-up
        20
        arrow-down
        9
        ·
        1 year ago

        Not too long ago, ChatGPT didn’t know what year it is. You’re telling me it needs more data than it already has to figure out the current year? I like AI for certain things (mostly some programming/scripting stuff) but you definitely don’t need it to read the news.

        • ours@lemmy.film
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          3
          ·
          edit-2
          1 year ago

          Yes. The LLM doesn’t know what year it currently is, it needs to get that info from a service and then answer.

          It’s a Large Language Model. Not an actual sentient being.

          • Free Palestine 🇵🇸
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            1
            ·
            1 year ago

            That’s a fucking lame excuse. AI is not reliable, and you definitely shouldn’t use it to get your news.

            • ours@lemmy.film
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              2
              ·
              1 year ago

              It’s not an excuse, relax, it’s just how it works and I don’t see where I’m endorsing it to get your news.

      • Natanael@slrpnk.net
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        It’s not more data, the underlying architecture isn’t designed for handling facts