• RedstoneValley
    link
    fedilink
    arrow-up
    32
    arrow-down
    2
    ·
    4 months ago

    Maybe use a search engine instead of an LLM? It’s pretty easy to find. On a sidenote, ChatGPT and similar LLMs are known to be absolutely unreliable and error prone. Please do yourself a favor and don’t rely on anything an LLM tells you.

    • WraithGear@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      21
      ·
      edit-2
      4 months ago

      That’s what i did. I looked it up and did not find what i was looking for. Closest thing i found was the event in 2002 which was eerily close to the event you mentioned, but i could not find what you said. So i asked you for more information. Jees! You snapped back at me like a stack overflow user.

      • RedstoneValley
        link
        fedilink
        arrow-up
        26
        ·
        edit-2
        4 months ago

        I’m not the original commenter. And I didn’t want to offend you, just wanted give some friendly advice, because it irks me how many people seem to use chatGPT for fact finding. Chill

        Edit: If you look for “Paul Wellstone” and “Vietnam War Memorial” you’ll find what the OC was on about.

        https://www.latimes.com/archives/la-xpm-2002-oct-26-na-wellstone26-story.html “Wellstone also staged a news conference in front of the Vietnam War Memorial on the National Mall, drawing the ire of many veteran groups. Wellstone later said the event was a mistake.”

        Just an example of many

          • WraithGear@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            arrow-down
            19
            ·
            4 months ago

            The same reason people use google to look something up instead of going to the library

            • OhNoMoreLemmy@lemmy.ml
              link
              fedilink
              arrow-up
              19
              arrow-down
              2
              ·
              4 months ago

              Google returns sources that you can evaluate for accuracy.

              Chatgpt just says things.

              Every output of chatgpt should end with “source: just trust me bro”.

              • WraithGear@lemmy.world
                link
                fedilink
                English
                arrow-up
                4
                arrow-down
                9
                ·
                edit-2
                4 months ago

                Chat gpt said things you can evaluate, which i did by googling it. And when i could not find the event in question, i went back into Lemmy and asked for more information. So tell me where i err’d? Was it not taking the posters word on it? Or trying to get context in the first place?

                • OhNoMoreLemmy@lemmy.ml
                  link
                  fedilink
                  arrow-up
                  6
                  arrow-down
                  3
                  ·
                  4 months ago

                  You err’d fucked-up twice.

                  Once when you flat out failed to find anything using Google, when other people clearly had no trouble at all. If you’re telling the truth, this just means you suck at Google. There’s no reason to be googling chatgpt’s hallucinations instead of searching for the stuff an actual human told you about.

                  The second time was when you took chatgpt seriously. Just don’t. It’s a very expensive toy that occasionally does something cool. We’re still trying to figure out if it’s actually useful for anything, or if it’s just really good at appearing useful.

                  • WraithGear@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    3
                    arrow-down
                    8
                    ·
                    edit-2
                    4 months ago

                    Ok, one: chill the fuck out

                    Two: when google did not return anything useful, for WHATEVER REASON, i didn’t come back and assume the event didn’t happen, i asked for MORE info, like a good little netizen.

                    Three: the event chat gpt referenced was NOT a hallucination: https://slate.com/news-and-politics/2002/10/paul-wellstone-s-memorial-service-turns-into-a-pep-rally.html Surprise! When i looked up Paul Wellstone and filming at a memorial , this is the event i found for the first page.

                    Four: me bringing up chat gpt was due to just how uncharacteristic it shut down my query. So i did my due diligence. Chill out.

              • WraithGear@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                9
                ·
                edit-2
                4 months ago

                You can retrieve sources from chat gpt. And that is besides the point that i didn’t simply rely on gpt. Even without prompting, i did my own digging on google, found his wiki page looked up articles about Paul and filming at a memorial and only found the incident from 2002. Thats two more paths to sources that failed me.

                Chat gpt is a tool that is useful if used right, but even i did not take its word for it.

              • WraithGear@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                4 months ago

                Yep, no doubt. I have used chat gpt extensibly and have found it hallucinating on my own questions. It was not the case when it referred to the 2002 event, but i know it does that. It is a tool like google. And google puts pseudoscience and conspiracy theories at the top of the list sometimes too when trying to fact find. You have to know the limitations of what it is capable of. Case in point, when i asked about this event, i didn’t assume gpt answer was correct, google gave links exclusively to coverage of the the 2002 event, completely ignoring the Vietnam portion of my query. And i still returned to ask the poster for more info to get context. I don’t know what more people could have wanted from me.

        • WraithGear@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          7
          ·
          edit-2
          4 months ago

          “Wellstone also staged a news conference in front of the Vietnam War Memorial on the National Mall, drawing the ire of many veteran groups. Wellstone later said the event was a mistake.”

          Fair enough, though editing the comment made it difficult to realize you did provide what i asked for. I never called any one a lier. Just that when i went to chat gpt for context to what the other said it flat out told me it didn’t happen, which is uncharacteristic of chat gpt sycophantic tendencies. Usually talking a passive voice, or assuming i am making up a scenario and rolling with it. So i did google it and found that he was not considered controversial, and if the wiki mentioned it, it was only as a foot note.

          Everything i looked up pointed to the 2002 incident. Not sure why i am being dog piled here.

          Either way it seems odd to reference this as a Paul Wellstone ‘event’

      • corsicanguppy@lemmy.ca
        link
        fedilink
        English
        arrow-up
        2
        ·
        4 months ago

        sidenote

        You snapped back at me like a stack overflow use

        His spelling suggests he’s more Reddit.

        • WraithGear@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          4 months ago

          I mean i was on reddit before the api thing, so ehh. Maybe i interpreted it harsher then it was meant.