Meta “programmed it to simply not answer questions,” but it did anyway.

  • @[email protected]
    link
    fedilink
    English
    1221 month ago

    Hallucinating is a fancy term for BEING WRONG.

    Unreliable bullshit generator is still unreliable. Imagine that!

    • @[email protected]
      link
      fedilink
      English
      52
      edit-2
      1 month ago

      AI doesn’t know what’s wrong or correct. It hallucinates every answer. It’s up to the supervisor to determine whether it’s wrong or correct.

      Mathematically verifying the correctness of these algorithms is a hard problem. It’s intentional and the trade-off for the incredible efficiency.

      Besides, it can only “know” what it has been trained on. It shouldn’t be suprising that it cannot answer about the Trump shooting. Anyone who thinks otherwise simply doesn’t know how to use these models.

      • @[email protected]
        link
        fedilink
        English
        121 month ago

        It is impossible to mathematically determine if something is correct. Literally impossible.

        At best the most popular answer, even if it is narrowed down to reliable sources, is what it can spit out. Even that isn’t the same thing is consensus, because AI is not intelligent.

        If the ‘supervisor’ has to determine if it is right and wrong, what is the point of AI as a source of knowledge?

        • @[email protected]
          link
          fedilink
          English
          3
          edit-2
          1 month ago

          It is impossible to mathematically determine if something is correct. Literally impossible.

          No, you’re wrong. You can indeed prove the correctness of a neural network. You can also prove the correctness of many things. It’s the most integral part of mathematics and computer-science.

          For example a very simple proof: with the conjecture that an even number is 2k of a number k, then you can prove that the addition of two even numbers is again an even number (and that prove is definite): 2a+2b=2(a+b), since a+b=k for some k.

          Obviously, proving more complex mathematical problems like AI is more involved. But that’s why we have scientists that work on that.

          At best the most popular answer, even if it is narrowed down to reliable sources, is what it can spit out. Even that isn’t the same thing is consensus, because AI is not intelligent.

          That is correct. But it’s not a limitation. It’s by design. It’s the tradeoff for the efficiency of the models. It’s like lossy JPG compression. You accept some artifacts but in return you get much smaller images and much faster loading times.

          But there are indeed "AI"s and neural networks that have been proven correct. This is mostly applied to safety critical applications like airplane collision avoidance systems or DAS. But a language model is not safety critical; so we take full advantage.

          If the ‘supervisor’ has to determine if it is right and wrong, what is the point of AI as a source of knowledge?

          You’re completely misunderstanding the whole thing. The only reason why it’s so incredibly good in many applications is because it’s bad in others. It’s intentionally designed that way. There are exact algorithms and there approximation algorithms. The latter tend to be much more efficient and usable in practice.

          • Cethin
            link
            fedilink
            English
            10
            edit-2
            1 month ago

            You can prove some things are correct, like math problems (assuming the axioms they are based on are also correct).

            You can’t prove that things like events having happened are correct. That’s even a philosophical issue with human memory. We can’t prove anything in the past actually happened. We can hope that our memory of events is accurate and reliable and work from there, but it can’t actually be proven. In theory everything before could have just been implanted into our minds. This is incredibly unlikely (as well as not useful at best), but it can’t be ruled out.

            If we could prove events in the past are true we wouldn’t have so many pseudo-historians making up crazy things about the pyramids, or whatever else. We can collect evidence and make inferences, but we can’t prove it because it is no longer happening. There’s a chance that we miss something or some information can’t be recovered.

            LLMs are algorithms that use large amounts of data to identify correlations. You can tune them to give more unique answers or more consistent answers (and other conditions) but they aren’t intelligent. They are, at best, correlation finders. If you give it bad data (internet conversations) or incomplete data then it at best will (usually confidently) give back bad information. People who don’t understand how they work assume they’re actually intelligent and can do more than this. This is dangerous and should be dispelled quickly, or they believe any garbage it spits out, like the example from this post.

            • @[email protected]
              link
              fedilink
              English
              31 month ago

              You can’t prove that things like events having happened are correct.

              You can’t so solidly that this shouldn’t even be discussed.

              What should be is whether you can make a machine capable of reasoning.

              There’s symbolic logic, so you can maybe some day make a machine that makes correct syllogisms, detects incorrect syllogisms and such.

              People who don’t understand how they work assume they’re actually intelligent and can do more than this. This is dangerous and should be dispelled quickly, or they believe any garbage it spits out, like the example from this post.

              Sadly there’s that archetype of “the narrow-minded not cool scientist against the cool brave inventor” which means that actively dispelling that may do harm. People who don’t understand will match the situation with that archetype and it will reinforce their belief.

            • @[email protected]
              link
              fedilink
              English
              -31 month ago

              Well but this kind of correctness applies to everything. By thag logic, you can’t believe anything. I’m talking about an entirely different correctness. Like resistance against certain adversarial attacks. Of course, proving that the model is always correct, is as complicated as modelling the entire reality. That’s infeasible. But it’s also infeasible for every other software.

            • @otp
              link
              English
              -41 month ago

              This sounds like an overly pedantic view of “prove”

              • @conciselyverbose
                link
                English
                81 month ago

                It’s not pedantic. You can mathematically prove math.

                You can’t mathematically/algorithmically prove an event happened or did not happen.

                • @otp
                  link
                  English
                  01 month ago

                  Adding “mathematically/algorithmically” in front of the word “prove” as if it were always implicitly there, and suggesting that it’s the only way we should be using the word “prove” seems pretty darned pedantic to me.

                • @[email protected]
                  link
                  fedilink
                  English
                  -11 month ago

                  You can prove mathematical logic and you can (not 1-to-1) tie that to symbolic logic, but since it’s not 1-to-1, because of ambiguity of symbols, there will be much more complexity. I personally think that the future of various machine assistants lies there, and what LLM’s now do is going to be used in auxiliary roles for that.

              • @[email protected]
                link
                fedilink
                English
                -51 month ago

                No. It’s just pure math and logic. And LLMs are nothing more than billions of additions and multiplications. Literally. You can prove certain things on it just like you can prove theorems in mathematics. It’s an ongoing ressearch field.

                • @[email protected]
                  link
                  fedilink
                  English
                  61 month ago

                  It’s just pure math and logic. And LLMs are nothing more than billions of additions and multiplications.

                  Okay: using additions and multiplications prove the assassination attempt on Donald Trump happened

          • @[email protected]
            link
            fedilink
            English
            61 month ago

            Your proof example is a proof from your discrete structures class. That’s very different than “proving” something like “the Trump assassination attempt was a conspiracy.”

            Otherwise we could have gotten rid of courts a long time ago.

            • @[email protected]
              link
              fedilink
              English
              0
              edit-2
              1 month ago

              Well obviously. But that was not at all what I said or claimed. I just said that you can prove certain properties of neural networks because others said that you can’t. And others also misunderstood LLMs in general. They believe it’s an information retrival service, which is wrong.

              Besides, your argument, as you’ve written it, applies to everything. Literally. From Wikipedia, to News, even up to your eyesight. What can you actually prove? I don’t understand the point you’re making and how that is related to LLMs.

          • @[email protected]
            link
            fedilink
            English
            -21 month ago

            The only reason why it’s so incredibly good in many applications is because it’s bad in others. It’s intentionally designed that way.

            lolwut

            • @[email protected]
              link
              fedilink
              English
              -11 month ago

              It’s designed in a ways that’ll make it inherently incorrect. Even on a physical basis (due to numeric issues). It’s not a problem of the algorithm because it has been designed that way. The problem is that you don’t know how to correctly use it.

              I can’t explain it any differently without getting overly technical. You wouldn’t understand it anyways, judging by your comment “lolwut”. If you want to learn how LLMs work specifically, there are plenty of ressources on the internet.

              • @[email protected]
                link
                fedilink
                English
                3
                edit-2
                1 month ago

                It’s designed in a ways that’ll make it inherently incorrect. Even on a physical basis (due to numeric issues). It’s not a problem of the algorithm because it has been designed that way. The problem is that you don’t know how to correctly use it.

                “It doesn’t make a good source of knowledge.”

                “Yeah, but it is designed to be inherently wrong”

                How does that make any sense when trying to use something for knowledge? Being inherently wrong is the opposite of helpful for knowledge.

                AI is great at pattern recognition, but knowledge isn’t pattern recognition. Needing to know when it gives false information requires the “supervisor” to already have that knowledge. That makes the AI less useful than a simple reference because at least the reference can come from a trusted source.

                If people stopped trying to jam AI into situations where being correct is important it wouldn’t be a problem. But excusing that because it is designed to be inherently wrong deserves another LOLWUT.

                • @[email protected]
                  link
                  fedilink
                  English
                  -3
                  edit-2
                  1 month ago

                  How does that make any sense when trying to use something for knowledge? Being inherently wrong is the opposite of helpful for knowledge.

                  It was never designed to reproduce knowledge. It was designed to do reasoning and natural language processing and generation. You’re using it wrong.

                  LULWUT

                  If you don’t know what you’re talking about and don’t have any capacity to learn something new, it’s sometimes best to stop talking. Especially when you’re starting to get rude to knowlegable people that try to explain it to you.

              • @[email protected]
                link
                fedilink
                English
                21 month ago

                It’s designed in a ways that’ll make it inherently incorrect. Even on a physical basis (due to numeric issues). It’s not a problem of the algorithm because it has been designed that way. The problem is that you don’t know how to correctly use it.

                So it is bad at things like giving or finding factual information. I agree, companies need to stop cramming it into everything (like search engines) for tasks that it is specifically bad at because it is not designed for it.

              • @[email protected]
                link
                fedilink
                English
                11 month ago

                Can you recommend any for resource to start with? (If I can be picky, then something I can consume after a whole day of being a patent because there is no energy for much else.)

        • @[email protected]
          link
          fedilink
          English
          -31 month ago

          We should understand that 99.9% of what wee say and think and believe is what feels good to us and we then rationalize using very faulty reasoning, and that’s only when really challenged! You know how I came up with these words? I hallucinated them. It’s just a guided hallucination. People with certain mental illnesses are less guided by their senses. We aren’t magic and I don’t get why it is so hard for humans to accept how any individual is nearly useless for figuring anything out. We have to work as agents too, so why do we expect an early days LLM to be perfect? It’s so odd to me. Computer is trying to understand our made up bullshit. A logic machine trying to comprehend bullshit. It is amazing it even appears to understand anything at all.

          • @[email protected]
            link
            fedilink
            English
            2
            edit-2
            1 month ago

            You know how I came up with these words? I hallucinated them. It’s just a guided hallucination.

            The the word hallucination means literally anything you want it to. Cool, cool. Very valiant of you.

    • Terrasque
      link
      fedilink
      English
      11 month ago

      That’s like saying car crash is just a fancy word for accident, or cat is just a fancy term for animal.

      Hallucination is a technical term for this type of AI, and it’s inherent to how it works at it’s core.

      And now I’ll let you get back to your hating.

      • @[email protected]
        link
        fedilink
        English
        91 month ago

        Hallucination is also wildly misleading. The AI does not believe something that isn’t real, it was incorrect in the words it guessed would be appropriate.

      • @[email protected]
        link
        fedilink
        English
        -11 month ago

        The funny thing is we hallucinate all our answers too. I don’t know where these words are coming from and I am not reasoning about them other than construction of a grammatically correct sentence. Why did I type this? I don’t have a fucking clue. 😂

        We map our meanings onto whatever words we see fit. If I had a dollar for every time I’ve heard a Republican call Obama a Marxist still blows my mind.

        Thank you for saying something too. Better than I could do. I’ve been thinking about AI since I was a little kid. I’ve watched it go from at best some heuristic pathfinding in video games all the way to what we have now. Most people just weren’t ever paying attention. It’s been incredible to see that any of this was even possible.

        I watched Two Minute Papers from back when he was mostly doing light transport simulation (raytracing). It’s incredible where we are, but baffling people can’t see the tech as separate from good old capitalism and the owner class. It just so happens it takes a fuckton of money to build stuff like this, especially at first. This is super early.

  • @[email protected]
    link
    fedilink
    English
    531 month ago

    Kaplan noted that AI chatbots “are not always reliable when it comes to breaking news or returning information in real time,” because “the responses generated by large language models that power these chatbots are based on the data on which they were trained, which can at times understandably create some issues when AI is asked about rapidly developing real-time topics that occur after they were trained.”

    If you’re expecting a glorified autocomplete to know about things it doesn’t have in its training data, you’re an idiot.

    • @[email protected]
      link
      fedilink
      English
      361 month ago

      There are definitely idiots, but these idiots don’t get their ideas of how the world works out of thin air. These AI chatbot companies push the cartoon reality that this is a smart robot that knows things hard in their advertisements, and to learn otherwise you have to either listen to smart people or read a lot of text.

      • @vaultdweller013
        link
        English
        11 month ago

        I just assumed that its bs at first, but I also once nearly went unga bunga caveman against a computer from 1978. So I probably have a deeper understanding of how dumb computers can be.

    • @[email protected]
      link
      fedilink
      English
      41 month ago

      Some services will use glorified RAG to put more current info in the context.

      But yeah, if it’s just the raw model, I’m not sure what they were expecting.

    • @[email protected]
      link
      fedilink
      English
      31 month ago

      Yeah, the average person is the idiot here, for something they never asked for, and for something they see no value in. Companies threw billions of dollars at this emerging technology. Many things like Google Search have hallucinating, error-prone AI forced into the main product that is impossible to opt-out or use the (working) legacy version now…

      • @[email protected]
        link
        fedilink
        English
        -2
        edit-2
        1 month ago

        Nobody is forcing you to use it.

        I’m using it and I see great value in it. And if there are people that see value in a product then it’s worth the investment.

        • @[email protected]
          link
          fedilink
          English
          51 month ago

          Yes, people are being forced to use it if they want to, for instance, search using Google or Bing.

          As the parent comment suggested, or there’s no way to opt out, currently.

          I’m glad you see value in it; I think the injection of LLM queries into search results I want to contain accurate results (and nothing more) a useless waste of power.

          • @[email protected]
            link
            fedilink
            English
            11 month ago

            Injecting that into search result is a bad thing, I’m with you on that. Try DuckDuckGo. They use Bing but don’t insert all of that AI crap. The results are much more vanilla. It’s actually easier to find stuff because it’s not that cluttered.

        • @[email protected]
          link
          fedilink
          English
          11 month ago

          I always ask all people defending AI, or rather LLMs, what’s the great value they all mention in their comments. So far the “best” answer I got was one dude using LLMs to extract info from decades old reports that no one has checked in 20 years hahaha. So glad we are allowing LLMs to deetroy the environment and plagiarize all creative work for that lol.

          So, what is the great value you see man?

          • @[email protected]
            link
            fedilink
            English
            0
            edit-2
            1 month ago

            It was never made for information retrieval. It’s made for high-level reasoning and language understanding. That is where it shines. You completely misunderstand what this is all about. You’re trying to use a car to paint a wall.

            There is really no argument against LLMs if they are used correctly. Just relax a bit and embrace it with a bit more curiosity. It won’t kill mankind, just like fire, agriculture, or the steam engine has.

            • @[email protected]
              link
              fedilink
              English
              11 month ago

              Me? I’m not using LLMs at all hahaha. I’m asking you, who says they have great value, to provide examples of their uses. I just provided pretty much the only one I have heard, which some random dude told me in a different thread. Everyone else, like you, just keeps it abstract and just bullshits and bullshits hahaha.

              • @[email protected]
                link
                fedilink
                English
                01 month ago

                Great use is subjective. But I use them to better understand university lectures. I can have a real discussion, ask questions, ask for examples and so on. I had countless situations where web searches would not have helped me because the ressources cannot do reasoning to explain intuitions. I’m also using it for coding. It’s awesome for boilerplate code. I also sometimes ask it to improve my existing code, so I can learn new coding practicesand tricks from that.

                None of these applications require the LLM to be correct 100% of the time. It’s still great value for me. And when I suspect that it’s wrong about something or it’s hallucinating or bad at explaining something, I’ll just do some web searches for validation.

                You might not find it useful because you’re using it wrong, or simply because you have no application for the value it can provide. But that doesn’t mean it’s all bad. OP certainly doesn’t know how to use it. I would never even think about asking it about historical events.

                • @[email protected]
                  link
                  fedilink
                  English
                  11 month ago

                  So literally you use it for information retrieval hahahaha. I did use copilot, codium, and the jetbrains one for a bit. But I had to disable each one, the amount fo wrong code simply doesn’t justify the little boilerplate it generates.

  • @[email protected]
    link
    fedilink
    English
    47
    edit-2
    1 month ago

    The shooting happened after the end of the training date. Like asking windows 95 clippy about 9/11 and it saying it didn’t happen.

      • @[email protected]
        link
        fedilink
        English
        1
        edit-2
        1 month ago

        Oh I’m far to pragmatic to believe that. If truth isn’t working, then what choice do you really have?

  • @[email protected]
    link
    fedilink
    English
    51 month ago

    Well, if the chatbot learned anything from Dementia Don the racist rapist with 34 felonies that can’t complete a coherent sentence, it learned that you never tell the truth.

  • @[email protected]
    link
    fedilink
    English
    -21 month ago

    Does the AI consistently say that, no matter who asks?

    Because if so, that’s not a hallucination.