Google has plunged the internet into a “spiral of decline”, the co-founder of the company’s artificial intelligence (AI) lab has claimed.

Mustafa Suleyman, the British entrepreneur who co-founded DeepMind, said: “The business model that Google had broke the internet.”

He said search results had become plagued with “clickbait” to keep people “addicted and absorbed on the page as long as possible”.

Information online is “buried at the bottom of a lot of verbiage and guff”, Mr Suleyman argued, so websites can “sell more adverts”, fuelled by Google’s technology.

    • @dependencyInjection
      link
      English
      139 months ago

      It depends what you’re using it for as to whether you need to fact check stuff.

      I’m a software developer and if I can’t remember how to do an inner join in SQL then I can easier ask ChatGPT to do it for me and I will know if it is right or not as this is my field of expertise.

      If I’m asking it how to perform open heart surgery on my cat, then sure I’m probably going to want several second opinions as that is not my area of expertise.

      When using a calculator do you use two different calculators to check that the first one isn’t lying?

      Also, you made a massive assumption that the stuff OP was using it for was something that warranted fact checking.

      I can see why you would use it. Why would I want to search Google for inner joins sql when it is going to give me so many false links that don’t give me the info in need in a concise manner.

      Even time wasting searches have just been ruined. Example: Top Minecraft Java seeds 1.20. Will give me pages littered with ads or the awful page 1-10 that you must click through.

      Many websites are literally unusable at this point and I use ad blockers and things like consent-o-matic. But there are still pop up ads, sub to our newsletter, scam ads etc. so much so that I’ll just leave the site and forego learning the new thing I wanted to learn.

    • @[email protected]
      link
      fedilink
      English
      29 months ago

      The new release of GPT-4 searches Bing, reads the results, summarizes, and provides sources, so it’s easier to fact check than ever if you need to.

    • @[email protected]
      link
      fedilink
      English
      -10
      edit-2
      9 months ago

      It’s pretty trivial to fact check an answer… You should start using this kind of bots more. Check perplexity.ai for a free version.

      Sources are referenced and linked.

      Don’t judge on chatgpt free version

      • GigglyBobble
        link
        fedilink
        29 months ago

        It’s pretty trivial to fact check an answer

        People don’t do it though and often parrot bullshit.

        • @[email protected]
          link
          fedilink
          English
          -6
          edit-2
          9 months ago

          People who do so aren’t smart enough to use internet anyway. With or without AI it wouldn’t change anything for them, they stay stupid and will continue acting stupidly

      • @[email protected]
        link
        fedilink
        English
        -19 months ago

        Perplexity.ai has been my go to for this reason.

        It often brings up bad solutions to a problem and checking the sources it references shows it regulary misses the gist of these sources.

        There sources it selects are often not the ones I end up using. They are starting point, but not the best starting point.

        What it is good for is for finding content when I don’t know the terminology of the domain. It is a starting point ready to lead me astray with exquisitely written content.

        Find trustworthy sources and use them.

          • @[email protected]
            link
            fedilink
            English
            1
            edit-2
            9 months ago

            That’s what’s usually gets said about lots of alternative fusion energy generation methods that later turn out to be impossible to have net-positive energy generation.

            And this is just one example. Another example: tons of “neat concept that shows potential” medical compounds end up dropped at the medical testing stage because of their nasty side effects or it turns out their “positive” effects are indistinguisheable from the placebo effect.

            The point being that you can’t actually extrapolative from “neat concept that shows potential” even to merelly “will work”, much less to “will be a great success”.

            PS: Equally, one can’t just say it’s not going to be a great success - being a “neat concept that shows potential” has a pretty low informational content when it comes to predicting the future, worse so when there are people monetarilly heavilly invested into it who have a strong interest in making it look like a “neat concept that shows potential” whilst hiding any early stage problem as they’re activelly poluting the information space around it.

            • @[email protected]
              link
              fedilink
              English
              2
              edit-2
              9 months ago

              You are mixing sci-fi level of cutting edge basic research (fusion), with commercial products (chatgpt). They are 2 very different type of proof of concepts.

              And both will likely revolutionize human society. Fusion will simply commercially become a thing in 30/50 years. AI has been on the market for years now. Generative models are also few years old. They are simply becoming better and now new products can be built on top of them

              (btw I already use chatgpt 4 productively every day for my daily work and it helps me a lot)

              • @[email protected]
                link
                fedilink
                English
                1
                edit-2
                9 months ago

                I seem to not have explained myself correctly.

                This specific tech you seem to be emotionally invested in is no different from the rest in this sense because it still faces in the real world the very same kind of risks and pitfalls as the rest - there are possible internal pitfalls inherent to every new technology (i.e. a problem we never knew about because we never used it with so many people in the real world before, becomes visible with widespread use) and there are possible external pitfalls inherent to how it fits in the complex world we live in (i.e. it turns out the use cases don’t make quite as much economic sense as was first tought or it indirectly generates more problems than it solves).

                Such Process and Fit risks are true for every early stage “revolutionary” tech (i.e. we never did it before, now that we do it, we discover problems we were not at all aware of before) - business guys might say that revolutionary tech starts with a lot more “unknown unknowns” than incremental improvements on existing tech - and is why the bean counters rarelly put money in revolutionary and instead go mainly for incremental improvements on proven tech. At times one or more of such “we had no idea this could happen problems” turn out to be insurmountable, sometimes they can be overcomed but the result is not especially great, sometimes they’re all overcomed without any nasty side-effects and the thing ends up being a world-changing tech: you can’t really tell upfront.

                In the case of LLMs, the two risky problems from what I’ve heard which might stop it from being “world changing” are in how LLMs being trained in material which includes LLM-generated material actually get worse (so as the Internet gets flooded with LLM-generated material passing for human-generated one, LLMs would get worse and worse) and the other is the so-called Hallucinations, which are really just the natural side effect of them being Language Models hence all that they do is generate compositions of language tokens that pass for human generated language, with no reasoning involved hence cannot validate through inductive or deductive reasoning said “compositions of language tokens”, so LLMs wouldonly usefull for altering format without touching the information (for example, turn lists of cold hard facts into fluffy longwinded text or do the opposite and summarize lots of fluffy text into just the facts) which would still have a big impact in certain professions but not necessarilly be “world changing” (or, even more interestingly, make over time people value “fluffy text” less and less, which would be “world changing” but not in a way that benefits the makers of LLM).

                Unless you want to deny the last 4 decades of History in Tech, you can’t logically extrapolate from an early “looks like it might be a success” to “it will be a success”, especially in the era of money-driven overhype we live in.