• @[email protected]
    link
    fedilink
    English
    351 month ago

    It will design the machines to build the autonomous robots that mine the cobalt… doing the jobs of several companies at one time and either freeing up several people to pursue leisure or the arts or starve to death from being abandoned by society.

    • @[email protected]
      link
      fedilink
      English
      701 month ago

      Have you seen the real fucking world?

      It’s gonna make the rich richer and the poor poorer. At least until the gilded age passes.

    • @funkless_eck
      link
      English
      111 month ago

      AI absolutely will not design machines.

      It may be used within strict parameters to improve the speed of theoretically testing types of bearing or hinge or alloys or something to predict which ones would perform best under stress testing - prior to acutal testing to eliminate low-hanging fruit, but it will absolutely not generate a new idea for a machine because it can’t generate new ideas.

      • @[email protected]
        link
        fedilink
        English
        31 month ago

        The model T will absolutely not replace horse drawn carts – Maybe some small group of people or a family for a vacation but we’ve been using carts to do war logistics for 1000s of years. You think some shaped metal put together is going to replace 1000s of men and horses? lol yeah right

        • @funkless_eck
          link
          English
          21 month ago

          apples and oranges.

          You’re comparing two products with the same value prop: transporting people and goods more effectively than carrying/walking.

          In terms of mining, a drilling machine is more effective than a pickaxe. But we’re comparing current drilling machines to potential drilling machines, so the actual comparison would be:

          • is an AI-designed drilling machine likely to be more productive (for any given definition of productivity) than a human-designed one?

          Well, we know from experience that when (loosely defined) “AI” is used in, for e.g. pharma research, it reaps some benefits - but does not replace wholesale the drug approval process and its still a tool used by - as I originally said - human beings that impose strict parameters on both input and output as part of a larger product and method.

          Back to your example: could a series of algorithmic steps - without any human intervention - provide a better car than any modern car designers? As it stands, no, nor is it on the horizon. Can it be used to spin through 4 million slight variations in hood ornaments and return the top 250 in terms of wind resistance? Maybe, and only if a human operator sets up the experiment correctly.

          • @[email protected]
            link
            fedilink
            English
            7
            edit-2
            1 month ago

            No, the thing I’m comparing is our inability to discern where a new technology will lead and our history of smirking at things like books, cars, the internet and email, AI, etc.

            The first steam engines pulling coal out of the ground were so inefficient they wouldn’t make sense for any use case than working to get the fuel that powers them. You could definitely smirk and laugh about engines vs 10k men and be totally right in that moment, and people were.

            The more history you learn though, you more you realize this is not only a hubrisy thing, it’s also futile as how we feel about the proliferation of technology has never had an impact on that technology’s proliferation.

            And, to be clear, I’m not saying no humans will work or have anything to do – I’m saying significantly MORE humans will have nothing to do. Sure you still need all kinds of people even if the robots design and build themselves mostly, but it would be an order of magnitude less than the people needed otherwise.

            • @[email protected]
              link
              fedilink
              English
              11 month ago

              Maybe I’m pessimistic but all I see is every call center representative disappearing and that’ll be it

          • @[email protected]
            link
            fedilink
            English
            51 month ago

            I agree that AI is just a tool, and it excels in areas where an algorithmic approach can yield good results. A human still has to give it the goal and the parameters.

            What’s fascinating about AI, though, is how far we can push the algorithmic approach in the real world. Fighter pilots will say that a machine can never replace a highly-trained human pilot, and it is true that humans do some things better right now. However, AI opens up new tactics. For example, it is virtually certain that AI-controlled drone swarms will become a favored tactic in many circumstances where we currently use human pilots. We still need a human in the loop to set the goal and the parameters. However, even much of that may become automated and abstracted as humans come to rely on AI for target search and acquisition. The pace of battle will also accelerate and the electronic warfare environment will become more saturated, meaning that we will probably also have to turn over a significant amount of decision-making to semi-autonomous AI that humans do not directly control at all times.

            In other words, I think that the line between dumb tool and autonomous machine is very blurry, but the trend is toward more autonomous AI combined with robotics. In the car design example you give, I think that eventually AI will be able to design a better car on its own using an algorithmic approach. Once it can test 4 million hood ornament variations, it can also model body aerodynamics, fuel efficiency, and any other trait that we tell it is desirable. A sufficiently powerful AI will be able to take those initial parameters and automate the process of optimizing them until it eventually spits out an objectively better design. Yes, a human is in the loop initially to design the experiment and provide parameters, but AI uses the output of each experiment to train itself and automate the design of the next experiment, and the next, ad infinitum. Right now we are in the very early stages of AI, and each AI experiment is discrete. We still have to check its output to make sure it is sensible and combine it with other output or tools to yield useable results. We are the mind guiding our discrete AI tools. But over a few more decades, a slow transition to more autonomy is inevitable.

            A few decades ago, if you had asked which tasks an AI would NOT be able to perform well in the future, the answers almost certainly would have been human creative endeavors like writing, painting, and music. And yet, those are the very areas where AI is making incredible progress. Already, AI can draw better, write better, and compose better music than the vast, vast majority of people, and we are just at the beginning of this revolution.

        • @funkless_eck
          link
          English
          41 month ago

          can

          might

          sure. But, like I said, those are subject to a lot of caveats - that humans have to set the experiments up to ask the right questions to get those answers.

          • @[email protected]
            link
            fedilink
            English
            11 month ago

            That’s how it currently is, but I’d be astounded if it didn’t progress quickly from now.

            • @[email protected]
              link
              fedilink
              English
              21 month ago

              OpenAI themselves have made it very clear that scaling up their models have diminishing returns and that they’re incapable of moving forward without entirely new models being invented by humans. A short while ago they proclaimed that they could possibly make an AGI if they got several Trillions of USD in investment.

              • @[email protected]
                link
                fedilink
                English
                11 month ago

                5 years ago I don’t think most people thought ChatGPT was possible, or StableDiffusion/MidJourney/etc.

                We’re in an era of insane technological advancement, and I don’t think it’ll slow down.

                • @[email protected]
                  link
                  fedilink
                  English
                  3
                  edit-2
                  1 month ago

                  Okay but the people who made the advancements are telling you it has already slowed down. Why don’t you understand that? A flawed Chatbot and some art theft machines who can’t draw hands aren’t exactly worldchanging, either, tbh.

                  • @[email protected]
                    link
                    fedilink
                    English
                    01 month ago

                    There are other people in the world. Some of them are inventing completely new ways of doing things, and one of those ways could lead to a major breakthrough. I’m not saying a GPT LLM is going to solve the problem, I’m saying AI will.

                  • @[email protected]
                    link
                    fedilink
                    English
                    -11 month ago

                    This is such a rich-country-centric view that I can’t stand. LLMs have already given the world maybe it’s greatest gift ever – access to a teacher.

                    Think of the 800 million poor children in the world and their access to a Kahn academy level teacher on any subject imaginable with a cellphone/computer as all they need. How could that not have value and is pearl clutching drawing skills becoming devalued really all you can think about it?

            • @funkless_eck
              link
              English
              1
              edit-2
              1 month ago

              i would be extremely surprised if before 2100 we see AI that has no human operator and no data scientist team even at a 3rd party distributor - and those things are neither a lie, nor a weaselly marketing stunt (“technically the operators are contractors and not employed by the company” etc).

              We invented the printing press 584 years ago, it still requires a team of human operators.

              • @[email protected]
                link
                fedilink
                English
                11 month ago

                A printing press is not a technology with intelligence. It’s like saying we still have to manually operate knives… of course we do.

                • @funkless_eck
                  link
                  English
                  0
                  edit-2
                  1 month ago

                  the comment I originally replied to claimed AI will design the autonomous machines.

                  It will not. It will facilitate some of the research done by humans to aid in the designing of willfully human operated machinery.

                  To my knowledge the only autonomous machine that exists is a roomba, which moves blindly around until it physically strikes an object, rotates a random degree and continues in a new direction until it hits something else.

                  Even then, it is controlled with an app and on more expensive models, some boundary setting.

                  It is extremely generous to call that “autonomy.”

                  • @[email protected]
                    link
                    fedilink
                    English
                    11 month ago

                    I was in a self-driving taxi yesterday. It didn’t need to bump into things to figure out where it was.

    • ObliviousEnlightenment
      link
      fedilink
      English
      51 month ago

      either freeing up several people to pursue leisure or the arts or starve to death from being abandoned by society.

      You know EXACTLY which one it’s gonna be.

    • @[email protected]
      link
      fedilink
      English
      21 month ago

      It isn’t the intelligence of the machine designer that is the issue, it is the middlemen and the end user.

      Continuously having to downgrade machines. Wouldn’t want some sales rep seeing something new.

    • @Grandwolf319
      link
      English
      11 month ago

      Hahaha, current ML is basically good guessing, that doesn’t really transfer to building machines that actually have to obey the laws of physics.

      • @[email protected]
        link
        fedilink
        English
        01 month ago

        is it good guessing that you know when you step out of your bed without looking you won’t fall to your death?

        • @[email protected]
          link
          fedilink
          English
          141 month ago

          Work a blue collar job your whole life and tell me it’s possible. Machines suck ass. They either need constant supervision, repairs all the time, or straight up don’t function properly. Tech bros always forget about the people who actually keep the world chugging.

          • @[email protected]
            link
            fedilink
            English
            11 month ago

            They suck because your employer wouldn’t pay me more for a better machine. Chemical is where it is at, outside of powerplants and some of the bigger pharms the chemical operator is a dead profession. Entire plants are automated with the only people doing work are doing repairs or sales.

        • @[email protected]
          link
          fedilink
          English
          141 month ago

          LLMs aren’t going to be designing anything; they’re just fancy auto complete engines with a tendency to hallucinate facts they haven’t been trained on.

          LLMs are preventing real advancements in AI by focusing the attention and funding into what’s evidently a dead end.

              • nickwitha_k (he/him)
                link
                fedilink
                English
                21 month ago

                I hope not. I want more types of sentient beings to exist. But, I also don’t believe any company is actually working towards AGI.

              • @[email protected]
                link
                fedilink
                English
                11 month ago

                No, the existence of humans inherently disproves that. We just have hardware so advanced many still think it’s magic.

                Now, if you said it was a pipe dream within the next decade? I’d agree.

                • @[email protected]
                  link
                  fedilink
                  English
                  -11 month ago

                  All the money’s going into the LLM bubble, so there won’t be any left for actual AI research until it bursts.

                  • @[email protected]
                    link
                    fedilink
                    English
                    11 month ago

                    Saying something like that doesn’t make it true. That’s not proof.

                    Are you claiming that absolutely nobody is working on AGI because LLMs exist and are hot right now?

                • @[email protected]
                  link
                  fedilink
                  English
                  1
                  edit-2
                  1 month ago

                  LLMs are incapable of “recognising” any patterns they haven’t been trained on.

                  And they don’t really even recognise those, they’re just fancy auto complete engines, simply outputting the highest scored token from their training base based on their input.

                  They’re pattern matching machines; there’s no recognition, inner modelling of new knowledge, self referencing, or understanding of any kind, merely blind statistics.

                  They’re just bigger and fancier Eliza’s, and just as distant as Eliza was from any practical form of intelligence, artificial or natural.

                  While I personally do believe that achieving AGI¹, on a Turing machine is possible, LLMs and how they work are an excellent example in support of John Searle’s arguments against it with his Chinese room though experiment.

                  1— Or at least something equivalent to human intelligence, or better, in the measures by which we consider ourselves to be intelligent, though it’s arguable whether we can really be considered intelligent at all, or we’re just better, more complex, Chinese rooms.

                  • @[email protected]
                    link
                    fedilink
                    English
                    3
                    edit-2
                    1 month ago

                    But since we don’t understand how cognition works in living beings almost at all – who’s to say that’s not how ‘actual thinking’ works other than 'I know it when I see it!"