• AwesomeLowlander
    link
    fedilink
    English
    arrow-up
    42
    arrow-down
    1
    ·
    7 hours ago

    The measure, aimed at reducing potential risks created by AI, would have required companies to test their models and publicly disclose their safety protocols to prevent the models from being manipulated to, for example, wipe out the state’s electric grid or help build chemical weapons.

    How exactly do LLMs do that? If you’ve given an LLM’s pseudorandom output control over your electrical grid, no regulation will mitigate your stupidity.

    • schizo@forum.uncomfortable.business
      link
      fedilink
      English
      arrow-up
      38
      arrow-down
      3
      ·
      6 hours ago

      How exactly do LLMs do that?

      Too many people are confused and think a LLM is an actual AI, and not just a tarted up ELIZA bot from 1968.

    • bamfic@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      6 hours ago

      Could he understand the halting problem? I doubt he does, but the legislators evidently don’t either

    • oce 🐆@jlai.lu
      link
      fedilink
      English
      arrow-up
      3
      ·
      6 hours ago

      I think it’s more about asking it the steps to create a bomb or how to disrupt the grid, for example, where to cut the major edges.

      • AwesomeLowlander
        link
        fedilink
        English
        arrow-up
        11
        ·
        6 hours ago

        asking it the steps to create a bomb

        That sounds like a self-correcting issue right there

        • dual_sport_dork 🐧🗡️@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          ·
          4 hours ago

          That, and the Internet has been teaching people how to create bombs since the dial-up days. I don’t predict that LLM’s will be either a benefit or a detriment to that particular strain of natural selection.

          • AwesomeLowlander
            link
            fedilink
            English
            arrow-up
            2
            ·
            4 hours ago

            Is it more of a public safety issue than if they actually build a working one from a legit bomb manual and deploy it?

    • UnderpantsWeevil@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      3
      ·
      5 hours ago

      How exactly do LLMs do that?

      If you hook an LLM up as an interface replacement for a manual/analog Power Plant interface and start asking the translator to intuit decisions based on fuzzy inputs, you can create a cascade of errors that result in grid failure.

      If you’ve given an LLM’s pseudorandom output control over your electrical grid, no regulation will mitigate your stupidity.

      This rule would prevent a business or public regulator from doing such a thing without proving out safeguards.

      And the governor vetoed it.