• Possibly linux@lemmy.zip
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    7
    ·
    5 months ago

    This isn’t entirely true. AI is usually trained on public data such as Wikipedia.

    AI is a tool. How you use it is what matters.

    • 31337
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      5 months ago

      It’s also trained on data people reasonably expected would be private (private github repos, Adobe creative cloud, etc). Even if it was just public data, it can still be dangerous. I.e. It could be possible to give an LLM a prompt like, “give me a list of climate activists, their addresses, and their employers” if it was trained on this data or was good at “browsing” on its own. That’s currently not possible due to the guardrails on most models, and I’m guessing they try to avoid training on personal data that’s public, but a government agency could make an LLM without these guardrails. That data could be public, but would take a person quite a bit of work to track down compared to the ease and efficiency of just asking an LLM.

    • Kilgore Trout@feddit.it
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      5 months ago

      Wikipedia requires attribution, which AI scrapers never give.

      It is “public” work, but under a license.