• reksas@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    9
    ·
    edit-2
    4 months ago

    While this is good thing, not being able to tell what is real and what is not would be disaster. What if every comment here but you were generated by some really advanced ai? What they can do now will be laughable compared to what they can do many years from now. And at that point it will be too late to demand anything to be done about it.

    Ai generated content should have somekind of tag or mark that is inherently tied to it that can be used to identify it as ai generated, even if only part is used. No idea how that would work though if its even possible.

    • conciselyverbose
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      1
      ·
      4 months ago

      You already can’t. You can’t close Pandora’s box.

      Adding labels just creates a false sense of security.

      • reksas@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 months ago

        it wouldnt be label, that wouldnt do anything since it could just be erased. It should be something like invisible set of pixels on pictures or some inaudible soundpattern on sounds that can be detected in some way.

        • conciselyverbose
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          4 months ago

          But it’s irrelevant. You can watermark all you want in the algorithms you control, but it doesn’t change the underlying fact that pictures have been capable of lying for years.

          People just recognizing that a picture is not evidence of anything is better.

          • reksas@sopuli.xyz
            link
            fedilink
            English
            arrow-up
            1
            ·
            4 months ago

            Yes, but reason why people dont already consider pictures irrelevant is that it takes time and effort to manipulate a picture. With ai not only is it fast it can be automated. Of course you shouldnt accept something so unreliable as legal evidence but this will spill over to everything else too

            • conciselyverbose
              link
              fedilink
              English
              arrow-up
              1
              ·
              4 months ago

              It doesn’t matter. Any time there are any stakes at all (and plenty of times there aren’t), there’s someone who will do the work.

              • reksas@sopuli.xyz
                link
                fedilink
                English
                arrow-up
                1
                ·
                4 months ago

                It doesnt matter if you cant trust anything you see? What if you couldn’t be sure if you weren’t talking to bot right now?

                • conciselyverbose
                  link
                  fedilink
                  English
                  arrow-up
                  2
                  ·
                  edit-2
                  4 months ago

                  Photos/video from unknown sources have already been completely worthless as evidence for a solid decade. If you used a random picture online to prove a point 5 years ago, you were wrong. This does not change that reality in any way.

                  The only thing changing is your awareness that they’re not credible.

                  • reksas@sopuli.xyz
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    4 months ago

                    What about reliable sources becoming less reliable? Knowing something is not credible doesn’t help if i can’t know what is credible