“404 Media previously reported Cox Media Group (CMG) was advertising a service that claimed to target ads based on what potential customers said near device microphones. Now, here is the pitch deck CMG sent to prospective companies. Google has kicked CMG off its Partner Program in response…”

  • Pasta Dental
    link
    fedilink
    arrow-up
    6
    ·
    edit-2
    2 months ago

    I simply think that until now (maybe they will start tomorrow), the PR and lawsuit risk of listening to people is too high, for the benefit they would get out of it. Much simpler metrics are enough for them to get a very good profile of the user. Voice data isn’t like in the test scenarios where the person will repeat 45x the word cat food, people talk about the weather and about gas prices which is pretty useless for creating an ad profile if you ask me. But the scary part is now with AI models and on device AI everything, local processing of the mic data into topics that then get sent to their servers is more concerning is not much more feasible.

    And for the lawsuits I am not sure they could write it off as a bug everywhere other than the us and Canada because there are actually normal laws in most other countries

    • ganymede@lemmy.ml
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      2 months ago

      the PR and lawsuit risk

      what risk? facebook & others conducted illegal human experiments. this is an enormous crime and was widely reported yet all fb had to do from a pr perspective was apologise.

      as we all know, fb even interfered with with the electoral process of arguably the world’s most powerful nation, and all they had to do was some rebranding to meta and it’s business as usual. this is exactly how powerful these organisations are. go up against a global superpower & all you need to do is change your business name??? they don’t face justice the same way anyone else would, therefore we cannot assess the risk for them as we would another entity - and they know it.

      So, while i personally disagree for above reasons, I can accept in your opinion they wouldn’t take the legal risk.

      simpler metrics are enough

      when has ‘enough’ ever satisfied these entities? we merely need to observe the rate of evolution of various surveillance methods, online, in our devices, in shopping centers to see ‘enough’ is never enough. its always increasing, and at an alarming rate.

      local processing of the mic data into topics that then get sent to their servers is more concerning is not much more feasible

      sorry i didn’t quite understand, are you saying its not feasible or it is feasible? from the way the sentence started i thought you were going to say it could be, but then you said ‘not much more feasible’?

      Voice data isn’t

      voice conversations are near-universally prized in surveillance & intelligence. There hasn’t been any convincing argument for any generalised exception to that.

      I am not sure they could write it off as a bug

      it’s already been written off as a bug. i didn’t follow that story indefinitely but i’m not aware of even a modest fine being paid in relation to the above story. if it can accidentally transcribe and send your conversations to your contact list without your knowledge or consent (literally already happened - with impunity(?)), they can 1000% “accidentally” send it to some ‘debug’ server somewhere.

      Are they actually doing it? It ofc remains to be seen. Imo the fallout if it was revealed would roughly look like this

      • A few people would say “no shit”
      • Most people would parrot the “ive done nothing wrong so i don’t care” line.
      • A few powerless people would be upset.