Per one tech forum this week: “Google has quietly installed an app on all Android devices called ‘Android System SafetyCore’. It claims to be a ‘security’ application, but whilst running in the background, it collects call logs, contacts, location, your microphone, and much more making this application ‘spyware’ and a HUGE privacy concern. It is strongly advised to uninstall this program if you can. To do this, navigate to 'Settings’ > 'Apps’, then delete the application.”

    • @kattfisk@lemmy.dbzer0.com
      link
      fedilink
      English
      344 months ago

      To quote the most salient post

      The app doesn’t provide client-side scanning used to report things to Google or anyone else. It provides on-device machine learning models usable by applications to classify content as being spam, scams, malware, etc. This allows apps to check content locally without sharing it with a service and mark it with warnings for users.

      Which is a sorely needed feature to tackle problems like SMS scams

      • @throwback3090@lemmy.nz
        link
        fedilink
        English
        114 months ago

        Why do you need machine learning for detecting scams?

        Is someone in 2025 trying to help you out of the goodness of their heart? No. Move on.

        • @Aermis@lemmy.world
          link
          fedilink
          English
          64 months ago

          If you want to talk money then it is in businesses best interest that money from their users is being used on their products, not being scammed through the use of their products.

          Secondly machine learning or algorithms can detect patterns in ways a human can’t. In some circles I’ve read that the programmers themselves can’t decipher in the code how the end result is spat out, just that the inputs will guide it. Besides the fact that scammers can circumvent any carefully laid down antispam, antiscam, anti-virus through traditional software, a learning algorithm will be magnitudes harder to bypass. Or easier. Depends on the algorithm

          • @throwback3090@lemmy.nz
            link
            fedilink
            English
            44 months ago

            I don’t know the point of the first paragraph…scams are bad? Yes? Does anyone not agree? (I guess scammers)

            For the second we are talking in the wild abstract, so I feel comfortable pointing out that every automated system humanity has come up with so far has pulled in our own biases and since ai models are trained by us, this should be no different. Second, if the models are fallible, you cannot talk about success without talking false positives. I don’t care if it blocks every scammer out there if it also blocks a message from my doctor. Until we have data on consensus between these new algorithms and desired outcomes, it’s pointless to claim they are better at X.

        • @kattfisk@lemmy.dbzer0.com
          link
          fedilink
          English
          44 months ago

          Blaming the victim solves nothing.

          Scamming is a rapidly growing industry that is becoming more professional and specialized all the time. Anyone can be scammed.

      • @cley_faye@lemmy.world
        cake
        link
        fedilink
        English
        24 months ago

        You don’t need advanced scanning technology running on every device with access to every single bit of data you ever seen to detect scam. You need telco operator to stop forwarding forged messages headers and… that’s it. Cheap, efficient, zero risk related to invasion of privacy through a piece of software you did not need but was put there “for your own good”.

        • @zlatko@programming.dev
          link
          fedilink
          English
          44 months ago

          I will perhaps be nitpicking, but… not exactly, not always. People get their shit hacked all the time due to poor practices. And then those hacked things can send emails and texts and other spam all they want, and it’ll not be forged headers, so you still need spam filtering.

    • Spaniard
      link
      fedilink
      English
      84 months ago

      If the app did what op is claiming then the EU would have a field day fining google.

    • @throwback3090@lemmy.nz
      link
      fedilink
      English
      4
      edit-2
      4 months ago

      graphene folks have a real love for the word misinformation (and FUD, and brigading). That’s not you under there👻, Daniel, is it?

      After 5 years of his antics hateful bullshit lies, I think I can genuinely say that word triggers me.

    • @dan@upvote.au
      link
      fedilink
      English
      44 months ago

      So is this really just a local AI model? Or is it something bigger? My S25 Ultra has the app but it hasn’t used any battery or data.

      • @teohhanhui@lemmy.world
        link
        fedilink
        English
        114 months ago

        Please, read the links. They are the security and privacy experts when it comes to Android. That’s their explanation of what this Android System SafetyCore actually is.

      • @loics2@lemm.ee
        link
        fedilink
        English
        54 months ago

        Have you even read the article you posted? It mentions these posts by GrapheneOS