Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
AI Privacy Apple

Apple To Analyze User Data on Devices To Bolster AI Technology 34

Apple will begin analyzing data on customers' devices in a bid to improve its AI platform, a move designed to safeguard user information while still helping it catch up with AI rivals. From a report: Today, Apple typically trains AI models using synthetic data -- information that's meant to mimic real-world inputs without any personal details. But that synthetic information isn't always representative of actual customer data, making it harder for its AI systems to work properly.

The new approach will address that problem while ensuring that user data remains on customers' devices and isn't directly used to train AI models. The idea is to help Apple catch up with competitors such as OpenAI and Alphabet, which have fewer privacy restrictions. The technology works like this: It takes the synthetic data that Apple has created and compares it to a recent sample of user emails within the iPhone, iPad and Mac email app. By using actual emails to check the fake inputs, Apple can then determine which items within its synthetic dataset are most in line with real-world messages.

Apple To Analyze User Data on Devices To Bolster AI Technology

Comments Filter:
  • by Sebby ( 238625 ) on Monday April 14, 2025 @09:24PM (#65306379) Journal

    So, with Apple's tricks [macintouch.com], will this be an optional "opt-in", then a "whoops, you got opt-in automatically without notice" with every subsequent updates?

    • Re: (Score:2, Interesting)

      by Anonymous Coward

      Of course. Nobody would opt-in to this bullshit but Apple's feet are on the coals so they, like every business, must squeeze the lifeblood out of every customer.

      • Really, really we need some regulator to get to the point where

        "Big tech wants common carrier liability waivers,right? Then, OK, you have to ensure that email, text messages, photos, Pii on the devices, online services is not used to train any AI models in original form, modified form or anonymized form"

        "And, note, this includes biometrics too, including typing speed, typing cadence, mouse movement paths,speed, click rates, etc."

    • Opt-in, and that preference sticking, is critical. I'm all for Apple taking a different approach from the "if you haven't completely prevented me from getting your data it is mine to use however I want" style generally in use.

      I wonder how much of the phone resources/battery life this on device processing will use.

      • by Sebby ( 238625 )

        I wonder how much of the phone resources/battery life this on device processing will use.

        That'd be my main reason to disable it - their entire software stack is already so bloated, buggy and slow, it makes using even a 2-year old device 'un-delightful' already. This would only make it worse.

  • Yet more reasons to stick with our 13 and 14 Pros

  • For example, a percentage of email communications that I receive includes confidentiality disclaimers in the footers, including legal requirements to not distribute the email to non-recipients, etc. I'm sure other slashdotters are in the same situation.

    An AI training program (which is presumably owned by Apple and licensed for your use on the iPhone) wouldn't be authorized to read these emails and will definitely not be authorized to arbitrarily act up on them.

    In particular, comparing the contents of co

    • In particular, comparing the contents of confidential emails against a synthetically produced external set of contents, so as to favour the more relevant, synthetically produced, samples, is a form of exfiltration of the data.

      Not really, there is no exfiltration. They are not copying anything of yours to their synthetics or anywhere else. What they are doing is ranking their synthetics for a match to your data. Anything matching your data was preexisting data in their synthetics. You can't exfiltrate something they already have.

      What you can do is confirm a guess when actual data matches preexisting synthetic data, as in your hangman letters example. Calling this exfiltration is a little misleading, its confirmation. So the ri

      • Not really, there is no exfiltration. They are not copying anything of yours to their synthetics or anywhere else. What they are doing is ranking their synthetics for a match to your data. Anything matching your data was preexisting data in their synthetics. You can't exfiltrate something they already have.

        The ranking is enough. You use the synthetics as a basis, and look for combinations of synthetics that recreate the unknown data. This falls within the purview of latent semantic analysis, and what is kn

        • Sure, but as I mentioned in my second paragraph, I just think "exfiltration" overstates things a little and that "confirming guesses" is more accurate. It's a subtle difference, but I think the fact that they can only confirm things they have hypothesized is an important limitation. Also "exfiltration" may suggest too much certainty.

          Yes, "guesswork" can be powerful, see WW2 Bletchley Park. They weren't always dealing with specific detailed orders to units. Sometimes it was something vague like a person i
      • Not really, there is no exfiltration. They are not copying anything of yours to their synthetics or anywhere else. What they are doing is ranking their synthetics for a match to your data. Anything matching your data was preexisting data in their synthetics. You can't exfiltrate something they already have.

        This is like arguing oracle attacks can't be used to decrypt ciphertexts. They are not copying... they are deriving... a distinction without a difference.

        What you can do is confirm a guess when actual data matches preexisting synthetic data, as in your hangman letters example. Calling this exfiltration is a little misleading, its confirmation. So the risk here is one of Personally Identifiable Information (PII). Is this match recorded alongside your PII? Or is this match recorded without any PII, such that Apple could not connect it to a user even if they wanted to?

        No this is actually exfiltrating data. No corporation would interpret unauthorized external queries "ranking synthetics" to their data as anything other than an attack because it is.

        • by drnb ( 2434720 )
          And "deriving" and "copying" suggest too much certainty. I think "confirming guesses" more accurately describes things.

          Can confirming guesses be powerful, yes, see my other response and my Bletchley Park example. Also note that without the ability to connect the guesswork to Personally Identifiable information there is little utility beyond training an AI.

          Personally Identifiable Information is the key issue and Apple needs to clarify things.
    • For example, a percentage of email communications that I receive includes confidentiality disclaimers in the footers, including legal requirements to not distribute the email to non-recipients, etc. I'm sure other slashdotters are in the same situation.

      Those are nonsensical, and roughly as binding as the following paragraphs:

      You are not authorized to read the paragraph above this one. If you read it anyway, you owe me $500 and need to turn yourself in to the nearest FBI headquarters to begin serving 6 months jail time in a federal minimum security prison.

      In addition, responding to this post will place you in debt to me for $1947 per word in your reply, plus $193 for every paragraph beyond two.

      • Sure thing. You should try that with a judge sometime, arguing that you wilfully ignored disclaimers even though you were perfectly capable of comprehending their meaning.
        • What I said above is basically legal consensus. Enforcing confidentiality requires a contract, and your putting words at the end of an email does not magically create a contract between you and some random recipient. You will be hard-pressed to find a lawyer that will advise you otherwise.

          As the Apex Law Group states rather well [apexlg.com]:

          Email disclaimers depend on contract law to bind the recipient to the disclaimer and impose a duty, often of non-disclosure. This theory, however, does not create a legally binding

    • My secret word is "bullshit".

      Now, tell me what the secret is.

    • by jeremyp ( 130771 )

      If that's your standard for those emails with confidentiality statements (they are not disclaimers), you already have a problem because your device and maybe email provider already read them to determine if they are spam. Also, unless the email body has been encrypted, they've been sat in SMTP server queues in plain text where nefarious people could read them.

      Those "disclaimers" aren't really worth anything.

  • by Anonymous Coward
    The opt in had better be off by default, and it needs to be sticky and stay off - not magically enabling a few months down the track. I do *not* wish apple to know the embeddings for the shops I buy from, with a little trickery. That's a sure-fire way to enable targeted ads. What else can they glean from private data if they compose a perfect email then ping which phones match? Critical of a politician in an email, apple look for matching emails and ICE appear on your step maybe?
  • Soon, everything you use will be out to harvest your effort. Retrocomputers are looking better all the time.
  • This is the sort of thing that would get written up as a clever(if probably not at the top of the urgency list) inferential attack if you could do it against remote hosts or VMs under the same hypervisor. Apple has had the temerity to declare the comparatively low bitrate of the leakage to be a privacy feature. Love that Cupertino attitude.

    What's great, of course, is that they are doing all this in a black box, whose security vs. people who aren't them they guard fairly jealously; and they are making no
  • Straight "fuck, no!" from me. Time to dump this rotten fruit into a bin.

  • ... not only have I turned off "Apple Intelligence" on any devices, I have also gone through every last detail to tell it not to learn from any apps on it and not to include it in any recommendations. It ultimately comes down to whether or not Apple and its iOS flavors will actually respect the user settings or not. At least by having turned off those switches it would give me a small amount of legal standing against Apple for violating user preferences/settings... not that I would do that kind of fight o
  • Reminds me of NSA hearings years ago where they claimed they didn't "collect" nearly everyone's data they actually collected because according to the NSA the word collecting does not count as collecting until you look at it or use it.

    Nobody on earth is even going to try and parse the distinction without a difference being made here. All they will hear is Apple is rummaging through their shit to train AIs and they won't be wrong in hearing that.

Nothing is faster than the speed of light ... To prove this to yourself, try opening the refrigerator door before the light comes on.

Working...