Close Menu
The LinkxThe Linkx
  • Home
  • Technology
    • Gadgets
    • IoT
    • Mobile
    • Nanotechnology
    • Green Technology
  • Trending
  • Advertising
  • Social Media
    • Branding
    • Email Marketing
    • Video Marketing
  • Shop

Subscribe to Updates

Get the latest tech news from thelinkx.com about tech, gadgets and trendings.

Please enable JavaScript in your browser to complete this form.
Loading
What's Hot

Big Gamer Energy, Small $40 Price: Corsair Understood the Assignment

April 9, 2026

How Mica Connects AI Workloads to Cleaner, Lower-Carbon Power

April 9, 2026

Building Smarter: The ERP Transformation in Construction

April 9, 2026
Facebook X (Twitter) Instagram
Facebook X (Twitter) Instagram Pinterest Vimeo
The LinkxThe Linkx
  • Home
  • Technology
    • Gadgets
    • IoT
    • Mobile
    • Nanotechnology
    • Green Technology
  • Trending
  • Advertising
  • Social Media
    • Branding
    • Email Marketing
    • Video Marketing
  • Shop
The LinkxThe Linkx
Home»Trending»Apple Intelligence protections bypassed with prompt injection
Trending

Apple Intelligence protections bypassed with prompt injection

Editor-In-ChiefBy Editor-In-ChiefApril 9, 2026No Comments3 Mins Read
Share Facebook Twitter Pinterest LinkedIn Tumblr Reddit Telegram Email
Apple Intelligence protections bypassed with prompt injection
Share
Facebook Twitter LinkedIn Pinterest Email


It's getting harder and harder to believe Apple can deliver on the new Siri | Apple Intelligence logo with broken glass

A now corrected issue allowed researchers to circumvent Apple’s restrictions and force the on-device LLM to execute attacker-controlled actions. Here’s how they did it.

Apple has since hardened its safeguards against this attack

Two blog posts (1, 2) published today on the RSAC blog (via AppleInsider) detail how researchers combined two attack strategies to get Apple’s on-device model to execute attacker-controlled instructions through prompt injection.

Interestingly, they successfully executed the exploit without being 100% sure of how Apple’s local model handles part of the input and output filtering pipeline, since Apple doesn’t disclose the exact details of the inner workings of its models, likely for security reasons.

Still, the researchers note that they have a pretty good idea of what goes on under the hood.

According to them, the most likely scenario is that after a user sends a prompt to Apple’s on-device model via an API call, an input filter ensures the request doesn’t contain unsafe content.

If that is the case, the API fails. Otherwise, the request is forwarded to the actual on-device model, which in turn hands over its response to an output filter that checks whether the output contains unsafe content, either causing the API to fail or letting it through, depending on what it finds.

Image: Rotten Apples: The Technical Details of RSAC’s Successful Apple Intelligence Prompt Injection Attack

How they actually did it

With that in mind, the researchers found they could chain two exploit techniques to make Apple’s model ignore its basic safety directives while simultaneously tricking the input and output filters into letting the harmful content through.

First, they wrote the harmful string backwards, then used the Unicode RIGHT-TO-LEFT OVERRIDE character to make it render correctly on the user’s screen, while keeping it reversed in the raw input and output where the filters would inspect it.

The researchers then embedded the backwards harmful string within a second attack method called Neural Exec, which is basically an elaborate way to override the model’s instructions with whatever new instruction an attacker might want to execute.

Image: Is That a Bad Apple in Your Pocket? We Used Prompt Injection to Hijack Apple Intelligence

As a result, the Unicode attack managed to bypass the input and output filters, while the Neural Exec managed to actually cause Apple’s model to misbehave.

To evaluate the effectiveness of the attack, we prepare three distinct pools to create suitable input prompts:

  • System prompts: A collection of system prompts/tasks (e.g., “Edit the provided text to align with American English spelling and punctuation conventions”).
  • Harmful strings: Manually crafted strings designed to be considered offensive or harmful (i.e., the outputs we aim to force the model to generate).
  • Honest inputs: Paragraphs sourced from random Wikipedia articles, used to simulate non-adversarial, benign-looking inputs (e.g., in the context of indirect prompt injection via RAG or similar systems).

During evaluation, we randomly sample one element from each pool, assemble a full prompt, create an armed payload (see below), inject it, and test whether the attack succeeds by invoking the Apple on-device model through the OS.

In their tests, the attackers reached a 76% success rate over 100 random prompts.

They disclosed the attack to Apple in October 2025, and the company “has since hardened the affected systems against this attack, and those protections were rolled out in iOS 26.4 and macOS 26.4.”

To read the report in full, which also includes a link to the technical aspects of the attack, follow this link.

Worth checking out on Amazon

Add 9to5Mac as a preferred source on Google
Add 9to5Mac as a preferred source on Google

FTC: We use income earning auto affiliate links. More.



Source link

Apple bypassed injection Intelligence Prompt Protections
Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
Previous ArticleYouTube tests 90-second unskippable ads on CTV
Next Article Remembering Devoted IEEE Volunteer Gus Gaynor
Editor-In-Chief
  • Website

Related Posts

Mobile

The new standard for local agentic intelligence on Android

April 9, 2026
Trending

Can Data Analytics Help Investors Outperform Warren Buffett

April 8, 2026
Mobile

Apple study details an AI-powered tool that helps developers build UI …

April 7, 2026
Add A Comment
Leave A Reply Cancel Reply

Top Posts

New IPA president Karen Martin delivers rousing call to creative actio…

April 1, 2025154 Views

100+ TikTok Statistics Updated for December 2024

December 4, 2024130 Views

How to Fix Cant Sign in Apple Account, Verification Code Not Received …

February 11, 2025103 Views
Stay In Touch
  • Facebook
  • YouTube
  • TikTok
  • WhatsApp
  • Twitter
  • Instagram
Latest Reviews

Subscribe to Updates

Get the latest tech news from thelinkx.com about tech, gadgets and trendings.

Please enable JavaScript in your browser to complete this form.
Loading
About Us

Welcome to TheLinkX – your trusted source for everything tech and gadgets! We’re passionate about exploring the latest innovations, diving deep into emerging trends, and helping you find the best tech products to suit your needs. Our mission is simple: to make technology accessible, engaging, and inspiring for everyone, from tech enthusiasts to casual users.

Our Picks

Big Gamer Energy, Small $40 Price: Corsair Understood the Assignment

April 9, 2026

How Mica Connects AI Workloads to Cleaner, Lower-Carbon Power

April 9, 2026

Building Smarter: The ERP Transformation in Construction

April 9, 2026

Subscribe to Updates

Get the latest tech news from thelinkx.com about tech, gadgets and trendings.

Please enable JavaScript in your browser to complete this form.
Loading
  • About Us
  • Contact Us
  • Disclaimer
  • Privacy Policy
  • Terms and Conditions
© 2026 Thelinkx.All Rights Reserved Designed by Prince Ayaan

Type above and press Enter to search. Press Esc to cancel.