FAQ

Apple's Gemini-Powered Siri Arrives in iOS 26.4 — But the Full Upgrade Is Still Months Away

iOS 26.4 is rolling out with the first phase of Apple's $1 billion-per-year Google Gemini integration, bringing on-screen awareness, contextual understanding, and email summarization to Siri. But the transformational conversational AI Apple promised when it announced the partnership in January has been split across multiple OS updates, with full Phase 2 capabilities pushed to iOS 26.5 in May and iOS 27 in September — frustrating early adopters and raising questions about Apple's AI execution.

5 min read

For five years, Apple’s most consistent AI story was an embarrassing one: Siri, the world’s most distributed voice assistant by installed base, was consistently outclassed by competitors with a fraction of Apple’s resources. Last January’s announcement that Apple had partnered with Google to replace the AI engine behind Siri with a custom version of Gemini — at a reported cost of $1 billion per year — was the company’s clearest acknowledgment yet that building a competitive large language model in-house had been a strategic failure.

iOS 26.4, now rolling out to devices worldwide, delivers the first tangible results of that partnership. But if you were expecting Siri to feel fundamentally different overnight, the reality is more complicated — and in some ways, more honest about the difficulty of rebuilding a decade-old voice assistant from the inside.

What’s Actually in iOS 26.4

The Phase 1 Siri upgrade in iOS 26.4 centers on four capabilities that mark a genuine departure from the previous generation of Siri’s functionality:

On-screen awareness allows Siri to see and understand whatever is currently displayed on an iPhone or iPad screen, and to reason about it contextually. Ask Siri to “add this event to my calendar” while looking at a restaurant website and it will parse the address, phone number, and hours and pre-populate the calendar entry without requiring any additional input. This was technically possible in limited scenarios with previous Siri versions, but the Gemini integration makes it fluid and reliable across a much broader range of UI contexts.

Contextual understanding extends Siri’s memory across a conversation session. The new Siri can maintain the thread of a multi-turn exchange, refer back to earlier statements without the user restating them, and resolve pronouns and references that the old Siri would have treated as isolated commands. “Find that restaurant we talked about” now actually works.

Email summarization and drafting puts Gemini’s language capabilities directly into Apple Mail, offering concise summaries of long email threads and intelligent reply drafting that understands the sender, the context of the conversation, and the user’s past communication patterns with that contact.

Basic cross-app actions let Siri execute simple task chains that cross application boundaries — booking a ride to a location mentioned in a message, creating a note from a recipe on a website, or adding a product seen in Photos to a shopping list. These are the building blocks of the agentic Siri experience that Apple has been describing since the original Apple Intelligence announcement in 2024.

The Gap Between Promise and Delivery

The frustration is not that iOS 26.4 is bad — the Phase 1 features represent a real and meaningful improvement over the Siri that preceded them. The frustration is the distance between what Apple announced in January and what is actually shipping now.

The January 12 announcement, framed as a joint statement from Apple and Google, described a “completely reimagined Siri” powered by Gemini that would deliver a “more personalized” experience, with Gemini’s conversational capabilities running through Apple’s Private Cloud Compute infrastructure to maintain “Apple’s industry-leading privacy standards.” The custom Gemini model at the center of the deal is reportedly eight times larger than Apple’s existing cloud AI models — a 1.2 trillion parameter system purpose-built for Apple’s specific requirements.

What iOS 26.4 delivers is Phase 1 of a planned two-phase rollout. According to multiple reports, the more profound conversational AI — the kind that can genuinely reason across all of a user’s data, understand long-horizon requests, and respond with the fluency that Gemini demonstrates in Google’s own products — has been pushed to iOS 26.5 (expected in May) for partial delivery and iOS 27 (September) for the full experience. The reason most frequently cited is integration complexity: routing Gemini’s inference through Apple’s Private Cloud Compute architecture, while maintaining the privacy guarantees Apple has committed to, required more engineering work than the original timeline anticipated.

The Privacy Architecture Challenge

The privacy engineering problem deserves more than a dismissive note. Apple’s Private Cloud Compute is genuinely novel infrastructure — a system that allows on-device and cloud AI inference to be performed in a way that Apple itself cannot access the user’s data, verified through cryptographic attestation that security researchers have audited publicly. Integrating a model as large as the custom 1.2 trillion parameter Gemini into this framework is not a trivial API integration.

The engineering challenge is that Gemini’s conversational capabilities — particularly its memory and long-context reasoning — depend on infrastructure that assumes the model can maintain state and access broad context. Apple’s privacy architecture is designed precisely to prevent that kind of persistent access to user data. Reconciling these two design philosophies, without compromising either Gemini’s capabilities or Apple’s privacy guarantees, is the problem that pushed Phase 2 to a later release.

When it ships, Phase 2 is intended to enable a Siri that can synthesize across a user’s entire data footprint — messages, emails, photos, health data, calendar, and third-party apps — with the conversational fluency of Gemini, while running through the same Private Cloud Compute attestation model. The privacy story, if Apple executes it, would be unprecedented: a trillion-parameter conversational AI that operates at frontier capability levels with zero-knowledge guarantees for the AI provider.

The Commercial Consequences

For Apple, the stakes of getting Siri right extend beyond product embarrassment. The company’s primary hardware premium — the argument that iPhone and Mac command higher prices than Android alternatives — has historically rested on software quality and ecosystem integration. As AI assistants become increasingly central to how people interact with their devices, a Siri that demonstrably lags Gemini on Google’s own Android devices, or ChatGPT on iPhones where users download OpenAI’s app, erodes the core value proposition.

The $1 billion annual payment to Google is an extraordinary commitment — larger than most AI companies’ entire annual revenue — and reflects Apple’s calculation that the cost of licensing Gemini is lower than the reputational cost of shipping another generation of mediocre Siri features. But licensing alone does not guarantee successful integration. The Phase 2 delays are a reminder that the hard work is not acquiring access to capable AI, but weaving it into a product experience that feels seamless.

Early adopters who have installed iOS 26.4 describe the Phase 1 Siri improvements as “noticeably better, not transformationally better” — a verdict that is probably fair. The real test will come in September, when iOS 27 ships with the full Phase 2 capabilities. If Apple delivers on what it promised in January, the narrative shifts from “Apple’s AI embarrassment continues” to “Apple solved the privacy problem no one else attempted.” If Phase 2 slips again, the $1 billion annual Gemini deal will look like an expensive acknowledgment of a competitive deficit that Apple still cannot close on its own timeline.

The history of Apple product cycles suggests that first-generation shipping is rarely the full story. But its users have been waiting for a Siri upgrade for years. Their patience, like the Phase 2 rollout, has a deadline.

Apple Siri Google Gemini iOS 26.4 Apple Intelligence AI mobile

Related Stories

Apple Sets April 28 iOS 26 SDK Deadline as EU Forces Open App Distribution

Apple has announced that all App Store submissions must use the iOS 26 SDK starting April 28, 2026, forcing developers to update their toolchains. Simultaneously, EU Mobile Software Competition Act compliance in iOS 26.2 is unlocking alternative app marketplaces, third-party payment processing, and independent distribution for the first time — the biggest structural change to the App Store since its 2008 launch.

5 min read

Apple's Secret AI Server Chip Gets a Glass Upgrade — and Samsung Is Supplying It

Apple is advancing development of its first custom AI server chip, codenamed Baltra, while Samsung Electro-Mechanics has begun supplying experimental glass substrate samples for packaging. The collaboration marks Apple as the last major tech hyperscaler to build its own AI silicon, while the glass substrate technology represents a potential generational leap in chip packaging that could reshape AI hardware economics in the 2027–2028 timeframe.

5 min read