Google has crossed a line that AI assistants have been edging toward for years. With Gemini task automation, the company’s AI can now open applications, navigate their interfaces, and complete multi-step tasks on your behalf, all while you get on with something else. The early use cases are deliberately mundane, but the technology behind them is anything but.
Rolling out from March 3 as part of Google’s 2026 Pixel Drop, it went live on the Samsung Galaxy S26 series from March 11, with Pixel 10 users next. The beta is currently limited to the US and Korea, with more countries and regions set to follow.
What is Gemini Task Automation and how does it work?
Long-press the power button, give Gemini a natural language instruction: “book me a ride to the airport,” “reorder my last DoorDash meal”, and Gemini takes over. It launches the relevant app inside a “secure, virtual window” and navigates it autonomously: tapping, scrolling, filling fields. The automation runs in an isolated environment processed in the cloud; the rest of your device remains entirely out of reach, and users must explicitly opt in before the feature activates.
A live notification narrates every step, and you can view progress, take control, or stop at any point. Before any final action, such as placing an order or confirming a booking, Gemini pauses and hands that last step back to you. That human-in-the-loop checkpoint reflects something Google has been explicit about. As CEO Sundar Pichai put it at the AI Impact Summit in February: “Trust is the bedrock of adoption.”
In beta, task automation covers food delivery (DoorDash, Grubhub), grocery ordering, and rideshare (Uber): a narrow starting point while Google gathers feedback ahead of a broader rollout.
Why this is a turning point for the future of work
The significance isn’t the DoorDash integration. It’s what the architecture demonstrates: an AI that reads a live app interface it wasn’t trained on, navigates it in real time, makes contextual inferences, and hands off at the moment of commitment. That’s not a chatbot. That’s an agent, and agents are what the next phase of workplace AI looks like.
Map those capabilities onto everyday knowledge work: logging a call in a CRM, updating a helpdesk ticket, rescheduling a meeting across calendar and video tools. These multi-step tasks are where working time quietly disappears. The data from organisations already deploying Gemini-powered agents makes the opportunity concrete. Telus has 57,000 staff saving 40 minutes per AI interaction, while Danfoss cut order response times from 42 hours to near real-time by automating 80% of transactional decisions.
As Google Cloud’s Global MD for Strategic Industries, Anil Jain, put it in the company’s 2026 AI Agent Trends Report:
“We’re moving away from abstract, future-gazing possibilities, and focusing on delivering tangible business value right now.”
Pichai was equally direct at the AI Impact Summit: “AI will undeniably reshape the workforce, automating some roles, evolving others and creating entirely new careers.”
What businesses should be asking about agentic AI right now
For organisations, how Google has designed for safety matters as much as the capability itself. The feature rests on three principles: control (automations start and stop on your command), transparency (every step narrated, the virtual window always accessible), and access (structural isolation, so Gemini cannot reach beyond the task window).
As we’ve covered previously at UC Today, trust remains the central challenge in enterprise agentic AI adoption, and any agent with application-layer access creates a new risk surface. Technology leaders should be pressing vendors on exactly this: what does the agent see, what can it do, and where are the hard limits?
Task automation is still a beta, scoped deliberately to low-stakes scenarios, but expansion is a plan, not a possibility. As Jain wrote:
“In 2026, we’ll see businesses connecting agents according to their needs, running entire workflows from start to finish.”
The question for organisations is no longer whether agentic AI will reach their workflows. It’s whether their strategy and governance are ready when it does. The architecture is in place. The future of work starts with ordering a coffee, and scales from there.






