Google Gemini Now Controls Your Screen Galaxy S26 Users Get First Taste of AI Automation.
The future of hands-free mobile interaction has arrived. Google has begun rolling out Gemini Screen Automation, a cutting-edge feature that allows the AI to navigate and control smartphone apps on behalf of the user. Currently, the feature is live for Samsung Galaxy S26 Series users in the United States and South Korea, while Pixel 10 Series owners are surprisingly still on the waiting list.
From Chatbot to Personal Assistant
Gemini Screen Automation transforms the AI from a simple conversationalist into a functional agent. By typing a natural language prompt, users can sit back while Gemini takes over the screen to:
Order Food: Navigate through delivery apps to select items and reach the checkout.
Shop Online: Find products and add them to carts across e-commerce platforms.
Book Rides: Open ride-hailing apps and set pickup/drop-off locations automatically.
While the current list of supported apps remains limited, the "magic" lies in Gemini's ability to understand UI elements and interact with them just like a human would.
Usage Tiers and Quotas
Google has introduced a tiered system for this high-compute feature:
Free Users: 5 uses per day.
Google AI Plus: 12 uses per day.
Pro Subscribers: 20 uses per day.
Ultra Subscribers: 120 uses per day.
As the rollout continues, the industry is watching closely to see which other flagship devices will gain access to this game-changing AI capability.
What sets Screen Automation apart from traditional voice commands is that it doesn't just open apps; it "sees" the screen the way a human sees it. This technology uses what's called Multimodal Perception to analyze on-screen buttons and text in real-time, allowing it to bypass complex steps in unfamiliar apps.
It's worth noting why the Pixel 10 received this feature after Samsung. Analysts view this as a "AI Exclusive Partnership" strategy between Google and Samsung to boost Galaxy S26 sales globally. At the same time, Google is refining security on the Pixel, as allowing AI to control the screen poses a significant risk to personal data privacy.
Security is paramount. Google has implemented a "Visual Sandbox," where users see a blue frame around the screen while the AI is working and can manually override it immediately if the AI starts doing something unusual, such as attempting to confirm a payment without user verification.
If this feature becomes popular, app developers worldwide will need to adapt their UI designs to be more AI-ready, because in the future, their "users" may not always be humans, but AI agents performing transactions instead.
YouTube CEO Neal Mohan Announces New 50MB Thumbnail Limit to Enhance Big-Screen Experience.
Source: 9to5Google

Comments
Post a Comment