Google's long-awaited AI vision is finally becoming a reality with the release of

Visual Impact: Watching the Phone "Use Itself"
Differently from traditional API integration,
Smart Taxi Hailing: When you give the instruction "Hail a taxi to the airport,"
will automatically open Uber, confirm the specific terminal (if there are multiple terminals, it will ask proactively), and automatically fill in the destination. Ordering Food: When processing the instruction "Order me a coffee and a croissant," the AI will independently scroll through the screen to find specific items on the Starbucks menu (such as Flat White), and even handle complex scrolling selections like a human would.
Security Logic: Human Control at Key Points
To avoid the risks associated with autonomy, Google has implemented a strict human review mechanism in the automation process:
Explicit Operation: Users can watch
Last Confirmation: Before submitting an order or payment, the system will stop at the payment screen, waiting for the user to verify the details and manually click "Confirm," ensuring that each transaction is completed under controlled conditions.
Currently, this feature is prioritized for delivery and ride-hailing applications. For
Although AI occasionally appears somewhat "clumsy" in scrolling menus and identifying options, this automation model that does not require deep API adaptation and instead works directly with UI interactions greatly expands the application boundaries of AI assistants. With algorithm iterations, we are moving away from the era of repeatedly switching between apps and entering a truly intelligent stage where all small tasks can be completed with a single sentence.
