a system governed by AI agents

The way we use mobile apps could be entering a new stage. Until now, the Android experience has been based on something very simple: opening applications and performing step-by-step actions within them. However, Google is exploring a different model, in which artificial intelligence acts as an intermediate layer between what we ask for and what apps can do. In that scenario, we won’t always be the ones scrolling through menus or completing processes manually. In many cases, it will be enough to express what we want to do so that the system will try to solve it for us, coordinating different phone functions. The next step in Android. In a post on the official developer blogthe company presents new capabilities designed so that applications can work directly with assistants and AI systems. These functions are designed so that tools like Gemini can discover and execute certain actions within some apps. The project is still in an early phase, but it suggests a very specific direction: begin to reconfigure Android as an environment in which artificial intelligence can help complete tasks. What do we understand by agent. In the field of AI, an agent is a designed system to move from response to action. While early digital assistants functioned as consultation tools, agents attempt to understand an intention and plan how to carry it out. To do so, they combine several capabilities: understanding natural language, evaluating the context and deciding what steps are necessary to fulfill a request. It is not just about generating text or suggestions, but about organizing a small chain of decisions oriented towards a specific objective. If we follow the reasoning that Google presents in its publication, the change does not only affect AI, but also how applications are conceived within Android. For years, the main objective of any app was to get the user to open it and complete all the necessary actions within it. However, now that criterion is beginning to shift. In this new scenario, success begins to be measured less by getting us to open an app and more by its ability to help complete a task, even when the user does not directly interact with its entire interface. One of the first pieces of change. The first path that Google proposes to move in this direction goes through something it calls AppFunctions. It is not a user-visible function as such, but rather a set of tools with which developers can expose functions and data of their apps to intelligent assistants such as Gemini. The example mentioned by the Android blog itself is quite illustrative: on the recently introduced Galaxy S26 seriesGemini can access Samsung Gallery features to locate specific photos based on a natural language request, such as asking to show images of a pet. In that case, the assistant interprets the request, activates the corresponding Samsung Gallery function and returns the result without requiring the user to manually navigate through the gallery. The other way of Google. Along with direct integrations, the company is preparing a second formula to extend this model to more applications. As he explains, it is an interface automation system that will allow Gemini to take care of generic multi-step tasks without depending on a specific connection between the app and the assistant. Instead of relying on a function previously exposed by the application, the AI ​​acts directly on the interface. Google notes that this initial preview will be tested on the Galaxy S26 series and some Pixel 10within the Gemini app and with a limited selection of delivery, grocery and transportation applications in the United States and Korea. The company also ensures that the user will be able to follow the process through notifications or a live view, resume manual control at any time and receive notifications before sensitive actions, such as a purchase. Looking to the future. If Google’s announcement makes anything clear, it is that Android is beginning to prepare for a different stage. The functions presented are still in development and their deployment will be gradual, but they point to a specific direction: an operating system in which artificial intelligence plays an increasingly active role in the way we perform daily actions on mobile phones. Pixel and Samsung appear for now as the most visible references, although Google suggests that it wants to bring these capabilities to more manufacturers as the ecosystem evolves. As is often the case with these types of changes, the final result will depend on how the tools, integrations and the response of the users themselves evolve. Images | Google In Xataka | The iPhone has been a “made in China” phone for decades. Now it is changing countries at full speed: India

These are Samsung’s augmented reality glasses. Project Moohan is coming, and it will be governed by Android XR

The celebration of Samsung Unpacked 2025 has provided not one, but two surprises. The first, the fleeting presentation of the Samsung Galaxy S25 Edgethe ultra-thin mobile that we were able to see in the exhibition area after the event. The second, the augmented reality glasses that Samsung has been developing for some time. They are the result of Project Moohan that Samsung spoke briefly about at the event, and that will precisely compete with devices like the Apple Vision Pro or the striking Orion of Meta. The design of the glasses is in fact very similar to that of Apple’s glasses, although it also has certain design lines apparently inherited from the Meta Quest Pronow discontinued. During the presentation, Samsung mentioned its intention to bring AI options to other form factors such as connected glasses – Ray-Ban Meta type, we imagine – and augmented reality glasses. Samsung actually announced that it was working on augmented reality glasses in collaboration with Google. That is the goal of Project Moohan, who also stands out for be governed by Android XR. This operating system will be a competitor to visionOS, Apple’s proposal in the Vision Pro. At the moment the augmented reality glasses could be seen on display and we were able to photograph them, but there are no more details about them: we have to wait. In Xataka | Mark Zuckerberg believes that in 2030 we will not take our smartphones out of our pockets so much: we will do almost everything from our glasses

Log In

Forgot password?

Forgot password?

Enter your account data and we will send you a link to reset your password.

Your password reset link appears to be invalid or expired.

Log in

Privacy Policy

Add to Collection

No Collections

Here you'll find all collections you've created before.