Gemini 2.5 Computer Use
Introducing the Gemini 2.5 Computer Use model, a specialized agent model built on top of Gemini 2.5 Pro’s visual reasoning capabilities, designed to interact directly with user interfaces (UIs). It is exposed via a new computer-use tool in the Gemini API, with inputs that include the user’s request, a screenshot of the UI environment, and a history of recent actions. The model generates function calls corresponding to UI actions like clicking, typing, or selecting, and may request user confirmation for higher-risk tasks. After each action is executed, a new screenshot and URL are fed back into the model to continue the loop until the task completes or is halted. It is optimized primarily for web browser control and shows promise for mobile UI interaction, though it is not yet suited for desktop OS-level control. In benchmarks across web and mobile control tasks, Gemini 2.5 Computer Use outperforms leading alternatives, delivering high accuracy at lower latency.
Learn more
Lux
Lux is a powerful computer-use AI platform that enables agents to operate software just like a human user—clicking, typing, navigating, and completing tasks across any interface. It offers three execution modes—Tasker, Actor, and Thinker—giving developers the ability to choose between step-by-step precision, near-instant task execution, or long-form reasoning for complex workflows. Lux can autonomously perform actions such as crawling Amazon data, running automated QA tests, or extracting insights from Nasdaq’s insider activity pages. The platform makes it possible to prototype and deploy real computer-use agents in as little as 20 minutes using developer-friendly SDKs and templates. Its agents are built to understand vague goals, execute long-running operations, and interact naturally with human-facing software instead of relying solely on APIs. Lux represents a new paradigm where AI goes beyond reasoning and content generation to directly operate computers at scale.
Learn more
OpenClaw
OpenClaw is an open source autonomous personal AI assistant agent you run on your own computer, server, or VPS that goes beyond just generating text by actually performing real tasks you tell it to do in natural language through familiar chat platforms like WhatsApp, Telegram, Discord, Slack, and others. It connects to external large language models and services while prioritizing local-first execution and data control on your infrastructure so the agent can clear your inbox, send emails, manage your calendar, check you in for flights, interact with files, run scripts, and automate everyday workflows without needing predefined triggers or cloud-hosted assistants; it maintains persistent memory (remembering context across sessions) and can run continuously to proactively coordinate tasks and reminders. It supports integrations with messaging apps and community-built “skills,” letting users extend its capabilities and route different agents or tools through isolated workspaces.
Learn more
Qwen2.5-VL
Qwen2.5-VL is the latest vision-language model from the Qwen series, representing a significant advancement over its predecessor, Qwen2-VL. This model excels in visual understanding, capable of recognizing a wide array of objects, including text, charts, icons, graphics, and layouts within images. It functions as a visual agent, capable of reasoning and dynamically directing tools, enabling applications such as computer and phone usage. Qwen2.5-VL can comprehend videos exceeding one hour in length and can pinpoint relevant segments within them. Additionally, it accurately localizes objects in images by generating bounding boxes or points and provides stable JSON outputs for coordinates and attributes. The model also supports structured outputs for data like scanned invoices, forms, and tables, benefiting sectors such as finance and commerce. Available in base and instruct versions across 3B, 7B, and 72B sizes, Qwen2.5-VL is accessible through platforms like Hugging Face and ModelScope.
Learn more