Qwen2.5-VL
Qwen2.5-VL is the latest vision-language model from the Qwen series, representing a significant advancement over its predecessor, Qwen2-VL. This model excels in visual understanding, capable of recognizing a wide array of objects, including text, charts, icons, graphics, and layouts within images. It functions as a visual agent, capable of reasoning and dynamically directing tools, enabling applications such as computer and phone usage. Qwen2.5-VL can comprehend videos exceeding one hour in length and can pinpoint relevant segments within them. Additionally, it accurately localizes objects in images by generating bounding boxes or points and provides stable JSON outputs for coordinates and attributes. The model also supports structured outputs for data like scanned invoices, forms, and tables, benefiting sectors such as finance and commerce. Available in base and instruct versions across 3B, 7B, and 72B sizes, Qwen2.5-VL is accessible through platforms like Hugging Face and ModelScope.
Learn more
Agent S2
Agent S2 is an open, modular, and scalable framework for computer-use agents developed by Simular. These autonomous AI agents interact directly with graphical user interfaces (GUIs) on desktops, mobile devices, browsers, and various software applications, mimicking human-like control via mouse and keyboard. Building upon the initial Agent S framework, Agent S2 enhances performance and modularity by integrating both frontier foundation models and specialized models. It achieves state-of-the-art results, notably surpassing previous benchmarks on OSWorld and AndroidWorld evaluations. Key design principles include proactive hierarchical planning, where the agent dynamically updates its plans after each subtask; visual grounding for precise GUI interaction using raw screenshots; an improved Agent-Computer Interface (ACI) that delegates complex tasks to specialized modules; and an agentic memory mechanism that enables continual learning from experience.
Learn more
Gemini 2.5 Computer Use
Introducing the Gemini 2.5 Computer Use model, a specialized agent model built on top of Gemini 2.5 Pro’s visual reasoning capabilities, designed to interact directly with user interfaces (UIs). It is exposed via a new computer-use tool in the Gemini API, with inputs that include the user’s request, a screenshot of the UI environment, and a history of recent actions. The model generates function calls corresponding to UI actions like clicking, typing, or selecting, and may request user confirmation for higher-risk tasks. After each action is executed, a new screenshot and URL are fed back into the model to continue the loop until the task completes or is halted. It is optimized primarily for web browser control and shows promise for mobile UI interaction, though it is not yet suited for desktop OS-level control. In benchmarks across web and mobile control tasks, Gemini 2.5 Computer Use outperforms leading alternatives, delivering high accuracy at lower latency.
Learn more
Dialogflow
Dialogflow from Google Cloud is a natural language understanding platform that makes it easy to design and integrate a conversational user interface into your mobile app, web application, device, bot, interactive voice response system, and so on. Using Dialogflow, you can provide new and engaging ways for users to interact with your product. Dialogflow can analyze multiple types of input from your customers, including text or audio inputs (like from a phone or voice recording). It can also respond to your customers in a couple of ways, either through text or with synthetic speech. Dialogflow CX and ES provide virtual agent services for chatbots and contact centers. If you have a contact center that employs human agents, you can use Agent Assist to help your human agents. Agent Assist provides real-time suggestions for human agents while they are in conversations with end-user customers.
Learn more