Computer Use & GUI Agents, when the interface itself becomes the tool.

Every API in the previous chapter is a privileged, pre-designed gateway. But most of the world's software doesn't have an API — it has a screen. Tax software, legacy enterprise dashboards, medical record systems, insurance portals: all interface-first, all effectively invisible to the function-calling paradigm. GUI agents change that. By watching a screen and controlling a mouse and keyboard, an agent can operate any software any human can operate — which is essentially all of it. This chapter covers how that works, why it's hard, and where the benchmarks put us today.

Prerequisites

This chapter builds on Tool Use & Function Calling (Ch 05) for the agent execution loop, and on Vision-Language Models (Part VII Ch 06) for the multimodal perception backbone that powers screen understanding. The benchmarks section connects to Planning & Reasoning (Ch 03) since multi-step web navigation is primarily a planning challenge. Familiarity with the basics of HTML DOM structure is helpful for the accessibility tree sections.

Why GUI Agents

Motivation · The API Gap

The tool-calling paradigm of Chapter 05 rests on a crucial assumption: that every capability an agent needs has been exposed as a structured API. In practice, this assumption fails constantly. Most enterprise software is GUI-first. Small businesses use QuickBooks, Salesforce, or custom internal tools that were never designed with API consumers in mind. Governments serve citizens through web portals. Healthcare systems store records in ancient EHRs with proprietary interfaces. The world runs on GUIs, and GUIs have no universal API.

GUI agents bridge this gap. Rather than requiring software vendors to expose structured interfaces, a GUI agent observes the screen, infers the interface structure, and drives the software through the same mouse and keyboard actions a human would use. This unlocks an enormous surface area of automation that was previously inaccessible — every application becomes, in principle, automatable.

The Robotic Process Automation Connection

This idea is not new. Robotic Process Automation (RPA) tools — UiPath, Automation Anywhere, Blue Prism — have automated GUI workflows for enterprise for over a decade. Classical RPA works by recording a human's interaction with a specific application version and replaying it — brittle to any change in layout, labelling, or workflow. LLM-based GUI agents replace the brittle replay approach with dynamic understanding: the agent reads the current screen and reasons about what to do, rather than following a pre-recorded script. This makes them far more robust to interface changes and capable of handling novel situations the recording never anticipated.

The Interface as Universal API

Every piece of software designed for human use has an implicit interface contract: it shows the user information, offers controls, and responds to input. A sufficiently capable GUI agent can read any interface and drive any control — which means the interface is the API, and the agent is the adapter. This reframes GUI agents not as a niche automation technique but as a fundamental expansion of what agents can operate on.

Observation Spaces

Perception · Three Modalities

A GUI agent must perceive the current state of the interface in order to decide what to do next. Three distinct observation representations are used in practice, each with different information content, token cost, and availability.

Raw Pixels
Screenshots

Full-resolution or downsampled screenshots, typically passed to a vision-language model. Contains all visible information including images, icons, layout, and rendered text. Works on any application. High token cost; requires visual grounding to identify actionable elements.

Accessibility Tree
A11y / AXTree

Structured tree of interactive elements exposed by the OS accessibility APIs (ARIA roles, element labels, coordinates). Text-only, compact, highly actionable. Works for applications that implement accessibility. Can miss custom-rendered elements and canvas-based UIs.

HTML / DOM
Page Source

The raw or cleaned HTML source of a web page, giving the agent full access to element IDs, classes, ARIA labels, and text content. Highly reliable for web tasks. Not available for desktop applications or pages with heavy client-side rendering that doesn't annotate elements.

Hybrid Observations

The strongest systems combine modalities. The typical hybrid: screenshot for spatial context and visual layout, accessibility tree or DOM for reliable element identification and labels, and sometimes additional structured metadata (page title, URL, tab list). The agent reasons using text (token-efficient) and uses the screenshot to verify or localise elements that the structured representations misidentify or omit. SoM (Set-of-Mark) prompting, introduced by Yang et al. (2023), renders the accessibility tree element IDs directly as numbered overlay boxes on the screenshot — giving the vision model both spatial context and element identifiers simultaneously.

Set-of-Mark (SoM) Prompting

SoM overlays numeric tags on every interactive element in a screenshot, then passes the tagged image to a VLM. The model outputs an action referencing an element by its number ("click element 14") rather than by pixel coordinates. This decouples visual localisation from semantic reasoning: the model no longer needs to predict precise coordinates, only to identify which tagged element to interact with. SoM-based agents show dramatically better grounding accuracy than coordinate-prediction baselines on complex UIs.

Pixels vs. Accessibility Trees

Trade-offs · Representation Choice

The choice between pixel-based and accessibility-tree-based observation is one of the central design decisions in GUI agent engineering. Neither dominates; the right choice depends on the target application, the available model capabilities, and the task structure.

Property Raw Pixels Accessibility Tree HTML/DOM
Coverage Universal — any rendered UI Requires a11y support Web only
Token cost High (image tokens) Low (compact text tree) Medium–high (full HTML)
Precise coords Implicit in image — must predict Explicit bounding boxes Requires JS to resolve
Dynamic content Always up to date May lag rendering May miss JS-rendered state
Canvas / SVG Fully visible Often inaccessible Inaccessible
Model requirement VLM mandatory Text LLM sufficient Text LLM sufficient
Best for Desktop apps, games, custom UIs Standard web / native UIs Web scraping & form-filling

The Accessibility Tree as Ground Truth

For web and standard native applications, accessibility trees provide more reliable grounding than pixel-based approaches. An element's ARIA role, name, and bounding box are precisely specified — there is no ambiguity about whether the model has correctly localised a "Submit" button. Errors in a11y-based agents are more often semantic (wrong decision about what to click) rather than localisation errors (clicking the wrong pixel). This makes them easier to debug and more robust to rendering differences across screen resolutions and zoom levels.

When Pixels Are Necessary

Pure-pixel approaches are unavoidable for: (1) applications that don't implement accessibility APIs (legacy desktop software, many games, custom kiosk UIs); (2) elements rendered on canvas or in WebGL, which have no DOM representation; (3) CAPTCHA-like challenges that are deliberately inaccessible to screen readers; (4) tasks where the visual presentation contains information not encoded in the structure (colour-coded status indicators, sparkline charts, icon meanings). VLMs with strong spatial reasoning are essential for these cases.

Action Spaces

Control · What an Agent Can Do

An agent's action space defines the vocabulary of operations it can perform on the interface. The design of the action space significantly affects task complexity: a low-level space (individual pixel clicks and keystrokes) is maximally expressive but requires far more steps and planning; a high-level space (semantic actions like "fill form field X with value Y") is efficient but requires mapping from the model's output to concrete UI elements.

Low-Level Actions

click(x, y) right_click(x, y) double_click(x, y) drag(x1,y1, x2,y2) hover(x, y) scroll(x, y, dir, amt) key(key_name) type(text) wait(ms)

Low-level actions operate directly on the OS input system via libraries like PyAutoGUI, xdotool (Linux), or Win32 SendInput (Windows). They work on any application but require precise coordinate prediction — a 5-pixel error can cause a click to miss an interactive element entirely on dense UIs.

Mid-Level Actions

Mid-level actions abstract over element identity rather than coordinates: click_element(element_id), type_in_field(field_label, value), select_option(dropdown_id, option_text). In browser contexts, these are implemented via the Playwright or Puppeteer APIs; in desktop contexts, via accessibility APIs (AT-SPI on Linux, UIAutomation on Windows). Mid-level actions are more reliable than coordinate-based actions because they don't depend on precise spatial prediction.

High-Level Actions

High-level actions encode task-specific operations: search(query), navigate_to(url), submit_form(), login(username, password). These are composed of multiple lower-level operations and implemented in the executor layer. They dramatically reduce the number of steps in a task trajectory but sacrifice generality — a new application requires a new set of high-level action definitions.

The most capable modern GUI agents use a hybrid: high-level actions for common patterns (navigation, form submission, text selection) and fall back to mid- or low-level actions for anything not covered. This mirrors how expert human computer users operate — keyboard shortcuts and menus for common operations, precise mouse control only when necessary.

Visual Grounding

Localisation · Connecting Text to Pixels

Visual grounding is the problem of mapping a semantic description ("the Submit button", "the search field", "the notification bell icon") to a concrete location on the screen — a bounding box or pixel coordinate. It is the fundamental perception challenge of GUI agents: without accurate grounding, even perfect high-level reasoning produces wrong actions.

Grounding via VLM Direct Prediction

The simplest approach asks the VLM to predict the target element's bounding box or centre coordinates directly. Given a screenshot and a description of the desired element, the model outputs \((x, y)\) coordinates (normalised to \([0,1]^2\) relative to image dimensions). The accuracy of this approach depends heavily on the spatial precision of the VLM's training. GPT-4V and Claude 3 are mediocre at this task; models specifically fine-tuned for grounding (CogAgent, SeeClick, UGround) are significantly better.

Grounding as Conditional Localisation
\[(x^*, y^*) = \underset{(x,y)}{\arg\max} \; P_\theta\!\left((x,y) \mid I_t, q, \mathcal{H}_{1:t-1}\right)\]
The agent predicts the screen coordinates \((x^*, y^*)\) of the target element conditioned on the current screenshot \(I_t\), the action query \(q\) (e.g., "click the search bar"), and prior interaction history \(\mathcal{H}\). Models trained on large-scale GUI interaction traces learn to resolve ambiguous element references using contextual history.

Grounding via SoM + Element ID

SoM (Set-of-Mark) side-steps direct coordinate prediction: by rendering numbered tags over all interactive elements before presenting the screenshot to the model, grounding reduces to classification over a small set of numbered elements. The coordinate resolution is delegated to the tag-rendering layer, which has ground truth. This is more reliable than VLM coordinate regression for most models and is the dominant approach in high-performance web agents as of 2025.

Grounding via OCR + Fuzzy Match

For text-heavy UIs, OCR (extracting text and its bounding boxes from a screenshot) provides reliable element candidates. The model describes the target element in text; OCR produces a list of text regions with coordinates; fuzzy matching selects the best candidate. This is cheap and accurate for button labels, menu items, and form fields — and fails for icon-only controls with no text label. Combined with VLM fallback for icon cases, OCR+fuzzy-match is a practical production pattern.

Browser Agents

Web · Playwright, CDP, Headless

Browser agents are GUI agents specialised for web interfaces. They use browser automation frameworks — Playwright, Puppeteer, or the Chrome DevTools Protocol (CDP) directly — to control a headless or headed browser. This provides both low-level control (simulate any user action) and rich structural access (full DOM, network requests, JavaScript execution).

The Browser as Structured Environment

Web browsers are unusually rich environments for agent operation. Beyond rendering the page, they expose: the full DOM tree (accessibility tree derivable from it), network request/response logs (letting the agent observe API calls), JavaScript execution (agent can query the DOM programmatically, fill forms directly, extract data without visual parsing), and browser state (cookies, local storage, tabs, history). An agent that can run JavaScript in the browser has access to a far richer observation than a screenshot alone.

// Browser agent — hybrid observation extraction

// Screenshot for visual context
screenshot = page.screenshot(full_page=False)

// DOM snapshot for element grounding
a11y_tree = page.accessibility.snapshot()

// Active element metadata
focus_info = page.evaluate("() => ({ tag: document.activeElement.tagName, id: document.activeElement.id })")

obs = {"image": screenshot, "tree": a11y_tree, "url": page.url, "focus": focus_info}

Key Browser Agent Challenges

Dynamic content: pages rendered heavily by JavaScript may show a nearly empty DOM until JS executes; agents must wait for page stability before extracting observations. Authentication: many web tasks require login, which involves CAPTCHA, two-factor authentication, and session management — all outside the agent's standard action space. Pop-ups and modals: cookie consent banners, newsletter pop-ups, and chat widgets interrupt task trajectories constantly. Infinite scroll: content that loads on scroll requires the agent to recognise when more content is available and scroll to reveal it, unlike paginated navigation. Each of these is a recurring failure mode in web navigation benchmarks.

Browser Use and the Emerging Ecosystem

The open-source browser-use library (2024) popularised a lightweight architecture: Playwright for browser control, a VLM for understanding the page, and a simple action loop. It achieved state-of-the-art results on several web benchmarks with a surprisingly compact implementation. Commercial offerings include Browserbase (cloud browser infrastructure), Playwright-as-MCP-server integrations, and integrated browser agents in platforms like Zapier and Make.

OS-Level Control

Desktop · Cross-Application

OS-level agents operate at the full desktop — not just within a browser, but across any application: native file managers, spreadsheets, IDEs, terminal emulators, media players, and custom enterprise software. They interact with the OS via two complementary channels: the accessibility API (structured element access) and raw input injection (simulating mouse and keyboard).

OS Accessibility APIs

Operating systems expose accessibility APIs for assistive technology (screen readers, switch access). These same APIs are available to agents. On macOS, the Accessibility API (via AXUIElement) provides a tree of every visible UI element with its role, label, value, and bounding box. On Windows, UI Automation (UIA) provides equivalent access. On Linux, AT-SPI covers GTK and Qt applications. These APIs make OS-level agents far more reliable than pure-pixel approaches for standard applications.

Claude's Computer Use

Anthropic's Computer Use capability (released October 2024 in beta) demonstrated the first commercially available OS-level agent from a frontier lab. The model receives a screenshot and issues low-level actions — computer.type, computer.key, computer.mouse_move, computer.left_click, computer.screenshot — running in a Docker containerised desktop environment. A distinguishing design choice: the model explicitly calls screenshot as a tool action to observe the result of each interaction, making the perception-action loop explicit and auditable. Early evaluations showed ~22% task completion on OSWorld, with failures concentrated on tasks requiring long multi-step desktop workflows.

Cross-Application Workflows

OS-level agents unlock workflows that span application boundaries — the most common real-world automation pattern. Copy data from an Excel spreadsheet, look it up in a CRM, paste results into an email, attach a PDF from the file system, and send — this requires four application switches and would need custom API integration in the tool-calling paradigm. For a GUI agent, it's one unified perception-action loop over the same desktop environment a human uses. The challenge is maintaining context across application switches when the agent's observation is always just the current screen.

The Context Loss Problem

When switching applications, the prior application's UI disappears from the screenshot. The agent must track in its working memory what state it left the prior application in — what row it was on in the spreadsheet, what field it was filling in the CRM. Long multi-application workflows strain working memory in exactly the same way long document tasks do. External state tracking (a simple text scratchpad the agent writes to) significantly improves cross-application task completion.

The Perception-Action Loop

Architecture · Observe-Reason-Act

GUI agents implement a specialised perception-action loop where each cycle consists of: observe the current screen state, reason about what to do next, act on the interface, and observe the result. This is the sense-plan-act loop from Chapter 01, instantiated over a visual interface rather than an abstract environment model.

Screen / Environment Submit Perception screenshot + a11y tree Reasoning VLM plans next action Action Executor mouse / keyboard / JS Working memory (task goal + action history) observe action command input event Loop repeats until task complete or max steps exceeded
GUI agent perception-action loop: each cycle observes the current screen (screenshot + accessibility tree), reasons using a VLM conditioned on task goal and history, selects an action, and dispatches it to the executor which drives the actual OS or browser input.

Step Frequency and Latency

Each cycle of the perception-action loop involves at least one VLM call, which adds 1–5 seconds of latency depending on model size and serving infrastructure. A 20-step web task therefore takes 20–100 seconds of pure model latency, before adding browser rendering time. This is acceptable for async automation (running in the background) but painful for interactive use. Several optimisations help: caching observations that haven't changed between steps, using smaller fast models for simple steps and reserving the frontier model for complex decisions, and predicting multiple steps ahead when the trajectory is predictable.

Error Recovery in GUI Loops

Unlike API tool calls, GUI actions can fail silently. A click that misses its target registers as a valid mouse event — no error is returned. The only signal is the next screenshot, which shows an unchanged or unexpected state. Robust GUI agents implement explicit state-change verification: after every action, compare the new screenshot with the prior one to confirm that the expected change occurred, and if not, re-examine the situation and retry with a corrected approach. This adds one screenshot call per action but dramatically improves reliability.

WebArena & Mind2Web

Web Benchmarks · Navigation Tasks

Web navigation benchmarks evaluate agent performance on realistic tasks across real-looking websites — shopping, booking, form-filling, information lookup, account management. They have become the primary measure of progress in browser agent capability.

WebArena

Zhou et al. (2023) introduced WebArena: a self-hosted benchmark with five realistic websites — an e-commerce platform (OneStopShop), a Q&A forum (Reddit clone), a GitLab code repository, a CMS (Content Management System), and a maps application. Tasks span all five sites and require multi-step navigation (up to 20 steps), form filling, cross-site information retrieval, and conditional logic. The evaluation is end-to-end: tasks are scored on whether the final page state matches the expected outcome, not on whether individual steps were correct.

Early GPT-4 baselines achieved only 14.9% task completion, exposing how hard real-world web navigation is even for frontier models. By 2025, the best published agents reached ~55–60% on the original WebArena tasks, with the remaining failures concentrated on tasks requiring long multi-step planning, understanding dynamic content, and handling authentication flows.

Mind2Web

Deng et al. (2023) took a different approach: rather than self-hosted sites, Mind2Web uses 2,000 tasks across 137 real websites from the public web, collected via human demonstrations. The evaluation measures element localisation accuracy (did the agent identify the right element to interact with?) and action type accuracy (did it choose the right action — click, type, select?), decomposing task performance into its component parts. Mind2Web revealed that element selection — even before action planning — is the primary bottleneck: models with similar reasoning capabilities differ enormously in their ability to correctly ground the target element in complex real-world pages.

Benchmark Sites Tasks Metric Best (2025)
WebArena 5 self-hosted 812 Task success rate ~58% (top agents)
Mind2Web 137 real websites 2,000 Element accuracy / step SR ~55% step SR
VisualWebArena 5 self-hosted 910 Task success (visual queries) ~35% (vision-required)
WorkArena ServiceNow 33 task types Task success rate ~40%
WebVoyager 15 real websites 643 Task success (GPT-4V judge) ~87% (w/ SoM)

The Gap Between Benchmarks and Deployment

Benchmark results overstate production performance in two important ways. First, benchmarks use controlled, reproducible environments — real websites change layout, introduce new pop-ups, and break without warning. Second, benchmark tasks are carefully scoped and unambiguous — real user requests are often vague ("find me a hotel in Paris for next weekend, reasonably priced") requiring the agent to make judgement calls that introduce additional failure modes. Production browser agent deployments typically report 30–50% lower success rates than the same system's benchmark score.

OSWorld & WindowsAgentArena

Desktop Benchmarks · Full OS Tasks

Desktop benchmarks are more demanding than web benchmarks: they require controlling native applications, managing files, and coordinating across multiple apps — without the DOM structure that makes web tasks more tractable.

OSWorld

Xie et al. (2024) introduced OSWorld: 369 computer tasks across Ubuntu Linux, Windows, and macOS, spanning web browsers, productivity software (LibreOffice, VS Code), file management, image editing (GIMP), and terminal/shell operations. Each task is grounded in a reproducible virtual machine snapshot, making evaluation deterministic. Tasks are long-horizon (median ~8 steps, max ~50) and require cross-application coordination.

The inaugural results were sobering: GPT-4V with screenshot observations achieved 11.8% task completion; Claude 3 (computer use API) reached ~22%. Human performance on the same tasks is ~72%, revealing a 50-percentage-point gap. The hardest task categories — terminal operations, image editing, complex file management — had near-zero model completion rates. OSWorld became the benchmark of record for OS-level agent evaluation and has driven rapid progress: by late 2024, the best systems reached ~30–38% completion.

WindowsAgentArena

Bonatti et al. (2024) extended the evaluation specifically to Windows, with 154 tasks across Windows 11 applications. Windows presents unique challenges: the Win32 and UWP application models have different accessibility API coverage, the Start menu and taskbar require specific interaction patterns, and many enterprise applications (Outlook, Teams, Excel) have extremely dense, information-rich UIs. Windows-specific results tend to lag macOS and Linux results by 5–10 percentage points for the same agent due to accessibility API coverage differences.

What OSWorld Reveals

Analysis of OSWorld failure modes identifies a consistent pattern: agents succeed at recognising what needs to be done, and fail at executing it reliably over long sequences. The per-step accuracy of top agents is ~75–85% — but compounding that over 15 steps gives an overall task success rate of only \(0.80^{15} \approx 3.5\%\) for a random task, explaining the low headline numbers. The critical research challenge is not improving single-step accuracy (already reasonably high) but error recovery and long-horizon consistency.

Compounding Error and the Long-Horizon Cliff

If a single-step error rate is \(\epsilon\), the probability of completing an \(n\)-step task without error is \((1-\epsilon)^n\). At \(\epsilon = 0.15\) (85% single-step accuracy), a 5-step task succeeds with 44% probability; a 20-step task with only 4%. This is the fundamental obstacle for long-horizon GUI agents. Progress requires either improving per-step accuracy dramatically, or building robust error detection and recovery that catches and corrects mistakes mid-trajectory.

SWE-bench & Coding Agents

Code Benchmarks · Real Repositories

SWE-bench (Jimenez et al., 2023) evaluates a specialized form of GUI/computer-use agent: the software engineering agent that must navigate a real codebase, understand an issue description, locate relevant code, write a fix, and verify it with tests. It uses 2,294 real GitHub issues from 12 Python open-source repositories, each paired with the test suite that was added along with the fix.

Why SWE-bench Is Hard

A successful SWE-bench solve requires: understanding a natural-language issue description; navigating a codebase with potentially thousands of files; identifying the root cause, which may span multiple files; writing a correct code patch that is consistent with the codebase's conventions; and ensuring the patch passes existing tests without breaking anything. This is a long-horizon, multi-modal reasoning task over a structured but complex environment. The "GUI" here is the file system, code editor, and terminal — all navigated through tool calls or direct file manipulation.

Progress

The original GPT-4 baseline solved 1.7% of issues. By mid-2024, Devin (Cognition AI) reached 13.9%. By 2025, several systems exceeded 40–49% — among them SWE-agent, OpenHands (formerly OpenDevin), and Claude-based systems using agentic scaffolding. The remaining 50%+ of failures tend to involve issues requiring deep architectural understanding, issues with ambiguous specifications, or fixes that require changes spanning many files and subsystems. SWE-bench Verified (a human-curated subset of 500 confirmed, correctly-specified issues) gives more reliable signal for progress measurement.

The Agent Scaffolding Factor

SWE-bench scores are as much a measure of the agent scaffolding (how the agent is prompted and structured) as of the underlying model. The same frontier model with different scaffolds can vary by 10–15 percentage points on the benchmark. Key scaffolding choices: whether the agent is given the full repo or only relevant files, how the model explores the codebase (file-by-file search vs. semantic code search), how test feedback is incorporated, and how many attempts the agent gets per issue. This makes benchmark comparisons across systems difficult unless scaffolding details are standardised.

Frontier & Open Problems

Frontier · What Comes Next

GUI agents represent one of the fastest-moving areas in applied AI research. The field has gone from academic demos to production deployments in roughly two years, driven by the availability of powerful VLMs and a clear commercial demand for flexible automation. Several open problems define the frontier.

Long-Horizon Reliability

As discussed, compounding error is the fundamental barrier. The most promising approaches are: (1) explicit self-verification after every action (take a screenshot, confirm the expected change occurred, re-plan if not); (2) hierarchical task decomposition with checkpoints, allowing recovery without restarting from scratch; (3) training on long-horizon trajectories with dense reward signals rather than only outcome supervision. None of these is fully solved, but progress is rapid.

Better Screen Understanding Models

General VLMs are trained primarily on natural images and text — they are not optimised for dense UI understanding, small text, icon recognition, or the specific spatial reasoning required for GUI grounding. Specialised GUI foundation models (CogAgent, SeeClick, UGround, Qwen-VL-UI) trained specifically on screen content have consistently outperformed general VLMs on grounding tasks. Training data — large-scale collections of (screenshot, action, outcome) trajectories — is the bottleneck; synthetic data generation from existing automation scripts is an active direction.

Multi-Session and Stateful Workflows

Most benchmark tasks complete in a single session. Real business workflows span days: initiate a request on Monday, follow up on Wednesday when a response arrives, complete the workflow on Friday. GUI agents for real workflows need persistent memory, the ability to recognise when a prior action has been fulfilled, and the capacity to resume interrupted workflows. This connects directly to the memory management architecture of Chapter 04, applied to the GUI context.

Safety and Oversight

A GUI agent with OS-level control can do arbitrary damage: delete files, send emails on behalf of users, make purchases, modify system settings. The safety requirements are substantially more serious than for text-only agents. Current best practices: sandbox execution in VMs for development/testing, request human confirmation before any irreversible action (file deletion, payment, external communication), maintain a full audit log of every action taken, and implement scope restrictions at the OS level (read-only filesystem outside the working directory, network egress whitelist). The challenge of giving agents enough capability to be useful while limiting their blast radius is one of the defining engineering problems of the GUI agent era.

The Universal Automation Horizon

If a GUI agent can reliably complete any task a human performs at a computer — at, say, 95% success rate on 50-step tasks — the economic and social implications are profound. Every white-collar workflow that involves a computer becomes, in principle, automatable without custom integration work. The timeline to that capability is not settled: extrapolating from current benchmark progress suggests 3–5 years for specific domains, longer for the full breadth of knowledge-worker tasks. What is clear is that the trajectory is steep and the destination is transformative.

Key Papers