Yutori

Introducing n1: Yutori's Browser-Use Model

By the Yutori team on January 29, 2026

In November, we introduced Navigator, a state-of-the-art web agent that autonomously navigates websites to complete everyday tasks. At the heart of Navigator is n1, our browser-use model.

Today, we're making n1 available in our API — giving developers access to the best accuracy, pricing, and latency for browser automation among computer-use models.

n1 is optimized for web tasks through extensive mid-training and supervised fine-tuning—and, uniquely, reinforcement learning on real websites rather than just simulated environments—resulting in strong performance across a wide range of real-world web automation tasks.

Table of Contents

State-of-the-Art in Browser-Use

n1 achieves state-of-the-art performance on browser-use benchmarks, outperforming all other computer-use models on both Navi-Bench1 and Online-Mind2Web2. It reaches 83.5% accuracy on Navi-Bench, which comprises of 100 tasks across five real-world websites—Apartments, Craigslist, OpenTable, Resy, and Google Flights—and 78.7% on Online-Mind2Web, a benchmark evaluating real-world web agent performance with 300 tasks spanning 136 popular sites.

Accuracy comparison

Figure 1. n1 achieves state-of-the-art accuracy on browser-use benchmarks.

Browser Automation at Scale

n1 offers the most cost-effective pricing among computer-use models — up to 8x cheaper than alternatives while delivering better accuracy.

ModelInput (per 1M tokens)Output (per 1M tokens)
Claude 4.5 Opus$5.00$25.00
Claude 4.5 Sonnet$3.00$15.00
OpenAI Computer-Use$3.00$12.00
Gemini 2.5 Computer-Use$1.25$10.00
Yutori n1$0.75$3.00

Table 1. n1 is the most cost-effective computer-use model.

At $0.75 per million input tokens and $3.00 per million output tokens, n1 makes browser automation accessible at scale.

Speed matters too — every second of model latency adds up across the dozens of steps required to complete a web task. n1 delivers the fastest per-step latency among computer-use models.

Latency comparison

Figure 2. n1 achieves the lowest per-step latency among computer-use models.

Measured on Navi-Bench under identical conditions with a maximum of 75 steps per task, n1 achieves an average per-step latency of 3.6 seconds — 2.2x faster than Claude 4.5 Opus and 1.7x faster than Gemini 2.5 Computer-Use. Combined with its higher accuracy, n1 completes tasks both faster and more reliably.

Get Started

The n1 API provides direct access to Yutori's browser-use model so you can build your own browser automation pipelines.

We're excited to see what you build with it.

Addendum — March 9, 2026

Since this post was published in January 2026, several new frontier models with computer use capabilities have been released. An updated comparison of n1 with other frontier models on Navi-Bench is below.

Updated benchmark comparison: Yutori n1 vs. Claude 4.6 Opus, GPT-5.4, Nova Act Latest, Gemini 2.5 Pro

n1 achieves 91% on Navi-Bench v1, outperforming Claude 4.6 Opus (88%), GPT-5.4 (78%), Nova Act Latest (73%), and Gemini 2.5 Pro (53%).