Skip to main content

Google Lighthouse Has A New Agentic Browsing Category

· 4 min read
Matt Zeunert

Lighthouse 13.3 came out this week, and includes a new category called "Agentic Browsing". Let's take a look at the new audits included in this category.

What is the agentic browsing category?

While Lighthouse has always primarily been a performance testing tool, it also includes other website quality checks looking at website accessibility, SEO, and other best practices.

The new category runs several checks to see how AI agents can interact with your website.

Lighthouse scores with new Agentic Browsing category

The agentic browsing category in the Lighthouse report runs several audits checking:

  • That the website accessibility tree is well-formed
  • Whether content shifts after rendering
  • Correct WebMCP implementation
  • llms.txt guidelines

Agentic browsing report category

Two important caveats

The agentic browsing category is currently still marked as "under development". There's still a lot of uncertainty about what websites need to focus on in order for AI agents to use them effectively.

Also, you won't fail this category just because you haven't integrated new AI features. example.com gets a green 2/2 score!

How to check your agentic browsing score

You can either test your website using an online tool or run Lighthouse on your own device.

Online testing tools

To check your own website you can use the free DebugBear website quality checker:

  1. Enter your website URL and run a test
  2. Open the Lighthouse tab in the sidebar

PageSpeed Insights and Chrome DevTools still use an older version of Lighthouse, but this should change in the coming months.

Finding the agentic browsing score in the DebugBear Lighthouse tab

Locally on your own computer

To test your website locally from your own computer you can use the Lighthouse command line tool:

  1. Install the latest version of Lighthouse: npm install -g lighthouse@latest
  2. Run a test, e.g. lighthouse --view https://web.dev/

Building agent-friendly websites

Google recently published a guide on how to build agent-friendly websites and there is some overlap between the new Lighthouse checks and the recommendations in the article.

One key take-away from the guide is that AI agents interact with website content in different formats:

  1. Screenshots
  2. HTML code
  3. The accessibility tree

What checks does the agentic browsing category actually run?

What AI agent issues can Lighthouse actually detect? The new category combines existing report data with several new audits.

Accessibility tree is not well-formed

The accessibility tree is a simplified representation of the page structure that is used by accessibility tools and AI agents. This audit data is based on the existing accessibility checks run by Lighthouse.

AI agents can process the accessibility tree more efficiently compared to the full HTML code or a website screenshot.

Invalid accessibility tree

WebMCP validation

WebMCP is an API that lets websites provide specific commands that agents can use to interact with them. Unlike general MCP servers, WebMCP is built for front-end interactions in the context of the visitor's browser session.

There's a declarative version of the API based on annotating HTML forms on the page. The WebMCP audit checks whether forms are annotated and if the annotations match the expected schema.

WebMCP form annotation validation result

Lighthouse also surfaces all WebMCP tools that have been registered on the page, whether they were added through the declarative API or programmatically with navigator.modelContext.registerTool.

WebMCP tools in the WebMCP maze demo game

llms.txt does not follow recommendations

llms.txt files are a proposed way for websites to provide information to AI agents and crawlers. However, it is not widely used by AI tools currently.

The new Lighthouse audit checks if the website provides an llms.txt file. If it does, it checks if the file is missing an H1 header, is too short, or doesn't contain any links.

llms.txt validation result in Lighthouse

Detecting layout shifts

It's hard to use a website if content shifts around unexpectedly. In 2020, Google introduced the Cumulative Layout Shift (CLS) metric to measure and report these shifts.

Why does this matter to AI agents? Google's guide explains:

Ensure stable layout. Agents that take screenshots will likely be confused if your website layout is constantly shifting

Ultimately this audit just provides existing CLS score reporting in a different category. An AI agent may interact with your website more quickly than a human visitor, so shifts that occur as content is loading may be more disruptive to an agent.

CLS score report

Test your website and keep it optimized

Whether you're optimizing your website for performance, SEO, or AI agents, DebugBear provides in-depth reporting on website quality and visitor experience. Identify key issues and get alerted when there's a regression.

Start your free trial today.

DebugBear audits dashboard

Illustration of website monitoringIllustration of website monitoring

Monitor Page Speed & Core Web Vitals

DebugBear monitoring includes:

  • In-depth Page Speed Reports
  • Automated Recommendations
  • Real User Analytics Data

Get a monthly email with page speed tips