What is BluePages?

An AI readiness directory for local businesses

Yellow Pages told humans where to find you.
BluePages tells AI where to find you.

AI assistants, search engines, and agents are increasingly how customers find businesses. But most local businesses are invisible to AI -- they have no structured data, no llms.txt, and no signals that help AI understand what they do.

BluePages crawls local business websites and scores them on 10 signals that matter for AI discoverability. The result is a score from 0-100 and a letter grade (A through F).

Scoring Methodology

Each website is evaluated on 10 signals, weighted by importance to AI systems.

15
pts

llms.txt

AI-readable site description file

The llms.txt standard is a machine-readable file (like robots.txt) that helps AI understand your business. It's the single highest-value signal.

5
pts

llms-full.txt

Extended AI documentation

15
pts

Schema.org

Structured data markup (JSON-LD)

JSON-LD structured data (Schema.org) tells search engines and AI exactly what your business is, where it's located, what hours you're open, and what services you provide.

5
pts

Schema Depth

FAQPage, sameAs, Wikidata, field richness

10
pts

Brand Facts

Machine-readable brand identity (/.well-known/brand-facts.json)

10
pts

robots.txt

Crawler access rules

A well-configured robots.txt allows AI crawlers like GPTBot and ClaudeBot to index your site while optionally blocking AI training scrapers.

10
pts

Sitemap XML

XML sitemap for discovery

10
pts

Meta Description

Search result summary text

5
pts

Open Graph

Social sharing metadata

5
pts

HTTPS

Secure connection (SSL/TLS)

5
pts

Page Load

Loading performance (<3s)

5
pts

Headings

Semantic heading structure (H1)

10
pts

Reliability

Contact info, hours, address visibility

Grading Scale

A
Excellent
80-100 pts
AI-ready. Well-structured for AI discovery.
B
Good
60-79 pts
Solid foundation with room to improve.
C
Average
40-59 pts
Basic web presence, missing AI signals.
D
Below Average
20-39 pts
Minimal online presence for AI.
F
Failing
0-19 pts
Essentially invisible to AI systems.

How It Works

1

Seed

Business names and website domains are collected from public directories and local knowledge.

2

Crawl

Our ethical crawler visits each website, fetching the homepage, llms.txt, robots.txt, sitemap, and other key resources. Rate-limited and respectful.

3

Score

10 specialized parsers analyze each resource and produce a signal score. These are summed into a total score (0-100) and letter grade.

4

Report

Results are stored and displayed with detailed signal breakdowns and specific, actionable recommendations for improvement.

Ethical Crawling

  • + Rate-limited (0.5s between requests per domain)
  • + Respects robots.txt directives
  • + Identifies itself clearly (BluePagesBot/1.0 user agent)
  • + Only fetches publicly accessible pages
  • + No personal data collection
  • + No AI training on crawled content

Built by Blue Octopus Technology

Software consulting for businesses that want to be found by AI