What is BluePages?
An AI readiness directory for local businesses
Yellow Pages told humans where to find you.
BluePages tells AI where to find you.
AI assistants, search engines, and agents are increasingly how customers find businesses. But most local businesses are invisible to AI -- they have no structured data, no llms.txt, and no signals that help AI understand what they do.
BluePages crawls local business websites and scores them on 10 signals that matter for AI discoverability. The result is a score from 0-100 and a letter grade (A through F).
Scoring Methodology
Each website is evaluated on 10 signals, weighted by importance to AI systems.
llms.txt
AI-readable site description file
The llms.txt standard is a machine-readable file (like robots.txt) that helps AI understand your business. It's the single highest-value signal.
llms-full.txt
Extended AI documentation
Schema.org
Structured data markup (JSON-LD)
JSON-LD structured data (Schema.org) tells search engines and AI exactly what your business is, where it's located, what hours you're open, and what services you provide.
Schema Depth
FAQPage, sameAs, Wikidata, field richness
Brand Facts
Machine-readable brand identity (/.well-known/brand-facts.json)
robots.txt
Crawler access rules
A well-configured robots.txt allows AI crawlers like GPTBot and ClaudeBot to index your site while optionally blocking AI training scrapers.
Sitemap XML
XML sitemap for discovery
Meta Description
Search result summary text
Open Graph
Social sharing metadata
HTTPS
Secure connection (SSL/TLS)
Page Load
Loading performance (<3s)
Headings
Semantic heading structure (H1)
Reliability
Contact info, hours, address visibility
Grading Scale
How It Works
Seed
Business names and website domains are collected from public directories and local knowledge.
Crawl
Our ethical crawler visits each website, fetching the homepage, llms.txt, robots.txt, sitemap, and other key resources. Rate-limited and respectful.
Score
10 specialized parsers analyze each resource and produce a signal score. These are summed into a total score (0-100) and letter grade.
Report
Results are stored and displayed with detailed signal breakdowns and specific, actionable recommendations for improvement.
Ethical Crawling
- + Rate-limited (0.5s between requests per domain)
- + Respects robots.txt directives
- + Identifies itself clearly (BluePagesBot/1.0 user agent)
- + Only fetches publicly accessible pages
- + No personal data collection
- + No AI training on crawled content
Built by Blue Octopus Technology
Software consulting for businesses that want to be found by AI