Skip to content

Intelligent Brand Index (IBI)

The Intelligent Brand Index (IBI) is a comprehensive scoring system designed by Advanced Analytica to evaluate how effectively a brand's website communicates with AI systems and large language models (LLMs),search engines, and accessibility tools. It combines structured data quality, discoverability, AI-readiness, and accessibility into a single score out of 100.

The IBI system contains the accessibility scoring component of the IBI, powered by Lighthouse CLI and axe-core.


Purpose of IBI

As AI-generated answers, voice assistants, and search overviews increasingly depend on machine-readable content, brands must ensure they are discoverable and interpretable by machines. IBI provides a benchmark for:

  • AI discoverability via structured data (e.g., JSON-LD + Schema.org)
  • Governance and crawlability (robots.txt, sitemaps, AI guidance files)
  • AI-readiness (semantic markup, content freshness, media accessibility)
  • Accessibility based on WCAG 2.1 AA conformance

IBI Scoring Framework

The full IBI score is composed of four main categories:

Category Max Points
A. Structured-Data Quality 50
B. Discoverability & Governance 30
C. AI-Friendliness & Media 20
D. Accessibility (WCAG 2.1 AA) 15
Total Raw Points 115

The final score is normalised to a 0–100 scale.


Expanded IBI Scoring Categories**

The IBI evaluates a website’s machine-readability and digital governance using a multi-dimensional score. It ensures your site is visually appealing, machine-readable and LLM-interpretable.


A. Structured Data Quality

Definition: Measures the presence, accuracy, and coverage of semantic metadata embedded on the site using formats like JSON-LD, Microdata, or RDFa, following Schema.org vocabularies.

Why It Matters: AI systems (including Google, ChatGPT, and voice assistants) increasingly depend on structured data to understand entities, products, events, and questions and asnwers (QAs). Well-formed data increases search visibility and enables features like rich snippets and knowledge graph inclusion.

Scoring Components:

Sub-Metric Description Pts
A-1. Schema.org Type Coverage Presence of types like Organization, Brand, Product, Event, FAQ, etc. 15
A-2. Field Completeness Degree to which required/recommended fields are populated (e.g., name, url) 10
A-3. Nested Entities & Relationships Use of nested schema (e.g., Product has Brand, Offer, Review) 10
A-4. JSON-LD Syntax Validity Checks for valid, parseable JSON-LD syntax 5
A-5. Duplicate / Conflicting Markup Penalizes repeated or contradictory structured data entries 5
A-6. Use of Custom Extensions Bonus for custom vocab extensions where applicable 5

B. Discoverability & Governance

Definition: Audits whether a site is technically configured for bots and agents to discover, crawl, and interpret the content — while also supporting newer AI agents and systems.

Why It Matters: Good governance helps ensure that crawlers (from search engines to AI scrapers) can properly index and access your content — improving visibility in SERPs, chat and voice search.

Scoring Components:

Sub-Metric Description Pts
B-1. Valid robots.txt file Should exist, be reachable, and contain appropriate disallow/allow directives 5
B-2. Sitemap Presence sitemap.xml is present and submitted 5
B-3. AI Guidance (ai.txt) Support for ai.txt or llms.txt to guide AI agents (e.g., OpenAI, Anthropic) 5
B-4. Canonical Tag Usage Correct canonical tag on pages to prevent duplicate content confusion 5
B-5. Page Indexability Page should not be noindex unless intended 5
B-6. Crawl Budget Optimization Avoid excessive redirects or dead links 5

C. AI-Friendliness & Media Semantics

Definition: Evaluates how machine-readable your content is, especially rich media like images, video, and interactive widgets.

Why It Matters: AI systems like GPT-4, Gemini, and Perplexity attempt to interpret content across modalities. Media needs metadata, captions, and context to be useful to machine reasoning.

Scoring Components:

Sub-Metric Description Pts
C-1. Descriptive Alt Text Contextually accurate alt attributes on visual content 5
C-2. Video Metadata & Captions Titles, transcripts, and subtitles present (especially for hero content) 5
C-3. Headless & Semantically Tagged Content Uses <article>, <section>, <main>, etc. to define intent 5
C-4. Freshness & Update Signals lastModified, updated timestamps, or freshness indicators for content 5

D. Accessibility

[Already well defined in your original message — retained with slight expansion below]

Definition: Assesses the ability of people with impairments (e.g., vision impairment, limited mobility) to navigate and consume the content, per WCAG 2.1 AA.

Why It Matters: Accessibility is not just ethical and often legally required — it improves semantic clarity and boosts SEO/AI-readability for all users and agents.

Scoring Components:

ID Metric Description Max Points
D-1 Alt text on images All meaningful images have non-empty alt attributes 3
D-2 Semantic heading order Headings follow logical h1 > h2 > h3 structure 2
D-3 Colour contrast (4.5:1 or better) Meets WCAG contrast standards for legibility 3
D-4 Keyboard navigation All interactive elements are accessible without a mouse 2
D-5 ARIA landmarks and skip links Uses roles like main, navigation, complementary 2
D-6 Critical accessibility violations < 5 Lighthouse/axe-core reports fewer than 5 critical issues 3

Final Score Computation

  • Each raw category is summed (max 115 pts).
  • Then normalised to a score out of 100:
ibi_score = round((raw_score / 115) * 100, 1)

How It Works

1. Run Lighthouse

python ibi_accessibility_scoring.py https://example.com output_dir/

This uses Lighthouse CLI (Node.js required) to generate a JSON report and extract accessibility data.

2. Extract Metrics

The script parses Lighthouse's audits and evaluates the presence and quality of WCAG-relevant elements.

3. Score Calculation

Each sub-metric is weighted according to the IBI framework and produces a total score out of 15.


Dependencies

  • Node.js with Lighthouse CLI (npm i -g lighthouse)
  • Python 3.8+
  • Chrome installed and available to CLI

Output Example

Accessibility Sub-Metrics:
  D1_alt_text: true
  D2_headings_order: true
  D3_contrast: false
  D4_keyboard_nav: true
  D5_landmarks: true
  D6_critical_violations: true

Total Accessibility Score: 13 / 15

Notes

  • This module evaluates accessibility only. For the full IBI score (structured data, governance, and AI-friendliness), integrate this into the broader pipeline.
  • Only public URLs are supported unless Lighthouse is run with custom Chrome sessions.