RatedWithAI
Accessibility scanner
Industry Analysis
At Axe-con 2026, Meta's engineering team revealed something that changes the accessibility conversation: they trained an AI coding tool on their design system's accessibility patterns and deployed 2,500 fixes across their codebase — with a 90% solve rate. Months of manual work, completed in weeks.
But here's what the headlines miss: Meta's AI didn't replace accessibility testing. It depended on it. The entire pipeline started with automated scanning to identify what was broken. Without that first step, the AI had nothing to fix.
This article breaks down exactly what Meta did, why it matters, and what it means for businesses of every size trying to achieve — and maintain — WCAG compliance.
Jesse Beach, Software Engineering Manager at Meta, took the stage at Axe-con 2026 (Deque Systems' annual digital accessibility conference, February 25-26, 2026) with a presentation titled "Accessible by Default: Scaling Design Systems with AI-Assisted Development."
The core revelation: Meta took their internal AI coding assistant, fed it examples of good accessibility fixes from their design system, and let it systematically apply those patterns across their entire codebase. The results were staggering.
Axe-con 2026 featured 45+ hours of sessions with attendees from approximately 100 countries. But Meta's presentation stood out because it moved the accessibility conversation from "how do we find issues" to "how do we fix them at scale" — and provided concrete data showing it's possible.
Key Quote — Jesse Beach, Meta
"The AI isn't inventing new accessibility patterns. It's applying our patterns — the ones defined in our Design System — consistently across the codebase."
This distinction matters enormously. Meta's AI wasn't guessing at accessibility solutions or applying generic fixes. It was implementing the same patterns Meta's own accessibility team had already defined and validated — just doing it at a speed and scale no human team could match.
Let's break down the data Meta shared:
2,500
Accessibility fixes deployed to production using AI-assisted development
5,000
Additional fixes queued and in review — double the already-deployed volume
~90%
Solve rate for accessibility label issues — 9 out of 10 problems fixed correctly by AI
Weeks
Time to complete what would have taken months of manual engineering work
To put this in perspective: Meta has one of the largest codebases in the world, serving billions of users across Facebook, Instagram, WhatsApp, and Threads. The fact that AI could meaningfully address accessibility at this scale isn't just impressive — it's a proof of concept for the entire industry.
The 90% solve rate is particularly significant. It means Meta's AI wasn't just finding easy wins — it was reliably fixing the vast majority of the accessibility label issues it encountered. The remaining 10% still required human judgment, but reducing the manual workload by 90% is transformative for any engineering organization.
Meta's approach wasn't magic. It was a systematic four-stage pipeline:
Automated Scanning: Find the Issues
Meta used automated accessibility scanning tools (built on standards like axe-core) to systematically identify WCAG violations across their codebase. This created an inventory of every accessibility issue — missing labels, improper ARIA attributes, contrast failures, and more.
Design System Patterns: Define the Solutions
Meta's design system already contained defined, validated accessibility patterns — the "right way" to implement labels, ARIA attributes, focus management, and other accessibility requirements. These patterns were the training data for their AI.
AI Application: Match Issues to Patterns
The AI coding assistant was trained on examples of good fixes — showing it what a correctly implemented accessibility label looks like, how ARIA attributes should be applied, and how the design system's patterns translate to code. Then it applied these patterns to the issues found by scanning.
Human Review: Validate and Deploy
Every AI-generated fix went through code review before deployment. The 90% solve rate means most fixes passed review successfully, but the human validation step ensured no incorrect changes reached production. The 10% that AI couldn't solve were routed to human engineers.
The critical insight: every step in this pipeline starts with automated scanning. Without knowing what's broken, AI has nothing to fix. This is the foundational truth that separates Meta's approach from overlay products that try to bypass the identification step entirely.
Meta's presentation reinforced a fundamental principle that the accessibility industry has been saying for years: you can't fix what you haven't found.
Consider the analogy: you wouldn't let a contractor start renovating your house without first conducting an inspection. The inspection tells you what's broken, where the structural issues are, and what's up to code. Only then can you create a plan to fix things.
Automated accessibility scanning serves the same purpose. Tools that evaluate your website against WCAG 2.2 criteria produce a detailed inventory of violations — each with a specific rule, location, and severity level. This inventory is what makes AI remediation (or manual remediation) possible.
Meta's team used this approach at massive scale. But the principle applies equally to a 10-page small business website:
Why This Matters
Even Meta — with thousands of engineers, billions of dollars, and some of the most advanced AI systems in the world — started with automated scanning. If Meta needs systematic accessibility testing to find their issues before AI can fix them, every business does.
Meta's AI approach might sound superficially similar to what accessibility overlay products claim to do. Both use AI. Both promise to fix accessibility issues. But the differences are fundamental — and the FTC's $1 million fine against accessiBe makes the distinction legally significant.
✅ Meta's AI Approach: Code-Level Fixes
❌ Overlay Approach: Browser-Layer Patches
The contrast is stark. Meta's approach validates the scan-then-fix methodology that the entire accessibility industry recommends. Overlays attempt to skip the hard part — understanding and fixing the actual problems in your code — and the FTC, courts, and disability community have all confirmed that doesn't work.
A crucial factor in Meta's success was their design system. Jesse Beach made this clear: the AI wasn't generating accessibility solutions from scratch. It was applying already-defined, already-validated patterns from Meta's design system consistently across their codebase.
This is important because it explains both why Meta succeeded and why the approach has limits. Meta's design system includes:
Most small and mid-size businesses don't have design systems. They use WordPress themes, Shopify templates, Squarespace sites, or custom-built websites without a formal pattern library. Does that mean Meta's approach is irrelevant to them?
Not at all. Here's what SMBs can take from this:
WCAG itself is the pattern library
Meta had their design system's patterns. Your equivalent is WCAG 2.1/2.2 — a comprehensive set of rules defining what "accessible" looks like for every type of web content. When a scanning tool tells you "image missing alt text" or "form field missing label," the fix is defined by WCAG standards. The "pattern" already exists.
Platform frameworks are design systems
If you use Shopify, Squarespace, or WordPress, the platform's template system is your de facto design system. Accessibility fixes follow platform-specific patterns. An AI coding assistant can learn those patterns just as Meta's learned theirs.
Scanning tools provide the fix instructions
Tools like RatedWithAI don't just tell you what's wrong — they tell you how to fix it. Each violation comes with specific remediation guidance. This is effectively a per-issue "pattern" that you or your developer can apply, and that AI coding assistants can increasingly help implement.
Meta's 90% solve rate is impressive, but it came with an important caveat: the AI was particularly effective at accessibility label issues. These are the most common and most pattern-based WCAG violations. Understanding what AI can and can't handle helps set realistic expectations.
AI Handles Well (Pattern-Based Issues)
AI Struggles With (Context-Dependent Issues)
The good news: the issues AI handles well are also the most common WCAG failures. Missing alt text, missing labels, and improper heading hierarchy account for a huge percentage of the violations that automated scanning tools detect. Fixing these — whether with AI or manually — addresses the majority of accessibility barriers most websites have.
Meta's AI accessibility breakthrough comes at a critical moment. At the same Axe-con 2026 conference, presenters shared alarming data about ADA website lawsuits:
In this environment, the speed of accessibility remediation matters. A business facing an ADA demand letter typically has 30-90 days to demonstrate good-faith remediation efforts. Meta's approach — scan issues, apply AI fixes, deploy quickly — offers a model for rapid response that could help businesses address compliance gaps before they become lawsuits.
Several states are responding to the lawsuit surge with reform legislation. California's SB 84 proposes a 120-day right-to-cure period. But even with cure periods, businesses need tools that can quickly identify and fix accessibility issues — exactly the scan-then-fix approach Meta validated.
Meta's presentation didn't happen in isolation. The entire accessibility industry is moving toward AI-assisted remediation:
Deque's Axe MCP Server
Deque (creators of axe-core, the most widely-used accessibility scanning engine) launched the Axe MCP Server — enabling AI coding assistants to directly run accessibility audits and get structured results. This is the infrastructure for "scan → AI fix" workflows.
axe-core Hits 3 Billion Downloads
The axe-core accessibility engine (which powers most automated accessibility scanning tools, including the one used by RatedWithAI) reached 3 billion npm downloads. This means automated accessibility scanning is already embedded in millions of development workflows worldwide.
AI Coding Assistants Adding Accessibility
GitHub Copilot, Claude, and other AI coding assistants are increasingly capable of generating accessible code when prompted correctly. The gap is narrowing between "identify the issue" and "implement the fix" — exactly what Meta demonstrated at scale.
The direction is clear: the future of accessibility is automated scanning + AI-assisted remediation + human validation. Not overlays. Not manual-only audits. A pipeline that combines the strengths of each approach.
You don't need Meta's engineering team to apply the principles behind their approach. Here's a practical playbook for any business:
Get Your Baseline Scan
Run an automated accessibility scan on your website. Tools like RatedWithAI evaluate your site against WCAG 2.1 and 2.2 criteria and produce a scored report with every violation identified. This is your issue inventory — the equivalent of what Meta's scanning produced.
Time: 5 minutes. Cost: Free scan available.
Prioritize by Impact
Focus on critical and serious violations first. Missing form labels, missing alt text, and missing ARIA attributes block entire user groups from accessing your content. Color contrast issues affect readability. Fix what matters most to actual users.
Time: 15 minutes to review your report and create a priority list.
Fix with AI Assistance (or Your Developer)
Take the specific violations from your scan and fix them. If you use an AI coding assistant (GitHub Copilot, Claude, etc.), you can provide the violation details and ask it to generate the fix — similar to what Meta did at scale. If you work with a developer, the scan report gives them exact locations and remediation guidance.
Time: 1-4 hours for most small business websites.
Re-Scan and Monitor
After fixing issues, scan again to verify the fixes worked and catch any new issues. Accessibility is ongoing — every content update, plugin install, or design change can introduce new violations. Regular scanning catches regressions before they become lawsuits.
Time: 5 minutes per scan. Recommended: monthly minimum.
Document Everything
Keep records of your accessibility scanning history, issues found, fixes applied, and remediation timeline. This demonstrates good-faith compliance efforts — which matters enormously if you ever receive an ADA demand letter. Meta documented their entire process; you should too.
Your RatedWithAI scan history serves as your accessibility documentation trail.
The total investment for a small business: one afternoon to scan, fix, and verify. Compare that to the cost of an ADA website lawsuit (average settlement: $5,000-$50,000+) or the annual cost of overlay products that the FTC has confirmed don't deliver compliance.
Meta's Axe-con presentation is a milestone, but it's the beginning of a trend, not the end. Here's where AI-powered accessibility is heading:
🔄 Real-Time Scanning + Fix Suggestions
Accessibility scanning tools will increasingly offer AI-generated fix suggestions alongside violation reports. Instead of just telling you "this image needs alt text," tools will analyze the image context and suggest appropriate alt text. Some tools already do this; it will become standard.
🤖 IDE-Integrated Accessibility
AI coding assistants in IDEs (VS Code, JetBrains, etc.) will flag accessibility issues as you write code — before it's even deployed. Deque's Axe MCP Server is the infrastructure for this. Expect accessibility linting to become as common as syntax highlighting.
📊 Continuous Monitoring + Auto-Remediation
The Meta model — scan, identify, fix, deploy — will become continuous. Websites will be monitored for accessibility regressions in real-time, with AI proposing (or automatically applying) fixes for pattern-based issues. Human review remains for complex cases.
🌍 International Compliance Convergence
With the European Accessibility Act now enforced and international laws converging on WCAG 2.1 AA, AI tools that can scan and fix against international standards will serve a global market. As Lainey Feingold noted at Axe-con 2026: "We need to start thinking about how the laws of other countries are part of the US digital legal landscape."
The bottom line: AI will make accessibility compliance faster, cheaper, and more reliable. But it won't eliminate the need for automated scanning — it will make it more important. Every AI-powered fix starts with a scan finding the issue.
How did Meta use AI to fix accessibility issues?
Meta trained their AI coding tool on examples of good accessibility fixes from their design system, then applied it systematically across their codebase. The AI didn't invent new accessibility patterns — it applied Meta's existing defined patterns consistently. This achieved a 90% solve rate for accessibility label issues, deploying 2,500 fixes with another 5,000 queued.
What was Meta's accessibility fix solve rate with AI?
Meta achieved approximately 90% solve rate for accessibility label issues using AI-assisted development. This means their AI tool successfully fixed 9 out of 10 accessibility problems it addressed. The key was systematic automated scanning to identify issues first, then applying AI to implement consistent fixes based on design system patterns.
Can AI replace accessibility testing tools?
No. Meta's presentation confirmed that AI fixes depend on automated scanning finding the issues first. AI is a remediation tool, not a detection tool. You still need systematic accessibility scanning (like axe-core or RatedWithAI) to identify WCAG violations before any AI tool can attempt to fix them. Meta's approach was specifically: scan → identify → apply AI fixes using design system patterns.
How does Meta's approach differ from accessibility overlays?
Meta's AI fixes the actual source code, making permanent changes based on design system patterns. Overlays inject JavaScript to modify pages in the browser without changing the underlying code. The FTC fined overlay vendor accessiBe $1 million for deceptive claims. Meta's fixes are permanent and code-level; overlay "fixes" are temporary, client-side patches that 800+ accessibility professionals have documented as ineffective.
Can small businesses use AI for accessibility like Meta?
Small businesses can follow the same principle: use automated scanning to find issues, then fix them systematically. While Meta has thousands of engineers and a mature design system, WCAG standards and platform frameworks (Shopify, WordPress, Squarespace) serve as the "design system" for SMBs. AI coding assistants like GitHub Copilot and Claude can help implement fixes based on scan results.
What accessibility issues can AI fix automatically?
AI is most effective at pattern-based violations: missing alt text, missing ARIA labels, form label associations, heading hierarchy issues, and role attributes. These are also the most common WCAG failures. Complex issues like keyboard navigation logic, dynamic content handling, and custom widget behavior still require human judgment.
Why is automated scanning still necessary with AI?
Meta's AI fix process started with automated scanning to find issues. Without systematic scanning, you don't know what's broken. Accessibility testing tools detect WCAG violations across your entire site, creating the inventory that AI or developers then fix. Scanning is the diagnosis; AI is one possible treatment.
What is Axe-con 2026?
Axe-con 2026 is Deque Systems' annual digital accessibility conference (February 25-26, 2026), featuring 45+ hours of sessions with attendees from approximately 100 countries. Key presentations included Meta's AI accessibility work, lawsuit statistics showing a 40% increase in ADA filings, and the launch of Deque's Axe MCP Server for AI-integrated accessibility testing.
Meta's AI fix pipeline began with automated accessibility scanning. Get your baseline scan with RatedWithAI — identify every WCAG violation on your site, with actionable fix guidance for each issue.
FTC Fined accessiBe $1 Million: Why Overlays Failed
How the FTC's enforcement action exposed overlay product limitations
ADA Lawsuits Up 40%: AI-Powered Pro Se Filings
How AI tools are reshaping accessibility litigation in 2025-2026
Accessibility Widgets: Do They Actually Work?
Analysis of overlay products and why code-level testing wins
Top 10 WCAG Failures (And How to Fix Them)
The most common accessibility issues — the same ones AI can help fix
Published February 26, 2026. Based on data from Meta's presentation at Axe-con 2026 (Deque Systems, February 25-26, 2026). Sources include Deque's Axe-con Day 2 recap, Seyfarth Shaw's ADA lawsuit data, the FTC's consent order against accessiBe, and the Overlay Fact Sheet.