216digital.
Web Accessibility

ADA Risk Mitigation
Prevent and Respond to ADA Lawsuits


WCAG & Section 508
Conform with Local and International Requirements


a11y.Radar
Ongoing Monitoring and Maintenance


Consultation & Training

Is Your Website Vulnerable to Frivolous Lawsuits?
Get a Free Web Accessibility Audit to Learn Where You Stand
Find Out Today!

Web Design & Development

Marketing

PPC Management
Google & Social Media Ads


Professional SEO
Increase Organic Search Strength

Interested in Marketing?
Speak to an Expert about marketing opportunities for your brand to cultivate support and growth online.
Contact Us

About

Blog

Contact Us
  • Google Lighthouse 100? Automated Testing Still Falls Short

    A 100 score from automated testing feels good. Your dashboard turns green. The report says you passed every check. It looks complete. On paper, everything looks compliant. It is the kind of result that gets shared in Slack, checked off in a ticket, and filed away as “resolved.”

    But that score does not mean people can use your site.

    Most automated testing tools are helpful. They catch real barriers and save time. The problem is what they cannot measure. In practice, automated checks tend to cover only a slice of accessibility—roughly 30 percent—because they are limited to what can be evaluated programmatically. The remaining work involves interaction, context, and human judgment. As standards evolve and legal expectations keep tightening, you have to be honest about whether the metrics you rely on still tell the truth—for your business and for your users.

    Here is where automated testing leaves gaps that can turn into barriers for users and real exposure for your team.

    What Google Lighthouse Checks (and What It Doesn’t)

    Google Lighthouse is an open-source tool that audits a web page and reports on several quality signals—most commonly performance, SEO, and accessibility. It is widely used because it is easy to run, easy to share, and it produces a single score that feels objective.

    As an accessibility tool, though, Lighthouse is limited.

    How Lighthouse Calculates Your Accessibility Score

    Like all automated accessibility tests, Lighthouse can miss barriers that affect users (false negatives). It can also flag patterns that are not actually barriers in context (false positives). That is not a knock on Lighthouse. It is a reminder that the tool is only as reliable as what can be measured from code alone.

    When Google Lighthouse scores accessibility, it runs a set of pass-or-fail checks and assigns weights to each one. Your final score is a weighted average, which means some failures carry much more impact than others.

    A clear example is severe ARIA misuse. Putting aria-hidden=”true” on the body element is heavily weighted because it removes page content from the accessibility tree. When that happens, a screen reader user may not be able to perceive the page at all. Lighthouse penalizes this hard, and it should.

    Where Lighthouse Scores Stop and User Experience Starts

    Notice what that scoring model reinforces. Lighthouse is evaluating machine-detectable code patterns. It is not validating the full user experience—whether a flow makes sense, whether focus order matches intent, whether labels hold up in context, or whether an interaction is usable with assistive technology.

    Google’s own guidance is clear: only a subset of accessibility issues can be detected automatically, and manual testing is encouraged. That is not a minor disclaimer. It defines the boundary of what the score means.

    If you use the score as a proxy for accessibility, you are using it outside its intended purpose.

    How Automated Accessibility Testing Evaluates Your Site

    Automated testing is built for consistency and repeatability. It excels at spotting structural issues that follow well-defined rules. In practice, that usually means it flags things like:

    • Missing alt attributes on images
    • Low color contrast ratios based on numeric values
    • Form fields with no programmatic label
    • Empty buttons or links with no text alternative
    • Missing language attributes on the html element
    • Obvious ARIA errors that break the accessibility tree

    Why “Pass” Does Not Mean “Helpful”

    Color contrast is another great example. A tool can measure foreground and background values, calculate the ratio, and report whether it meets the Web Content Accessibility Guidelines (WCAG) requirements.  For example,  SC 1.4.3 Contrast Minimum requires a 4.5:1 ratio for normal text. That matters for users with low vision and color vision differences.

    Contrast is another place where automated tools fall short. They can measure color contrast ratios, but they cannot evaluate readability in context. They cannot tell whether your font size and weight work well with that contrast choice, whether visual styling creates confusing groupings in navigation, or whether users can scan the page and understand it easily.

    That pattern shows up across most automated checks. Tools confirm that something is present in code; they do not confirm how well it works in context. The scan focuses on individual elements rather than the interactions between them, on static states rather than the workflows people have to move through.

    That coverage is useful, but it is thin. It reaches only a narrow slice of accessibility. The rest sits in the gap that automation cannot reach.

    The Limits of Automated Accessibility Testing

    The issues that stop people usually sit outside what automation can prove. They show up in behavior and context, not in markup alone. That is how a site can “pass” and still fail users.

    Keyboard Navigation and Focus Visibility

    A tool can confirm that an element is focusable and that a label exists. It cannot verify what using the page with a keyboard actually feels like.

    You still need to know:

    • All interactive elements can be reached by pressing Tab.
    • Focus indicators stay visible and easy to follow.
    • Complex widgets like date pickers, autocomplete fields, and modal dialogs work correctly with keyboard-only navigation.

    Those answers do not come from scanning markup. Keyboard testing requires human interaction and someone who understands how keyboard users move through web pages.

    Screen Reader Output and Meaning

    Automation can confirm that text alternatives and labels are present. It cannot confirm what a screen reader announces, in what order, and whether that output is useful in context.

    This is where “passes” hide confusion. A tool cannot tell whether the alt text says “image123” or “Yum yum” for a product photo. Both satisfy the requirement. Only one helps a user.

    A label can exist but be announced in a way that does not match the visible interface. Alt text can be technically present and still add noise instead of clarity. Errors can appear visually and never be announced at all. The code can look correct while the experience still breaks.

    Screen readers also differ. NVDA, JAWS, VoiceOver on macOS, VoiceOver on iOS, and TalkBack all interpret markup in slightly different ways. Automated testing does not account for those differences. It assumes a static model of accessibility, while users operate in dynamic environments.

    Understanding, Language, and Cognitive Load

    Tools do not measure whether your interface is understandable. They do not know when instructions are dense. They do not notice when terminology shifts from one step to the next or when navigation labels do not match what the page is actually doing.

    Key questions stay unanswered:

    • When someone scans the page, can they tell what to do next, or is it buried in jargon and extra complexity?
    • If they make a mistake, do they have a clear way to recover, or are they forced to start over?
    • As users change text size or zoom, does the layout hold together, or does it fall apart?
    • For people with cognitive disabilities, do your interface patterns feel consistent and understandable?

    Why Manual Accessibility Testing Still Sets the Standard

    Automated checks can tell you whether patterns exist in your code. They cannot tell you whether those patterns work when a person tries to complete a task with assistive technology.

    Manual testing is where you find the failures that stay invisible in a report. It is also where you verify that “accessible” holds up across the tools people actually use.

    In audits, we test with NVDA and JAWS on Windows, VoiceOver on macOS, VoiceOver on iOS, and TalkBack on Android. These tools do not behave the same way, even when the markup looks clean. We also test keyboard-only navigation, voice control, and zoom. Each component is evaluated against a checklist of over 260 items for full WCAG 2.2 coverage.

    This is often where perfect automated scores stop feeling meaningful. Forms can look correct on paper, yet labels that technically announce still fail voice control because the spoken target is unclear. Mobile layouts may meet target size rules, while the placement makes taps unreliable. Dynamic regions can update with no announcement at all, so screen reader users lose the thread. Navigation might be valid in markup and still be hard to use when landmarks are noisy, vague, or missing where people expect them.

    Manual testing connects those details back to the actual job a user is trying to do.

    The Cost of Relying Only on Automated Accessibility Tests

    Teams that stop at automated testing tend to learn about the remaining issues the hard way. A user hits a blocker, reports it, and now the problem is public. That carries reputational risk, can become legal risk, and often lands on your team as an urgent disruption instead of planned work.

    It is also avoidable.

    The cost curve is clear. A full audit that includes manual testing is typically cheaper than defending a claim, rebuilding components that shipped without assistive technology constraints in mind, or patching accessibility after customers have already churned. Teams sometimes rebuild the same feature more than once because the first pass did not account for how screen readers announce changes or how voice control targets labels.

    Automated testing is a starting point. A perfect score is baseline hygiene worth maintaining. It is necessary, and still nowhere near enough.

    Combining Automated and Manual Accessibility Testing

    Lighthouse scores and perfect automated testing results create false confidence. Genuine accessibility depends on both automated and manual testing. Automated checks belong in your everyday development pipeline, catching structural issues early and guarding against regressions. But don’t stop there. Manual testing with assistive technology then fills in the rest of the picture, showing whether people can actually complete tasks.

    A better approach is to treat automation as the first pass, but manual testing as the standard of proof. Run automated tests early and often, then make space for keyboard checks, screen reader passes, and voice control scenarios before you sign off on a release.

    If you want help putting that kind of testing strategy in place, 216digital can work alongside your team. Schedule an ADA Strategy Briefing with our experts to review your current workflow, understand your risk, and design an accessibility testing plan that pairs automated coverage with focused manual testing where it counts most.

    Greg McNeil

    February 17, 2026
    Testing & Remediation
    Accessibility, Accessibility testing, automated scans, automated testing, How-to, Website Accessibility
  • Consultants or Automated Platforms: What’s Right for You?

    Consultants or Automated Platforms: What’s Right for You?

    Making a website accessible isn’t always a straight path. There are sleek platforms that can scan every page in minutes, and seasoned consultants who can spot problems no algorithm would catch. Each offers value—but in very different ways.

    The challenge isn’t choosing which one is “better.” It’s knowing when to rely on quick automated checks and when your site needs the nuance only human expertise can provide. Getting that balance right can turn accessibility from a one-time project into a lasting part of how your site works.

    What Automated Platforms Do Well

    Automated accessibility platforms are essentially software tools that scan your site for compliance issues. Think of them as always-on monitors quietly running in the background. They can:

    • Scan your site regularly to flag new problems as they appear
    • Track your accessibility performance over time
    • Send alerts when something changes

    They’re fast, efficient, and cost-effective. Within minutes, they can show you where your site stands and give you a benchmark to measure progress. For many organizations, this kind of real-time insight is reassuring—especially after an initial round of accessibility improvements. Automated tools can help ensure new issues don’t creep in unnoticed.

    But while they’re powerful, they’re not perfect. automated platforms can catch many surface-level problems, like missing alt text or low-contrast color pairings. What they can’t do is understand the human experience of using your site. They don’t know if your navigation makes sense to someone using a screen reader, or whether form instructions are clear enough to avoid confusion. For those nuanced judgment calls, human expertise is essential.

    The Role of Accessibility Consultants

    Accessibility consultants offer something no machine can: experience, context, and human perspective. They don’t just tell you what’s broken—they explain why it matters and how to fix it in a way that fits your real-world workflows.

    A good consultant will:

    • Conduct thorough audits that go far beyond automated scans
    • Identify root causes, not just symptoms
    • Guide your team through remediation with practical, achievable steps
    • Provide training so you can build accessibility into your process in the future

    Consultants also bring critical legal and standards knowledge to the table. They stay on top of evolving regulations and know how guidelines like WCAG apply to your specific industry or audience. That insight can help you minimize legal risk while also creating a more welcoming experience for users with disabilities.

    In other words, they look at the big picture—something an automated tool can’t do.

    Why Consultant-Led Remediation Should Come First

    One of the most common missteps organizations make is starting with an automated platform before any human-led remediation. On paper, it seems logical: run a scan, see what’s wrong, and start fixing. But in practice, this often backfires.

    Automated scans can return long lists of issues—some legitimate, some false positives, and many without clear instructions for resolution. Without expert guidance, it’s easy to spend hours chasing the wrong problems or applying “fixes” that don’t actually help real users.

    Consultant-led remediation flips this process around for better results. Instead of reacting to a flood of automated alerts, you get a clear, prioritized plan from someone who understands both the technical and human aspects of accessibility. They focus on foundational issues first, ensuring the fixes are meaningful and sustainable.

    Once that groundwork is in place, automated platforms become incredibly useful. They act like a safety net, helping you maintain the progress you’ve made.

    Think of it like building a house: you wouldn’t install a security system before the walls are up. The system is valuable—but only once the structure is solid.

    When Automated Platforms Make Sense

    After you’ve remediated your site with the help of a consultant, automated platforms can become a valuable part of your ongoing strategy. Websites are living, changing systems. Every content update, plugin installation, or design tweak carries the potential to introduce new accessibility barriers.

    An automated platform helps you stay ahead of those problems by catching them early. They’re instrumental when:

    • You publish new content frequently
    • You’re rolling out design changes or new features
    • You want to show good-faith efforts with regular monitoring reports
    • You need an affordable way to keep watch between consultant reviews

    Used this way, automated tools act as a maintenance system. They can’t replace human testing, but they can help keep your site healthier between more in-depth reviews.

    How to Decide What’s Right for You

    Choosing between consultants and automated platforms becomes much easier once you know where you are in the accessibility journey.

    Starting from scratch? Bring in a consultant first. Their guidance will help you build a solid foundation and avoid the guesswork that leads to costly mistakes.

    Working through remediation? Stay the course with consultant-led support. Automated scans can muddy the waters here, flagging noise instead of what really matters.

    Site already in good shape? That’s the moment to add an automated platform. Let it keep an eye on new changes while consultants check in periodically for deeper reviews.

    For many organizations, the most effective approach is a blend of both—just in the correct order. Human expertise lays the groundwork. automated platforms help you maintain it.

    The Long-Term Payoff

    Web accessibility isn’t a box you check off once—it’s a long-term commitment. But it’s one that pays off in measurable ways: stronger legal compliance, broader audience reach, improved usability, and greater trust from customers and clients.

    Consultants give you the strategy, expertise, and training to start on solid ground. automated platforms give you the ongoing monitoring to protect that investment.

    When used together, they create a sustainable system. You get the precision of expert audits and the efficiency of automated monitoring. This balance reduces risk, improves user experience, and keeps you aligned with evolving standards as they change over time.

    Human First, Automation Second

    Choosing between consultants and automated platforms isn’t really about picking one over the other—it’s about knowing how they fit together. Automated tools can keep watch over the details, but it takes human expertise to build the kind of foundation that lasts.

    Start by getting that solid groundwork in place with a consultant-led audit and remediation. Once your site is truly accessible, an automated platform can help you keep it that way—quietly catching issues before they become problems.

    If you’re ready to map out what that first step should look like, schedule an ADA briefing with 216digital. It’s a chance to talk through where your site stands, what’s needed to meet compliance, and how to build a long-term plan that keeps accessibility on track.

    Greg McNeil

    September 17, 2025
    Testing & Remediation
    Accessibility, Accessibility Remediation, Accessibility testing, automated scans, automated testing, consultants, Web Accessibility, Web Accessibility Remediation, Website Accessibility
  • Automated Scans: The Good, the Bad, and the Gaps

    Creating a truly inclusive online experience means making sure everyone can comfortably navigate and enjoy your website. If your business aims to meet accessibility guidelines like WCAG and comply with laws such as the Americans with Disabilities Act (ADA), you’ve probably considered automated accessibility scans as a helpful first step. Popular tools like WAVE, PowerMapper, and Google Lighthouse can swiftly highlight accessibility issues, making them a practical starting point.

    But here’s the thing—automated scans aren’t magic. They typically catch only about 20-30% of accessibility problems covered by WCAG 2.1 AA, mostly basic mistakes. Many important issues need a human touch—manual testing and real judgment—to make sure your website truly serves everyone. This article breaks down what automated scans do well, where they fall short, and tackles some common misconceptions, helping you build a complete, effective approach to digital accessibility.

    The Good: Why Automated Accessibility Scans Are Helpful

    Automated accessibility scans offer a quick and practical way to uncover accessibility problems. Here’s why they’re so beneficial:

    1. Speed and Efficiency Automated tools can quickly scan your entire website within minutes, analyzing hundreds or even thousands of pages. They’re perfect for frequent or large-scale audits without overwhelming your resources.

    2. Consistent Results Unlike manual tests, automated scans apply the same criteria every time, ensuring consistent results across your entire site.

    3. Budget-Friendly Initial Assessment For businesses working with tight budgets, automated scans offer an affordable way to spot common issues, helping you take immediate action without the hefty costs of comprehensive manual audits.

    4. Quickly Spots Easy-to-Fix Issues Automated scans excel at identifying straightforward problems such as missing alternative text for images, insufficient color contrast, empty form labels, and poorly labeled buttons.

    5. Fits Seamlessly into Development Workflows Modern accessibility tools integrate directly with your existing development processes, allowing teams to catch and fix issues early on—saving time and reducing future remediation costs.

    The Bad: Limitations of Automated Scans

    To avoid a false sense of security, organizations need to understand the limitations of automated scans:

    1. Can’t Identify Contextual Issues Automated tools aren’t great at judging context. For example:

    • Alternative Text: They detect missing alt text but can’t determine if descriptions are actually helpful.
    • Keyboard Navigation: Tools may confirm elements are focusable but can’t judge ease of use.
    • Form Usability: Automated scans notice if labels exist, but can’t assess how easy forms are to use.

    2. Struggles with Dynamic Content Automated tools often miss dynamic and interactive content, such as:

    • ARIA implementations
    • Live region updates
    • Changes in modal dialogs or dropdowns

    3. False Positives and False Negatives Automated scans sometimes flag issues that aren’t real barriers (false positives) or overlook genuine problems (false negatives), potentially causing unnecessary work or leaving critical issues unaddressed.

    4. Not a Substitute for Real User Testing Automated scans can’t replicate the experience of actual users who rely on assistive technology. Only manual testing and real user feedback can fully validate accessibility.

    5. Risk of Legal Issues Relying solely on automated scans won’t guarantee full legal compliance. Courts consistently rule that automated tools alone aren’t enough, potentially exposing organizations to legal action.

    The Gaps: Bridging Automation with Human Insight

    To build a truly accessible website, combine automated testing with manual evaluation:

    1. Mix Automated and Manual Testing A comprehensive audit should include both automated and manual methods. Automated tools give a broad initial review, while manual testing covers the usability and context these tools miss.

    2. Perform Real-World Testing with Screen Readers and Keyboards Manual tests with screen readers and keyboard navigation ensure your website is practically accessible in real-world use.

    3. Get Expert Audits Professional accessibility audits blend automated tools, manual testing, and real-user scenarios, uncovering hidden issues automated scans might overlook.

    4. Continuously Monitor Accessibility Accessibility isn’t a one-time fix. Regular monitoring with tools like a11y.Radar helps maintain compliance even as your website evolves.

    5. Educate Your Team Training your developers, designers, and content creators on accessibility best practices minimizes future issues. Automated tools should guide—not replace—human judgment.

    Conclusion: Finding the Right Balance

    Automated accessibility scans are fantastic for quickly identifying common issues but aren’t enough on their own. Recognizing their limitations highlights why manual testing, expert reviews, and ongoing monitoring are vital. Businesses committed to digital accessibility should adopt a balanced approach, combining automated tools with human insights to ensure truly inclusive experiences. If you’re looking for guidance or support in building this balanced approach, the experts at 216digital can help you navigate your accessibility journey. Ultimately, accessibility is about genuinely understanding and meeting real users’ needs, going beyond mere compliance.

    Greg McNeil

    March 17, 2025
    Testing & Remediation
    Accessibility, automated scans, automated testing, manual audit, Manual Testing, Web Accessibility

Find Out if Your Website is WCAG & ADA Compliant







    By submitting this form, you consent to follow-up from 216 Digital by call, email, or text regarding your inquiry. Msg & data rates may apply. Reply STOP to opt out or HELP for help.

    216digital Logo

    Our team is full of professionals in Web Accessibility Remediation, eCommerce Design & Development, and Marketing – ready to help you reach your goals and thrive in a competitive marketplace. 

    216 Digital, Inc. BBB Business Review

    Get in Touch

    2208 E Enterprise Pkwy
    Twinsburg, OH 44087
    216.505.4400
    info@216digital.com

    Support

    Support Desk
    Acceptable Use Policy
    Accessibility Policy
    Privacy Policy

    Web Accessibility

    Settlement & Risk Mitigation
    WCAG 2.1/2.2 AA Compliance
    Monitoring Service by a11y.Radar

    Development & Marketing

    eCommerce Development
    PPC Marketing
    Professional SEO

    About

    About Us
    Contact

    Copyright © 2026 216digital. All Rights Reserved.