Our approach
Why automated accessibility testing isn't enough
Automated scans check code. We check whether real people can actually use your website.
What automation misses
Automated tools can tell you if an image has alt text. They can't tell you if that alt text makes sense to a blind person.
They can check if a form has labels. They can't check if someone using a screen reader can actually complete it.
That's why we test with the people accessibility is supposed to help.
Guiding principles
Our work is shaped by testing from disabled users and the impact on real journeys.
Real user impact
Automated scans catch a portion of issues. Manual feedback from disabled user testers reveals the barriers that block journeys.
Lived experience
Our disabled user testers bring daily experience with screen readers, magnification, and keyboard-only workflows.
Mission-led delivery
MyVision has a 150-year heritage supporting blind and visually impaired people, and that insight guides every audit.
70%
Barriers missed by automation alone
£274B
UK disabled spending power
71%
Leave sites with accessibility barriers
Triple impact
Every audit does three things.
Your clients
Get websites that work for disabled people.
Disabled testers
Get paid for expertise that automation can't replicate.
MyVision
Gets income to continue supporting visually impaired people.
That's not a side benefit—it's the whole point.
The business case
Agencies need audit results they can trust, supported by real user testing and technical validation.
- Automated scans can miss up to 70% of real-world barriers without manual testing.
- Disabled people and their families in the UK have an annual spending power of £274 billion.
- 71% of disabled customers leave sites immediately if they encounter difficulties.