Skip to content
This repository has been archived by the owner on Apr 7, 2022. It is now read-only.

Evaluating tools for automatic identification of accessibility fails

Notifications You must be signed in to change notification settings

ukgovdatascience/accessibility_fails

Repository files navigation

Accessibility Fails

Quick brief

The Accessibility team are using a number of automated services to check pages for their accessibility. They have made an assessment of ten of these tools here.

What they want to know is:

  • What are the best combinations of the different tools
  • How does this change when you consider different scoring criteria?

Scoring criteria

The three scoring criteria are:

  • error found
  • error found + warning reported
  • error found + warning reported + manual inspection

Outline of method

  • Hash the various accessibility fails to make them easier to work with
  • Create three dataframes and binarise for the three success criteria
  • Compute rowsums for all the possible combinations of accessibility tools
  • Visualise this data somehow

Issues

  • Test 16 and 81: 'Text language changed without required change in direction' were replicated in accessibility_fail_data.csv. The test at line 81 was appended with a 1 in teh csv to solve this issue.

About

Evaluating tools for automatic identification of accessibility fails

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published