Here‘s a quick overview of all of the locations in the codebase where you’ll find accessibility tests, and a brief overview of the purpose of all of them.
This is the primary place where we test accessibility code in Blink. This code should have no platform-specific code. Use this to test anything where there's any interesting web platform logic, or where you need to be able to query things synchronously from the renderer thread to test them.
Don‘t add tests for trivial features like ARIA attributes that we just expose directly to the next layer up. In those cases the Blink tests are trivial and it’s more valuable to test these features at a higher level where we can ensure they actually work.
Note that many of these tests are inherited from WebKit and the coding style has evolved a lot since then. Look for more recent tests as a guide if writing a new one.
Test files: third_party/blink/web_tests/accessibility
Source code to AccessibilityController and WebAXObjectProxy: content/web_test/renderer
First, you'll need to build the tests:
autoninja -C out/release blink_tests
Then, to run all accessibility web tests:
third_party/blink/tools/run_web_tests.py --build-directory=out --target=release accessibility/
Or, to run just one test by itself, without invoking the run_web_tests.py
script:
out/release/content_shell \ --run-web-tests third_party/blink/web_tests/accessibility/name-calc-inputs.html
For information on modifying or adding web tests, see the main web tests documentation.
This is the best way to write both cross-platform and platform-specific tests using only an input HTML file, some magic strings to describe what attributes you're interested in, and one or more expectation files to enable checking whether the resulting accessibility tree is correct or not. In particular, almost no test code is required.
More documentation on DumpAccessibilityTree
Test files: content/test/data/accessibility
Test runner: content/browser/accessibility/dump_accessibility_tree_browsertest.h
To run all tests:
autoninja -C out/release content_browsertests && \ out/release/content_browsertests --gtest_filter="All/DumpAccessibilityTree*"
Expectation baselines for each OS can be generated via:
tools/accessibility/rebase_dump_accessibility_tree_tests.py
There are many other tests in content/ that relate to accessibility. All of these tests work by launching a full multi-process browser shell, loading a web page in a renderer, then accessing the resulting accessibility tree from the browser process, and running some test from there.
To run all tests:
autoninja -C out/release content_browsertests && \ out/release/content_browsertests --gtest_filter="*ccessib*"
This tests the core accessibility code that's shared by both web and non-web accessibility infrastructure.
Code location: ui/accessibility and subdirectories; check the *_unittest.cc
files, next to every file that it's being tested.
List of sources: ui/accessibility/BUILD.gn
To run all tests:
autoninja -C out/release accessibility_unittests && \ out/release/accessibility_unittests
ChromeVox tests are part of the browser_tests suite. You must build with target_os = "chromeos"
in your GN args.
To run all tests:
autoninja -C out/release browser_tests && \ out/release/browser_tests --test-launcher-jobs=20 --gtest_filter=ChromeVox*
autoninja -C out/Default unit_tests browser_tests && \ out/Default/unit_tests --gtest_filter=*SelectToSpeak* && \ out/Default/browser_tests --gtest_filter=*SelectToSpeak*
We also have a page on Performance Tests.
Even this isn't a complete list. The tests described above cover more than 90% of the accessibility tests, and the remainder are scattered throughout the codebase. Here are a few other locations to check:
Across all tests there are some helpful flags that will make testing easier.
Run tests multiple times in parallel (useful for finding flakes): --test-launcher-jobs=10
Filter which tests to run: --gtest_filter="*Cats*"