From 5f8de423f190bbb79a62f804151bc24824fa32d8 Mon Sep 17 00:00:00 2001 From: "Matt A. Tobin" Date: Fri, 2 Feb 2018 04:16:08 -0500 Subject: Add m-esr52 at 52.6.0 --- testing/web-platform/tests/wai-aria/OWNERS | 1 + testing/web-platform/tests/wai-aria/README.md | 65 +++++++++++++++++++++++++++ 2 files changed, 66 insertions(+) create mode 100644 testing/web-platform/tests/wai-aria/OWNERS create mode 100644 testing/web-platform/tests/wai-aria/README.md (limited to 'testing/web-platform/tests/wai-aria') diff --git a/testing/web-platform/tests/wai-aria/OWNERS b/testing/web-platform/tests/wai-aria/OWNERS new file mode 100644 index 000000000..35ae0f6c0 --- /dev/null +++ b/testing/web-platform/tests/wai-aria/OWNERS @@ -0,0 +1 @@ +@halindrome diff --git a/testing/web-platform/tests/wai-aria/README.md b/testing/web-platform/tests/wai-aria/README.md new file mode 100644 index 000000000..66d6d087f --- /dev/null +++ b/testing/web-platform/tests/wai-aria/README.md @@ -0,0 +1,65 @@ +Wai-aria: Tests for the WAI-ARIA Recommendations +================================================ + +The [WAI ARIA Recommendations](https://www.w3.org/TR/wai-aria) +define extensions to HTML4/5 for support of extended semantics. These +semantics make it easier for Assistive Technologies to intepret and +present content to users with varying physical and cognitive abilities. + +The purpose of these tests is to help ensure that user agents support the +requirements of the Recommendations. + +The general approach for this testing is to enable both manual and automated +testing, with a preference for automation. + + +Running Tests +------------- + +In order to run these tests in an automated fashion, you will need to have a +special Assistive Technology Test Adapter (ATTA) for the platform under test. We will +provide a list of these for popular platforms here as they are made available. + +The ATTA will monitor the window under test via the platforms Accessibility +Layer, forwarding information about the Accessibility Tree to the running test +so that it can evaluate support for the various features under test. + +The workflow for running these tests is something like: + +1. Start up the ATTA for the platform under test, informing it of the location + of the test server and the user agent to be tested. +2. Start up the test driver window and select the wai-aria tests to be run + (e.g., the ARIA 1.1 tests) - click "Start" +3. A window pops up that shows a test - the description of which tells the + tester what is being tested. In an automated test, the test with proceed + without user intervention. In a manual test, some user input may be required + in order to stimulate the test. +4. The test runs. Success or failure is determined and reported to the test + driver window, which then cycles to the next test in the sequence. +5. Repeat steps 2-4 until done. +6. Download the JSON format report of test results, which can then be visually + inspected, reported on using various tools, or passed on to W3C for + evaluation and collection in the Implementation Report via github. + +**Remember that while these tests are written to help exercise implementations, +their other (important) purpose is to increase confidence that there are +interoperable implementations.** So, implementers are the audience, but these +tests are not meant to be a comprehensive collection of tests for a client that +might implement the Recommendation. + + +Capturing and Reporting Results +------------------------------- + +As tests are run against implementations, if the results of testing are +submitted to [test-results](https://github.com/w3c/test-results/) then they will +be automatically included in documents generated by +[wptreport](https://www.github.com/w3c/wptreport). The same tool can be used +locally to view reports about recorded results. + + +Writing Tests +------------- + +If you are interested in writing tests for this environment, see the +associated [CONTRIBUTING](CONTRIBUTING.md) document. -- cgit v1.2.3