Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1878803.1878813acmconferencesArticle/Chapter ViewAbstractPublication PagesassetsConference Proceedingsconference-collections
research-article

Testability and validity of WCAG 2.0: the expertise effect

Published: 25 October 2010 Publication History

Abstract

Web Content Accessibility Guidelines 2.0 (WCAG 2.0) require that success criteria be tested by human inspection. Further, testability of WCAG 2.0 criteria is achieved if 80% of knowledgeable inspectors agree that the criteria has been met or not. In this paper we investigate the very core WCAG 2.0, being their ability to determine web content accessibility conformance. We conducted an empirical study to ascertain the testability of WCAG 2.0 success criteria when experts and non-experts evaluated four relatively complex web pages; and the differences between the two. Further, we discuss the validity of the evaluations generated by these inspectors and look at the differences in validity due to expertise.
In summary, our study, comprising 22 experts and 27 non-experts, shows that approximately 50% of success criteria fail to meet the 80% agreement threshold; experts produce 20% false positives and miss 32% of the true problems. We also compared the performance of experts against that of non-experts and found that agreement for the non-experts dropped by 6%, false positives reach 42% and false negatives 49%. This suggests that in many cases WCAG 2.0 conformance cannot be tested by human inspection to a level where it is believed that at least 80% of knowledgeable human evaluators would agree on the conclusion. Why experts fail to meet the 80% threshold and what can be done to help achieve this level are the subjects of further investigation.

References

[1]
S. Abou-Zahra. Web accessibility evaluation. In S. Harper and Y. Yesilada, editors, Web Accessibility: A Foundation for Research, Human-Computer Interaction Series, chapter 7, pages 79--106. Springer, London, first edition, Sept. 2008.
[2]
F. Alonso, J. L. Fuertes, A.L. González, and L. Martínez. On the testability of wcag 2.0 for beginners. In Web for All - W4A 2010, Raleigh, USA, April 2010. ACM.
[3]
G. Brajnik. Comparing accessibility evaluation tools: a method for tool effectiveness. Int. Journal on Universal Access in the Information Society, 3(3-4):252--263, Oct. 2004.
[4]
G. Brajnik. Beyond conformance: the role of accessibility evaluation methods. In S. Hartmann, X. Zhou, and M. Kirchberg, editors, WISE 2008: 9th Int. Conference on Web Information Systems Engineering - 2nd International Workshop on Web Usability and Accessibility IWWUA08, LNCS 5176, pages 63--80, Auckland, New Zealand, Sept. 2008. Springer-Verlag. Keynote speech.
[5]
B. Caldwell, M. Cooper, L. G. Reid, and G. Vanderheiden. Web Content Accessibility Guidelines (WCAG) 2.0. W3C, 2008. http://www.w3.org/TR/WCAG20/.
[6]
M. Catani and D. Biers. Usability evaluation and prototype fidelity: users and usability professionals. In Proc. of the Human Factors and Ergonomics Society 42nd Annual Meeting, 1998.
[7]
K. P. Coyne and J. Nielsen. How to conduct usability evaluations for accessibility: methodology guidelines for testing websites and intranets with users who use assistive technology. http://www.nngroup.com/reports/accessibility/testing, Nielsen Norman Group, Oct. 2001.
[8]
DRC. The web: Access and inclusion for disabled people. Technical Report, Disability Rights Commission (DRC), UK, 2004.
[9]
A. D. N. Edwards. Assistive technologies. In S. Harper and Y. Yesilada, editors, Web Accessibility: A Foundation for Research, Human-Computer Interaction Series, chapter 10, pages 142--162. Springer, London, first edition, Sept. 2008.
[10]
S. L. Henry and M. Grossnickle. Just Ask: Accessibility in the User-Centered Design Process. Georgia Tech Research Corporation, Atlanta, Georgia, USA, 2004. On-line book: www.UIAccess.com/AccessUCD.
[11]
M. Hertzum and N. E. Jacobsen. The evaluator effect during first-time use of the cognitive walkthrough technique. In Proc. of HCI International on Human-Computer Interaction: Ergonomics and User Interfaces-Volume I, pages 1063--1067, 1999.
[12]
M. Hertzum and N. E. Jacobsen. The evaluator effect: a chilling fact about usability evaluation methods. Int. Journal of Human-Computer Interaction, 1(4):421--443, 2001.
[13]
M. Hertzum, N. E. Jacobsen, and R. Molich. Usability inspections by groups of specialists: Perceived agreement in spite of disparate observations. In CHI 2002 Extended Abstracts, pages 662--663. ACM, ACM Press, 2002.
[14]
K. Hornbæk and E. Frøkjær. A study of the evaluator effect in usability testing. Human-Computer Interaction, 23(3):251--277, 2008.
[15]
N. E. Jacobsen, M. Hertzum, and B. John. The evaluator effect in usability tests. In CHI '98, pages 255--256. ACM, 1998.
[16]
C. Jay, D. Lunn, and E. Michailidou. End user evaluations. In S. Harper and Y. Yesilada, editors, Web Accessibility: A Foundation for Research, Human-Computer Interaction Series, chapter 8, pages 107--126. Springer, London, first edition, September 2008.
[17]
T. Lang. Comparing website accessibility evaluation methods and learnings from usability evaluation methods. http://www.peakusability.com.au/about-us/pdf/website_accessibility.pdf, Visited May 2008, 2003.
[18]
C. Ling and G. Salvendy. Effect of evaluators' cognitive style on heuristic evaluation: Field dependent and field independent evaluators. Int. Journal of Human-Computer Studies, 67(4):382--393, 2009.
[19]
J. Nielsen. Finding usability problems through heuristic evaluation. In Proc. of CHI 1992, pages 373--380, Monterey, CA, USA, May 1992. ACM.
[20]
J. Nielsen. Usability Engineering. Morgan Kaufmann, San Francisco, 1993.
[21]
H. Petrie and O. Kheir. The relationship between accessibility and usability of websites. In Proc. CHI 2007, pages 397--406, San Jose, CA, USA, 2007. ACM.
[22]
J. Rubin and D. Chisnell. Handbook of Usability Testing. Wiley, second edition, 2008.
[23]
J. Slatin and S. Rush. Maximum Accessibility: Making Your Web Site More Usable for Everyone. Addison-Wesley, 2003.
[24]
J. Thatcher, M. Burks, C. Heilmann, S. Henry, A. Kirkpatrick, P. Lauke, B. Lawson, B. Regan, R. Rutter, M. Urban, and C. Waddell. Web Accessibility: Web Standards and Regulatory Compliance. FriendsofED, 2006.
[25]
W3C/WAI. Requirements for WCAG 2.0 Checklists and Techniques. http://www.w3.org/TR/2003/WD-wcag2-tech-req-20030207, 2003.
[26]
Y. Yesilada, G. Brajnik, and S. Harper. How Much Does Expertise Matter? A Barrier Walkthrough Study with Experts and Non-Experts. In Proc. of 11th Int. ACM SIGACCESS Conference on Computers and Accessibility - ASSETS 2009, pages 203--210, Pittsburgh, PA, Oct. 2009.

Cited By

View all
  • (2113)Intelligent Interaction in Accessible ApplicationsA Multimodal End-2-End Approach to Accessible Computing10.1007/978-1-4471-5082-4_5(93-117)Online publication date: 26-Mar-2113
  • (2024)User participation in digital accessibility evaluations: Reviewing methods and implicationsEducation for Information10.3233/EFI-240014(1-16)Online publication date: 12-Sep-2024
  • (2024)Accessibility Inspections Using the Web Content Accessibility Guidelines by Novice Evaluators: an Experience ReportProceedings of the XXIII Brazilian Symposium on Human Factors in Computing Systems10.1145/3702038.3702040(1-10)Online publication date: 7-Oct-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
ASSETS '10: Proceedings of the 12th international ACM SIGACCESS conference on Computers and accessibility
October 2010
346 pages
ISBN:9781605588810
DOI:10.1145/1878803
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 October 2010

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. evaluation
  2. expertise
  3. guideline
  4. web accessibility

Qualifiers

  • Research-article

Conference

ASSETS '10
Sponsor:

Acceptance Rates

Overall Acceptance Rate 436 of 1,556 submissions, 28%

Upcoming Conference

ASSETS '25

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)58
  • Downloads (Last 6 weeks)5
Reflects downloads up to 12 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2113)Intelligent Interaction in Accessible ApplicationsA Multimodal End-2-End Approach to Accessible Computing10.1007/978-1-4471-5082-4_5(93-117)Online publication date: 26-Mar-2113
  • (2024)User participation in digital accessibility evaluations: Reviewing methods and implicationsEducation for Information10.3233/EFI-240014(1-16)Online publication date: 12-Sep-2024
  • (2024)Accessibility Inspections Using the Web Content Accessibility Guidelines by Novice Evaluators: an Experience ReportProceedings of the XXIII Brazilian Symposium on Human Factors in Computing Systems10.1145/3702038.3702040(1-10)Online publication date: 7-Oct-2024
  • (2024)Tab to Autocomplete: The Effects of AI Coding Assistants on Web AccessibilityProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3688513(1-6)Online publication date: 27-Oct-2024
  • (2024)Automating Error Identification and Evaluating Web Accessibility for Differently Abled UsersProceedings of Data Analytics and Management10.1007/978-981-99-6547-2_22(275-292)Online publication date: 3-Jan-2024
  • (2023)Accessibility of Mobile Apps for Visually Impaired Users: Problems Encountered by User Evaluation, Inspections and Automated ToolsProceedings of the XXII Brazilian Symposium on Human Factors in Computing Systems10.1145/3638067.3638101(1-11)Online publication date: 16-Oct-2023
  • (2023)Accessibility Inspections of Mobile Applications by Professionals with Different Expertise Levels: An Empirical Study Comparing with User EvaluationsHuman-Computer Interaction – INTERACT 202310.1007/978-3-031-42280-5_9(135-154)Online publication date: 25-Aug-2023
  • (2022)Accessibility-Related Publication Distribution in HCI Based on a Meta-AnalysisExtended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems10.1145/3491101.3519701(1-28)Online publication date: 27-Apr-2022
  • (2022)Accessibility Feedback in Mobile Application Reviews: A Dataset of Reviews and Accessibility GuidelinesExtended Abstracts of the 2022 CHI Conference on Human Factors in Computing Systems10.1145/3491101.3519625(1-7)Online publication date: 27-Apr-2022
  • (2022)Accessibility evaluation of Arabic University websites for compliance with success criteria of WCAG 1.0 and WCAG 2.0Universal Access in the Information Society10.1007/s10209-022-00921-822:4(1199-1214)Online publication date: 3-Oct-2022
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media