Hi - I second the remark about the global indicator. I have worked closely for 
a couple of years with my university’s accessibility group to assess, identify 
and then fix accessibility issues within our public-facing interfaces. Our 
experience is that there is no efficacious method for automated / without human 
interpretation or intervention processes. There are several problems. 

First, you have to have a target level of accessibility, generally. There are a 
variety of standards, with differing levels of compliance within these 
standards. I suppose you could publish just a raw ranking of ’scores’ for 
libraries in relation to various standards, but is that meaningful to anyone 
anywhere? I am not sure.

Primarily because of issue 2, which is that you need to use multiple tools to 
acquire appropriate impressions of your libraries’ accessibility, in relation 
to your target. Many of these tools are commercial, some are browser plugins, 
some are server-based, with differing compliance assessment tools being used by 
different universities and the government. The bottom line for most 
universities is ‘are you in compliance with what the federal government 
requires for institutions receiving federal funding’ in a realistic sense, so 
that is the accessibility target, the Section 508 items as well as compliance 
with other laws and statutes that relate to web accessibility, some being state 
laws, depending on your state.

Issue 3 is that an automated review off accessibility issues will be - In my 
experience - highly erroneous depending upon the complexity of your site, and 
the technologies used. For instance, one site I assessed and ‘corrected’ was a 
a page that used Bootsrap / Angular /Node.js to implement large drop down menus 
that were effectively hidden from, but also organized for,  screen readers 
using various tags and positioning within the code. It was 100% effective in 
all supported browsers, and was parsed correctly by common screen readers, but 
was yet indicted by 1 or 2 assessment tools as being a serious issue, but they 
were not, and the automated claims of the assessment tools were wrong, and it 
required a human assessment of the reporting function to make a correct 
assessment. There were/ are many instance of this across all of our properties. 
If that had been a strictly automated process, we would have recorded an 
important false-positive, and given the interface a much lower score than it 
actually warranted.

Assessment of library accessibility is a mix of qualitative and quantitative. 
An accurate review would contain elements of both, not strictly one or the 
other., IMO.

> On Oct 23, 2020, at 8:37 AM, Caffrey-Hill, Julia <[email protected]> 
> wrote:
> 
> Hello Dr. Parthasarathi Mukhopadhyay,
> I can provide some partial thoughts, and there are other members who have 
> strong, knowledgeable perspectives that may want to chime in also.
> 
> Re: 2. 
> - For ARIA, there's consensus that a high number of ARIA found on a page is 
> not necessarily an indicator of accessibility and, to the contrary, a high 
> score is a red flag that may indicate abuse of ARIA tags. They are easily 
> mishandled. There are others in this community, namely Katherine Deibel, who 
> are prolific on this topic that I hope can chime in or link to part 
> presentations/resources.
> - For your study, as it relates to ARIA specifically, I recommend AXE browser 
> extension (https://www.deque.com/axe/). I don't think an API is available for 
> it, but it is good for validation, and I believe is suited to a quantitative 
> study. There is a learning curve on understanding it. Deque Systems, 
> according to their training, split off from the team behind WAVE, and built 
> out the tool's capacity for testing ARIA tags.
> 
> Re: 3
> - In terms of a globally recognized quantitative indicator, I'm not aware of 
> one. A combination of different tools is recommended, and they do have their 
> weak spots. I prefer mixed methods to test for web accessibility. 
> - For a large number of websites at a time, I understand the need for a 
> framework. For auditing our e-resources for accessibility, Towson University 
> adapted a framework from Princeton University, who in turn adapted it from 
> another library. My colleagues and I recently presented on how to do this 
> approach (Description: 
> https://wp.towson.edu/tcal/one-step-at-a-time-assessing-e-resources-for-accessibility-compliance/
>  Recording: https://www.youtube.com/watch?v=zQZjTeW-69E&feature=youtu.be  - 
> 40 mins) - I hope that's helpful and if so, I'd be interested to hear about 
> it.
> 
> All the best,
> Julia Caffrey-Hill
> Web Services Librarian
> Towson University
> 
> -----Original Message-----
> From: Code for Libraries <[email protected]> On Behalf Of Parthasarathi 
> Mukhopadhyay
> Sent: Thursday, October 22, 2020 7:55 AM
> To: [email protected]
> Subject: [CODE4LIB] Web accessibility and ARIA
> 
> [EXTERNAL EMAIL - USE CAUTION]
> 
> Hello all
> 
> We are trying to measure web accessibility of some Indian 
> institutes/universities/libraries in the form of a score and then rank those 
> institutes/universities/libraries against the score (still at the idea 
> plane). The plan is to fetch data through API in a data wrangling software 
> for further analysis. My questions are as follows:
> 
> 1) Are there other services (apart from WAVE) that provide results in JSON 
> format through API?
> 2) What is the significance of *ARIA* in determining such a score for web 
> accessibility? Does a higher number of ARIA indicate a better accessibility? 
> Or is converse true?
> 3) Is there any globally agreed-upon indicator for web accessibility?
> 
> Best
> 
> -----------------------------------------------------------------------
> Dr. Parthasarathi Mukhopadhyay
> Professor, Department of Library and Information Science, University of 
> Kalyani, Kalyani - 741 235 (WB), India
> -----------------------------------------------------------------------

Attachment: smime.p7s
Description: S/MIME cryptographic signature

Reply via email to