On 21/06/2011, at 2:51 PM, Pete Brunet <[email protected]> wrote:

> 
> 
> On 6/20/2011 10:20 PM, Silvia Pfeiffer wrote:
>> 
>> Hi Pete,
>> 
>> Before I address any of this, I think there is a confusion still and
>> I'd like to make this very clear so we don't talk past each other: The
>> "descriptions" that Alex and I have been talking about for the
>> purposes of this thread are not in audio format. Instead, they are
>> text and provided to the browser in exactly the same way as captions.
> Thanks Sylvia, Yes I understood that.
>> Here is an example of such a description file:
>> http://www.annodex.net/~silvia/itext/elephants_dream/audiodescription.srt
>> . It has cues like the following:
>> 
>> 00:00:00,000 --> 00:00:05,000
>> The orange open movie project presents
>> 
>> 00:00:05,010 --> 00:00:12,000
>> Introductory titles are showing on the background of a water pool with
>> fishes swimming and mechanical objects lying on a stone floor.
>> 
>> 00:00:12,010 --> 00:00:14,800
>> elephants dream
>> 
>> 00:00:26,100 --> 00:00:28,206
>> Two people stand on a small bridge.
>> 
>> They aren't actually useful unless voiced in parallel to the video
>> that is playing and rendered during the time that the cue is
>> providing.
>> 
>> 
>> On Tue, Jun 21, 2011 at 4:52 AM, Pete Brunet <[email protected]> wrote:
>>> Hi Sylvia, We probably have more to learn from you than you from us :-)
>> Well, we have to work together to solve this problem. :-)
>> 
>>> I think even in the case of HTML5 nothing has changed for those who are
>>> either deaf or blind but not both, i.e. an additional mode can be used to
>>> compensate for the sense that is impaired, e.g. captions for the deaf and
>>> audio descriptions for the blind.  However, in the case of those who are
>>> deaf/blind then a tactile mode is needed.  One solution is to make captions
>>> available to the screen reader (with its Braille support) and the audio
>>> descriptions available as text to the screen reader.
>>> 
>>> Are there other scenarios besides use by those who are deaf/blind where text
>>> descriptions are needed?
>> In the HTML spec, there is mention of hands-free applications that
>> could make use of it, too. But I suspect that would also require use
>> of a screen reader type additional application.
> I think hands-free implies the need for voice recognition so I'm not seeing 
> that as scenario that would require text descriptions.

This is hands-free for watching the video without looking and without user 
interaction, just plain playback. Thus,  it can be regarded as an identical 
situation to being blind. But I don't want to get hung up on this because we 
don't really care about such a need on this list. So, let's just focus on blind 
users.


> 
>> 
>>> If the only need for text descriptions is to provide access for those who
>>> are deaf/blind, what are the current solutions?  Transcripts?
>> For deaf-blind people I believe transcripts are the solution and they
>> are the solution still, even though a voicing of both, captions and
>> text descriptions will provide a solution for deaf-blind people, too.
>> I regard that as a minor use case though.
>> 
>>>   Are the
>>> existing solutions insufficient enough to justify the engineering effort
>>> associated with text descriptions and stream control?
>> Text descriptions are for blind people in general, not just for
>> deaf-blind people.
>> 
>> The current solution are audio descriptions and they are much harder
>> to produce than text descriptions. So, in the interest of gaining more
> Thanks.  That's good information and an important justification.  Now that I 
> see the usefulness in supporting text descriptions I'll review the proposal 
> and the issues that have been raised so far and post another response 
> tomorrow.

Thanks very much indeed! Apologies for not explaining earlier.

Silvia.

> 
>> accessibility to video content, text descriptions were created to help
>> achieve that. Both mechanisms: audio descriptions and text
>> descriptions, are supported in HTML5.
>> 
> 
>> 
>>>   From the business
>>> perspectives that development managers are bound by the cost of design and
>>> implementation would not be justifiable for such a small user base, unless
>>> there is a legal requirement.  Is there (or will there soon be) a legal
>>> requirement to provide text descriptions?
>> I assume you are talking about development managers of accessibility
>> software? I believe the implementation of such a feature is indeed a
>> business decision and screen readers are free to compete on the
>> grounds of one having more a11y features than another. However, we are
>> here only indirectly talking about software - we are instead talking
>> about a general, standardised means of making a HTML5 feature
>> available to AT. I believe that this standardisation effort is well
>> worth the effort so we don't get different screen readers implementing
>> support for audio descriptions in a different manner when they do
>> decide to implement it.
>> 
>> 
>>> Regarding the descriptions keyword of the kind attribute, the document says
>>> that it's meant for use when visual capability is unavailable and gives the
>>> example of driving or blind users and also mentions that the text is meant
>>> for synthesis.  However, for those who are blind (and not deaf/blind) audio
>>> can still be heard and thus there is no need for a text version of an audio
>>> description.  And for deaf blind users synthesis is not needed - tactile
>>> output (Braille) is needed.
>> I hope my above description explains why text descriptions are
>> different to audio descriptions and that support for both is required.
>> Audio descriptions will indeed already work with the current
>> specification of HTML5. But we want to make the simpler authoring task
>> of creating text descriptions a more effective means of delivering
>> accessibility to videos for blind users. Note that when there are text
>> descriptions available, we would *not* expect there to also be audio
>> descriptions available.
>> 
>>> 
>>>> 
>>>>> At least at this point I'm not in favor of the media control methods.
>>>>> Developers should provide accessible GUI controls.  The developer would 
>>>>> have
>>>>> to implement the access in any case and having access through the GUI 
>>>>> would
>>>>> eliminate adding the code for these new methods on both sides of the
>>>>> interface.  If the app developer does a correct implementation of the GUI
>>>>> there would be no extra coding required in ATs.
>>>> I guess the idea here was that there may be situations where AT needs
>>>> to overrule what is happening in the UI, for example when there are
>>>> audio and video resources that start autoplaying on a newly opened
>>>> page. However, I am not quite clear on this point either.
>>> I believe the AT user would be in the same situation as a non-AT user, i.e.
>>> all users would use the same means to stop autoplaying (if such means were
>>> available).
>> That is probably true: a browser setting to generally disallow
>> autoplaying or a shortcut key in the browser to stop any and all media
>> elements that are autoplaying would be a nice browser feature for any
>> user.
>> 
>> Just to clarify: I cannot explain why we need the API in 2.7.1
>> https://wiki.mozilla.org/Accessibility/IA2_1.3#Control_video.2Faudio .
>> 
>> I do think, however, that we need the interface in 2.7.2
>> https://wiki.mozilla.org/Accessibility/IA2_1.3#Text_cues .
>> 
>> Note that I created the second interface in that section, because I
>> believe that AT needs to know the start time, end time, and exact text
>> of the to-be-read cue. I included "id" so we can keep an identifier,
>> but that may not be necessary. Also, I included both a function to
>> grab the HTML version as well as the plaint text version of the cue so
>> we have the ability to render markup differently, such as "em" can
>> create emphasis in the voicing, or navigation markers can be used to
>> jump over earlier details in the cue to later ones. I am just
>> guessing, though, what kinds of information may be useful for the
>> screenreader to receive.
>> 
>> Also, note the need to listen to the "cuechange" event on the video's
>> description track and for access to setting/unsetting the
>> "pauseOnExit" IDL attribute of the cue from the screenreader.
>> 
>> I hope I've been able to clarify a few things...
>> 
>> Cheers,
>> Silvia.
>> 
> 
> -- 
> Pete Brunet
>                                                                               
>      
> a11ysoft - Accessibility Architecture and Development
> (512) 689-4155 (cell)
> Skype: pete.brunet
> IM: ptbrunet (AOL, Google), [email protected] (MSN)
> http://www.a11ysoft.com/about/
> Ionosphere: WS4G 
> _______________________________________________
> Accessibility-ia2 mailing list
> [email protected]
> https://lists.linux-foundation.org/mailman/listinfo/accessibility-ia2
_______________________________________________
Accessibility-ia2 mailing list
[email protected]
https://lists.linux-foundation.org/mailman/listinfo/accessibility-ia2

Reply via email to