IBM aims to help impaired access video content
- 28 March, 2007 11:31
IBM researcher Chieko Asakawa has been blind since she was 14 years old. Since joining IBM in Japan in 1985, she has worked on myriad projects to improve accessibility for the visually impaired. Asakawa, now a senior accessibility researcher at IBM's Tokyo Research Laboratory, has been working on nonvisual computer interfaces in an effort to improve Web accessibility and usability for the visually impaired and others with special needs. She helped develop the IBM Home Page Reader in 1997 and a digital Braille system and three key applications, including a Braille Editing System to allow users to easily input and edit Braille using an ordinary keyboard and monitor. In 2004, she and her team also previewed a disability simulator that helps Web designers ensure that their pages are accessible and usable by visually impaired users. And over the last year, she and a team of four researchers have been working to make it possible for blind and visually impaired users to access multimedia content online, using a keyboard to control media player software. She discussed her work last week via e-mail from Tokyo.
Excerpts from that interview follow:
As a visually impaired researcher and active Internet user, was this project motivated by your personal situation -- knowing that all kinds of Web content was out there that you and other visually impaired people couldn't access?
Increasingly, I have been facing difficulties where I simply could not access Web content easily. I recently conducted a survey and found that most of the tested Web sites with multimedia content were not accessible. Based on this result and my personal experience, I fear that if we don't take any action, it will broaden the digital gap between the sighted and the blind. Today, accessibility for static HTML has been well established in various aspects, including technical, guidelines and regulations. And indeed, static Web contents have helped us narrow the gap. I felt that there should be ways to help narrow the gap concerning multimedia content's accessibility.
The tool set provides the user with keyboard-controlled ways to run some media player applications, such as starting the video, stopping it, rewinding it, etc.?
Without the tool set, these functions can't be controlled with a keyboard? The tool is compatible with Windows Media Player and Flash. Users only need to know a unified shortcut key operation to run video and animation. Previously, these functions could not be controlled by using a keyboard since the images that are up on the Web sites are only controllable by pointing and clicking a mouse, especially for embedded players in Web pages. Only very rarely, there might be a case where there are play/stop and volume up/down buttons that can be operated with a keyboard. However, it is hardly possible to find out the existence of such buttons while video is playing due to conflicts with screen reading software.
Once you identified the problem, how did you come up with your ideas for accessing streaming media, video and other visual content online?
First, I thought if we can provide a function to separately control volume, at least, we will be able to hear multimedia sound and screen reader sound. Technically, it was not easy to provide such a function. However, given the inability to access multimedia content pages, we did not give up; we thought we could find the way. Second, it's really frustrating to wait until inaccessible pages become accessible [by page designers who know to prepare content for visually impaired users]. If we can make inaccessible pages accessible by providing external metadata, we can significantly shorten the time to access such pages. We developed a tool [that] analyzes and adapts external metadata dynamically, and it can generate accessible pages on the fly on the client side.
Can you describe the tool set that you developed? What computer language is it written in? How does it work? How does it interact with other programs such as media players?
The tool is mostly written in Java, and works as a standalone application for users. When a user starts the application, it gives the same experience as when you open Internet Explorer and surf any Web pages. Once a user opens a Web page, the browser automatically analyzes multimedia objects inside the page, then the browser [establishes] a connection to each multimedia object. Currently, it has adaptors for [Adobe] Flash and Windows Media Player. A part of the adaptors are written in C++.
Usually, multimedia content is only designed for mouse operation, so it is impossible or too hard to operate by using a keyboard. The tool has a function to provide an alternative text-based interface to the content based on manually created XML metadata. Someone needs to create the metadata manually, but once created, usability of the site is drastically improved. It also provides functions to add audio descriptions to movies based on XML metadata. Audio descriptions are usually created by content creators as an additional sound track of a movie. While descriptions should be authored manually in advance, our tool provides a cost effective way for anyone who wants to create [and add] the audio description [later] to help blind users. Without making any changes to the content itself, the attached XML metadata information can synchronize with the video.
Lastly, the tool has a function [that] controls speech rate. Visually impaired users are well accustomed to high-speed voice, since we are using screen readers [that read text aloud for the user] every day. As for videos, it is better to provide speed control functions, but it was technically difficult. So, we developed some technology to control speech rate of any types of players, including Flash, Media Player and any types of media content can be sped up by using the tool.
Page BreakWhat is the project's status today? Is it ready for download by interested users? When will it be possible for others to try it?
We plan to open source this tool in the near future, and once that is done, we hope to discuss in detail how to make it available for downloading by interested users. Technically speaking, basic functions of the tool are almost ready to be tried out by users, but first, we would like to conduct some usability tests to make sure to what extent the tool should provide information to users.
Using the program, a visually impaired user can also use screen-reader software that will describe what is occurring on the screen in a video? How does that work?
This tool can work with screen reading software. Screen reader software is an eye for visually impaired users. Without having the tool, users face a hard time not being able to listen to screen reader software since the streaming video [soundtracks] disturb it. The tool lets users use shortcut key operations to control the level of both sounds from screen reader software and from the video soundtrack. Also, users can use shortcut operations to adjust volume of both sounds.
How has the tool kit changed your use of the Internet so far? Can you describe how it has enhanced or changed your experience with these forms of media?
It has changed it a lot. Before having this tool, I avoided pages with multimedia content since I knew I will hardly get information from such Web sites. This tool has given me opportunities to access Web sites, and it made me realize that I was missing a lot of good information [that] has been made available on the Net. By not bothering with multimedia rich contents, I simply was not aware of that fact, and the tool made me realize it. I now know good educational animation content made available as part of the Japanese e-government initiative, I can understand a lot more about Disney, which I sincerely adore, and I can now review online TV news while I am on business trips to update myself in real time of what's going on around the globe.
It's wonderful to be able to access video and animation on the Net to broaden my horizon. But, to make that truly happen, we are hoping that volunteers, content creators, developers and content providers show their interest in creating metadata [that] can be attached to the tool to help make users' experience of visiting the multimedia content sites more comfortable and seamless.
What other research and additional features can you envision being brought into the tool kit in the future? Are there still capabilities you'd like to see added for users?
I do think we need to provide easier methods to author metadata to help increase the metadata, which will provide smooth and comfortable user experience as users visit Web sites. Also, we want to provide ways to help developers to learn and check their content's accessibility easily and effectively. We are trying to make other media players controllable, such as QuickTime and RealPlayer. It will widen coverage of the tool.
Are you using this application regularly now in your Web browsing?
I am using this tool to access multimedia content, such as YouTube and news TV sites in my leisure hours. But currently, I am usually using it to test the tool. While testing, I am finding every day how multimedia content pages are designed to be visually attractive, but are not being developed with any consideration of accessibility. It will still take some time [for visually impaired users] to completely enjoy some multimedia sites. However, we want to try our best to make it happen.