AI technology tools hold remarkable promise for providing more accessible, equitable, and inclusive learning experiences for students with disabilities.
The impact of artificial intelligence (AI) on various areas of higher education—including learning design, academic integrity, and assessment—is routinely debated and discussed. Arguably, one area that is not explored as critically or thoroughly is the impact of AI on digital accessibility, inclusion, and equity. Several exciting technological developments in this space offer promise and optimism inside and outside of higher education. These advancements afford people with disabilities more equitable access to the same educational services and resources offered to students without disabilities. Ironically, students with disabilities—who stand to gain the most from emerging AI tools and resources—are often the most disadvantaged or least able to use them. More concerning still is that few people in the disabled community have been asked to advise on the development of these products. A 2023 survey of assistive technology users found that fewer than 7 percent of respondents with disabilities believe there is adequate representation of their community in the development of AI products—although 87 percent would be willing to provide end-user feedback to developers.
As an accessibility advocate with a hearing impairment, I have been keenly interested in AI advancements that have the potential to provide people with disabilities more equitable access to educational content—especially since the release of ChatGPT and other large language models (LLMs). However, the use of AI in educational technology and instruction is hardly new. The Programmed Logic for Automatic Teaching Operations (PLATO) system was developed at the University of Illinois in the 1960s, and Jaime Carbonell developed SCHOLAR at Stanford University in the 1970s. Both computer learning tools are early forms of AI. More contemporary edtech developments were introduced in the 2000s, such as ALEKS, Newton, and intelligent tutoring systems (ITS)—all popular, widely used student-facing AI courseware platforms. AI-generated automatic captions for web conferencing also became widely available in the early to mid-2010s, along with significant advancements in automatic speech recognition (ASR) technology. In 2009, Google introduced automatic captioning for YouTube videos. This groundbreaking development demonstrated the potential for using speech recognition to generate captions in real time or from recorded content. Although this technology was initially panned for its inaccuracy and high error rate, it gradually influenced similar capabilities in other technologies, including web conferencing platforms.
The relatively recent release of LLMs has ushered in a surge in AI product development in this space. A few recently introduced edtech products and services are described below. While this list is not exhaustive, it captures capabilities that were thought impossible only a short time ago.
Automated Image Descriptions
For screen readers to accurately decipher the content of pictures, images, and diagrams, content authors must add descriptions, labels, or alt text (also referred to as alternative text). With the advent of LLMs, AI technologies can auto-generate these descriptions. Several tools that generate image descriptions are in early development and release. For example, Arizona State University recently launched a new AI-image description utility that uses ChatGPT-4o to analyze user-uploaded images and produce robust alternative text descriptions. This tool can also analyze and extract embedded text (i.e., text that is not machine-readable) from slides and images.
Accessibility advocate and developer Cameron Cundiff created a Non-Visual Desktop Access (NVDA) add-in that provides semantically rich image descriptions of any website, software product, or desktop icon. This tool uses the vision capabilities in the Google Gemini API to analyze and generate robust image descriptions that can be read back through the speech synthesizer in NVDA.
Astica.ai used its Vision API technology to develop an image description tool that generates captioned images, brand identification, and automatic content moderation. Users can upload complex images, and Astica.ai will automatically scan and identify elements and generate detailed alt-text descriptions.
Researchers at MIT developed VisText to help people generate captions and descriptions of complex charts and graphs—among the most challenging image types for assistive technology to describe. This tool is particularly useful for describing complex patterns and trends within chart data.Footnote5
Darren DeFrain, an English professor at Wichita State University, led a team of developers that created Vizling, a mobile device app designed to make multimodal media, such as comics, maps, graphic novels, and art, accessible for blind and low-vision readers. Screen-reading products have difficulty parsing comics and graphic novels because their panel-based layout and use of speech balloons do not conform to predictable patterns.
Audio Description Generation
U.K.-based WPP is working with Microsoft to develop advanced audio description tools built on GPT4. This technology generates enhanced audio descriptions of user-uploaded videos and images. The company is also working collaboratively with the Rijksmuseum, the national museum of the Netherlands, to provide enhanced audio descriptions for its collection of nearly one million works of art, opening the door to libraries with extensive special collections. This tool is expected to be available soon.
Support for Cognitive and Physical Disabilities
Microsoft UK recently introduced a series of vignettes showcasing how AI technology is being used to support people with various cognitive and physical disabilities. Nearly all of the featured use cases have direct applications in higher education.
In 2023, Microsoft partnered with OpenAI to develop Be My AI, a digital visual assistant within the Be My Eyes app. Be My AI is powered by OpenAI’s Vision API, which contains a dynamic new image-to-text generator. Be My AI users can send images and ask questions via the Be My AI app. An AI-powered virtual volunteer answers any questions about the images and provides instantaneous visual assistance for a variety of tasks. This technology provides enhanced opportunities for learners who are blind or have low vision.
Goodwin University in Connecticut is experimenting with AI products to support neurodivergent students. For example, the university recommends GitMind for assistive notetaking, mind mapping, and brainstorming.
The University of Central Florida, in conjunction with United Cerebral Palsy (UCP) of Central Florida, has developed “ZB”—an AI-driven socially assistive robot—as part of Project RAISE. ZB is designed to help students with disabilities develop and improve their social skills and can even teach them how to code. “He hangs out with students in their classes, affirming them with positive messages,” according to a Kansas City PBS news story.
Inclusive Design Support
GPT Accessibility CoPilot, developed by Joe Devon, co-founder of Global Accessibility Awareness Day (GAAD) and chair of the GAAD Foundation, is a tool that helps content developers and instructional designers by analyzing the code structure in web and content pages and matching it against WCAG 2.2 Success Criteria. If the code does not meet the criteria, Accessibility CoPilot provides suggestions for improving it.
Ask Microsoft Accessibility is a free tool that can be used by faculty and students to develop accessible course content. Users can type a question such as, “How do I make Excel files more inclusive?” and the AI assistant provides several solutions in near real time. This product is in early release.
Procter & Gamble is using an AI-assisted QR Code technology called Navilens to assist people who are blind or have low vision. Navilens can be used to locate products among dense shelving and read the instructions for use or list of ingredients. This technology is also available to venues that require wayfinding and sign-reading services. Navilens is free to download and use, and the company is currently offering its proprietary codes to schools. The company has partnered with Microsoft to provide greater autonomy to users of a specialized headset developed by ARxVision.
Coding and Development Support
GitHub recently launched Copilot, a code completion tool developed in conjunction with Microsoft and OpenAI. GitHub Copilot Chat is a complementary chat interface that can help programmers learn about accessibility and improve the accessibility of their code.
Accessibility and training company Deque announced the release of Axe DevTools AI, a suite of tools that can be used by web developers to test and correct the digital accessibility of web content and other website elements. For example, Colorvision (just one of the tools in the suite) automatically checks for incompatible color contrast. At the Axe-Con 2024 conference, Gregg Vanderheiden, professor emeritus at the University of Maryland, predicted that AI-powered tools would provide near ubiquity to all digital products and that these products would adapt to the user’s accessibility preferences in real time.
Translations, Captions, Lip Reading, and Speech Recognition
LLMs have made possible a variety of new translation, caption, lip reading, and speech recognition tools. For example, Microsoft Copilot+ PCs include live translation in nearly every language. Previously, this technology was available only in certain productivity products, such as PowerPoint; however, it is now poised for wide availability across various Microsoft productivity products.
SRAVI (Speech Recognition App for the Voice Impaired) is an AI-powered lip-reading app developed by Fabian Campbell-West, co-founder and CTO of Liopa, a software development company in Belfast, Ireland. SRAVI was initially developed to help ICU and critical care patients who have lost the ability to speak communicate more effectively with their families and health care providers. In 2023, the app was being tested on patients who had undergone a total laryngectomy procedure. Liopa is a spin-out from Queen’s University Belfast and its Centre for Security Technologies. Although the company was dissolved earlier this year, the SRAVI app is still available for download.
Ava is a mobile app that allows people who are deaf or hard of hearing to take part in group conversations in English, Dutch, French, German, Italian, or Spanish. The app provides limited conversational support for twenty spoken languages. People who are engaged in a conversation can open Ava on their phones and then speak as the app listens. Ava converts spoken words into text in nearly real time, rendering each speaker’s words in a different color to help those who need to read along follow the chat.
The University of Illinois is working with Microsoft, Google, Amazon, and several nonprofit organizations on the Speech Accessibility Project, an interdisciplinary initiative “to make voice recognition technology useful for people with a range of speech patterns and disabilities.” University researchers are recording individuals who have Parkinson’s disease, Down syndrome, cerebral palsy, stroke, and amyotrophic lateral sclerosis. Those recordings are used to train an AI automatic speech recognition tool. According to the Speech Accessibility Project website, “Before using recordings from the Speech Accessibility Project, the tool misunderstood speech 20 percent of the time. With data from the speech accessibility project, this decreased to 12 percent.”
Conclusion
These are just a few of the exciting developments emerging from the intersection of AI and accessibility. Many people in the higher education community are rightfully cautious about the use of AI. However, numerous products and services that promise more equity and inclusion for people with disabilities are currently available or in development.