The impact of artificial intelligence (AI) on various areas of higher education\u2014including learning design, academic integrity, and assessment\u2014is routinely debated and discussed. Arguably, one area that is not explored as critically or thoroughly is the impact of AI on digital accessibility, inclusion, and equity. Several exciting technological developments in this space offer promise and optimism inside and outside of higher education. These advancements afford people with disabilities more equitable access to the same educational services and resources offered to students without disabilities. Ironically, students with disabilities\u2014who stand to gain the most from emerging AI tools and resources\u2014are often the most disadvantaged or least able to use them. More concerning still is that few people in the disabled community have been asked to advise on the development of these products. A 2023 survey of assistive technology users found that fewer than 7 percent of respondents with disabilities believe there is adequate representation of their community in the development of AI products\u2014although 87 percent would be willing to provide end-user feedback to developers.<\/p>\n
As an accessibility advocate with a hearing impairment, I have been keenly interested in AI advancements that have the potential to provide people with disabilities more equitable access to educational content\u2014especially since the release of ChatGPT and other large language models (LLMs). However, the use of AI in educational technology and instruction is hardly new. The Programmed Logic for Automatic Teaching Operations (PLATO) system was developed at the University of Illinois in the 1960s, and Jaime Carbonell developed SCHOLAR at Stanford University in the 1970s. Both computer learning tools are early forms of AI. More contemporary edtech developments were introduced in the 2000s, such as ALEKS, Newton, and intelligent tutoring systems (ITS)\u2014all popular, widely used student-facing AI courseware platforms. AI-generated automatic captions for web conferencing also became widely available in the early to mid-2010s, along with significant advancements in automatic speech recognition (ASR) technology. In 2009, Google introduced automatic captioning for YouTube videos. This groundbreaking development demonstrated the potential for using speech recognition to generate captions in real time or from recorded content. <\/sup>Although this technology was initially panned for its inaccuracy and high error rate, it gradually influenced similar capabilities in other technologies, including web conferencing platforms.<\/p>\n The relatively recent release of LLMs has ushered in a surge in AI product development in this space. A few recently introduced edtech products and services are described below. While this list is not exhaustive, it captures capabilities that were thought impossible only a short time ago.<\/p>\n For screen readers to accurately decipher the content of pictures, images, and diagrams, content authors must add descriptions, labels, or alt text (also referred to as alternative text). With the advent of LLMs, AI technologies can auto-generate these descriptions. Several tools that generate image descriptions are in early development and release. For example, Arizona State University recently launched a new AI-image description utility<\/a> that uses ChatGPT-4o to analyze user-uploaded images and produce robust alternative text descriptions. This tool can also analyze and extract embedded text (i.e., text that is not machine-readable) from slides and images.<\/p>\n Accessibility advocate and developer Cameron Cundiff created a Non-Visual Desktop Access (NVDA) add-in<\/a> that provides semantically rich image descriptions of any website, software product, or desktop icon. This tool uses the vision capabilities in the Google Gemini API to analyze and generate robust image descriptions that can be read back through the speech synthesizer in NVDA.<\/p>\n Astica.ai<\/a> used its Vision API technology to develop an image description tool that generates captioned images, brand identification, and automatic content moderation. Users can upload complex images, and Astica.ai will automatically scan and identify elements and generate detailed alt-text descriptions.<\/p>\n Researchers at MIT developed VisText to help people generate captions and descriptions of complex charts and graphs\u2014among the most challenging image types for assistive technology to describe. This tool is particularly useful for describing complex patterns and trends within chart data.Footnote<\/span>5<\/a><\/sup><\/p>\n Darren DeFrain, an English professor at Wichita State University, led a team of developers that created Vizling,<\/a> a mobile device app designed to make multimodal media, such as comics, maps, graphic novels, and art, accessible for blind and low-vision readers. Screen-reading products have difficulty parsing comics and graphic novels because their panel-based layout and use of speech balloons do not conform to predictable patterns.<\/p>\n U.K.-based WPP is working with Microsoft to develop advanced audio description tools built on GPT4. This technology generates enhanced audio descriptions of user-uploaded videos and images. The company is also working collaboratively with the Rijksmuseum, the national museum of the Netherlands, to provide enhanced audio descriptions for its collection of nearly one million works of art, opening the door to libraries with extensive special collections. <\/sup>This tool is expected to be available soon.<\/p>\n Microsoft UK recently introduced a series of vignettes<\/a> showcasing how AI technology is being used to support people with various cognitive and physical disabilities. Nearly all of the featured use cases have direct applications in higher education.<\/p>\n In 2023, Microsoft partnered with OpenAI to develop Be My AI, a digital visual assistant within the Be My Eyes app. Be My AI is powered by OpenAI’s Vision API, which contains a dynamic new image-to-text generator. Be My AI users can send images and ask questions via the Be My AI app. An AI-powered virtual volunteer answers any questions about the images and provides instantaneous visual assistance for a variety of tasks. This technology provides enhanced opportunities for learners who are blind or have low vision.<\/p>\n Goodwin University in Connecticut is experimenting with AI products to support neurodivergent students. For example, the university recommends GitMind for assistive notetaking, mind mapping, and brainstorming.<\/p>\n The University of Central Florida, in conjunction with United Cerebral Palsy (UCP) of Central Florida, has developed “ZB”\u2014an AI-driven socially assistive robot\u2014as part of Project RAISE. ZB is designed to help students with disabilities develop and improve their social skills and can even teach them how to code. “He hangs out with students in their classes, affirming them with positive messages,” according to a Kansas City PBS news story.<\/p>\n GPT Accessibility CoPilot,<\/a> developed by Joe Devon, co-founder of Global Accessibility Awareness Day (GAAD) and chair of the GAAD Foundation, is a tool that helps content developers and instructional designers by analyzing the code structure in web and content pages and matching it against WCAG 2.2 Success Criteria. If the code does not meet the criteria, Accessibility CoPilot provides suggestions for improving it.<\/p>\n Ask Microsoft Accessibility<\/a> is a free tool that can be used by faculty and students to develop accessible course content. Users can type a question such as, “How do I make Excel files more inclusive?” and the AI assistant provides several solutions in near real time. This product is in early release.<\/p>\n Procter & Gamble is using an AI-assisted QR Code technology called Navilens<\/a> to assist people who are blind or have low vision. Navilens can be used to locate products among dense shelving and read the instructions for use or list of ingredients. This technology is also available to venues that require wayfinding and sign-reading services. Navilens is free to download and use, and the company is currently offering its proprietary codes to schools. The company has partnered with Microsoft to provide greater autonomy to users of a specialized headset developed by ARxVision.<\/sup><\/p>\n GitHub recently launched Copilot, a code completion tool developed in conjunction with Microsoft and OpenAI. GitHub Copilot Chat is a complementary chat interface that can help programmers learn about accessibility and improve the accessibility of their code.<\/sup><\/p>\n Accessibility and training company Deque announced the release of Axe DevTools AI,<\/a> a suite of tools that can be used by web developers to test and correct the digital accessibility of web content and other website elements. For example, Colorvision (just one of the tools in the suite) automatically checks for incompatible color contrast. At the Axe-Con 2024 conference, Gregg Vanderheiden, professor emeritus at the University of Maryland, predicted that AI-powered tools would provide near ubiquity to all digital products and that these products would adapt to the user’s accessibility preferences in real time.<\/sup><\/p>\n LLMs have made possible a variety of new translation, caption, lip reading, and speech recognition tools. For example, Microsoft Copilot+ PCs include live translation in nearly every language.<\/a> Previously, this technology was available only in certain productivity products, such as PowerPoint; however, it is now poised for wide availability across various Microsoft productivity products.<\/p>\n SRAVI (Speech Recognition App for the Voice Impaired) is an AI-powered lip-reading app developed by Fabian Campbell-West, co-founder and CTO of Liopa, a software development company in Belfast, Ireland. SRAVI was initially developed to help ICU and critical care patients who have lost the ability to speak communicate more effectively with their families and health care providers. In 2023, the app was being tested on patients who had undergone a total laryngectomy procedure. Liopa is a spin-out from Queen’s University Belfast and its Centre for Security Technologies. Although the company was dissolved earlier this year, the SRAVI app is still available for download.<\/sup><\/p>\nAutomated Image Descriptions<\/h2>\n
Audio Description Generation<\/h2>\n
Support for Cognitive and Physical Disabilities<\/h2>\n
Inclusive Design Support<\/h2>\n
Coding and Development Support<\/h2>\n
Translations, Captions, Lip Reading, and Speech Recognition<\/h2>\n