Since the most recent explosion of the widespread generative artificial intelligence (AI), it has seemed that a brand new AI tool has been created every week.
With different success, AI offers solutions for productivity, creativity, research and accessibility: products, services and other content for individuals with disabilities more more usable.
The Prize winner 2024 Super Bowl display for Google Pixel 8 is a moving example of how the most recent AI technology can overlap with disabilities.
Under the direction of Blind Director Adam Morse, it shows a function operated with AI, through which audio notes, haptic feedback (where vibrating sensations provide information to the user) and animations to support users of blind and low when taking photos and videos.
https://www.youtube.com/watch?v=Wyptzifqodq
The commercial was welcomed to be including and representative disabilities. It also showed a growing capability for and interest in AI to be able to generate more accessible technologies.
AI can be able to challenge how audio description is created and the way it might sound. This is the main focus of our research team.
The audio description is a track of the narrative that describes necessary visual elements of visual media, including television programs, movies and live appearances. Synthetic voices and fast, automated visual descriptions can lead to a different audio description on our screens. But will users lose in one other way?
AI because the eyes of the people
Proliferate AI-powered accessibility tools. Among them is Microsoft's See AIAn app that turns your smartphone right into a speaking camera by reading text and identifying objects. The app Be my AI Used virtual assistants to explain photos taken by blind users. It is a AI version of the unique app is my eyes, through which the identical task was done by human volunteers.
There are an increasing number of AI software options for text-to-speech and document reading. To produce audio description.
The audio description is a necessary function to make visual media accessible to the blind or visually impaired goal groups. But his benefits transcend.
Studies are increasingly showing Audio description advantages other groups of disabled people And Mainstream audience without disabilities. Audio description will also be a creative approach to proceed Develop or improve a visible text.
Traditionally, the audio description was created with human voices, screenwriters and production teams. Last 12 months, nevertheless, several international streaming services, including Netflix and Amazon Prime I began offering audio description that’s not less than partially generated with AI.
However, there are a lot of problems with the present AI technologies, including their ability to generate misinformation. These tools should be assessed and improved critically.
Is AI for Audio description Jobs?
There are several ways of how AI can influence the creation – and the tip result – of the audio description.
With AI tools, streaming services can receive Synthetic voices for “reading” an audio description script. There is the potential for various levels of automationwhile I give Users the likelihood to adapt audio description meet your specific needs and preferences. Would you want your cooking show to be told in a British accent? With AI you possibly can change this by pressing a button.
In the audio description of the industry, many are concerned that AI could undermine the standard, creativity and professionalism that folks bring to the equation.
For example, the language learning app Duolingo has recently announced that it goes with it Development “Ai First”. As a result, many contractors lost jobs that may now allegedly be done by algorithms.
On the one hand, AI could help expand the choice of audio descriptions which might be available for a lot of media and live experiences.
However, the AI audio description may also cost jobs as an alternative of making them. The worst result can be a considerable amount of audio description of less quality, which might undermine the worth of making.
Landbild/Shutterstock
Can we trust ai to explain things well?
Industry effects and the technical details on how AI could be utilized in the audio description are one thing.
What is currently missing is research that focuses on the perspectives of the users and takes into consideration their experiences and desires for future audio description.
Accuracy and trust on this accuracy is of crucial importance for the audience with blind and low vision.
Cheap and sometimes free AI tools at the moment are widespread to summarize, transcribe and translate. But it’s a well -known problem that generative AI has difficulty remaining factual. Known as “hallucinations”, these plausible inventions also multiply when the AI tools are not asked to create something recent – Like a straightforward audio transcription.
If AI tools simply produce content as an alternative of constructing existing materials accessible, this might be distant and disadvantaged blind and sensitive consumers.
We can use AI for accessibility – with caution
AI is a comparatively recent technology, and in order that it’s an actual profit in relation to accessibility, your accuracy and reliability should be absolutely absolutely. Blind and low-vision users must give you the chance to modify on AI tools with confidence.
In the present “AI Rush” to make the audio description cheaper, faster and more available, it will be important that the people you wish essentially the most are closely involved in the availability of technology.