An AI-powered voice control system enhancing digital accessibility for individuals with physical disabilities
The increasing adoption of voice-enabled technologies has highlighted their potential to empower individuals with physical disabilities, enabling them to independently navigate digital platforms. However, challenges such as noise interference, limited command vocabulary, and platform incompatibility persist.
Our research product, which makes use of state-of-the-art technologies including AI, Natural Language Processing, and Automatic Speech Recognition (ASR), is a Voice-Controlled Integration System. Key concerns including cross-platform usability, noise resilience, and speech recognition accuracy are addressed by the system, which seeks to offer hands-free control over computers and communication platforms for amputees. Through the integration of features like as voice-controlled media playing, speech-to-text, and application navigation, the solution provides increased accessibility and independence with a certain degree of precision.
Voice technologies have a clear future in helping people with physical disabilities specially amputees communicate with the digital world independently. However, a few challenges remained, such as noise interference, limited contextual understanding, and a deficiency in platform compatibility.
According to recent studies, around 40% of physically disabled individuals use voice commands for web browsing, yet only 2% of websites are fully assistive technology compatible. Such statistics signify the need for vigorous and universal solutions to accessibility challenges.
There are currently more than 1 billion disabilities in the world, so there is an increasing demand for technologies that can make things accessible and usable.
Voice-activated systems based on development in Automatic Speech Recognition and Natural Language Processing can be very promising, but they are often challenged in "real-life" environments, especially in noisy conditions or while supporting multilingual users. Fixing these issues is significant to return the benefits of such systems into the more extensive user community.
Our project aims to provide a voice-controlled system "SONIX" tailored specifically for individuals with physical disabilities who have lost the ability to use their hands due to congenital conditions or accidents. The methodology is divided into three key phases: Input, Processing, and Output.
90% success rate
80% success rate
70% success rate
80% success rate
95% average accuracy
85-95% accuracy
92% accuracy in noisy environments
55% accuracy with diverse accents
The application runs on the Microsoft Windows operating system and is tested by four disabled individuals with different accents and voices. An eye gesture virtual mouse allows the user to write in Microsoft Word with formatting capabilities, manage and conduct video meetings, and operate installed applications.
In addition to sending and receiving emails, this model allows you to open any mail browser, check your junk folder, clean out your inbox, and browse the internet without any issues. Also, the model might eventually be able to zip any file, copy it, and transfer it across the system as well as platform independent with suitable and reliable system security.
Future Enhancements: