The project involves an AI system that senses SEMG surface muscle signals and allows classifying voluntary gestures of a person to control and manage a music app remotely. The system allows controlling audio and video devices by recognizing hand gestures and does not need to be re-trained every time. Results show a success rate above 94% in controlling the music app using hand gesture recognition and AI techniques.
Creating a new interface to control music apps, electronic devices or apps for people with and without restricted mobility.
Interaction support (chatbots, virtual assistants and others), Customization, Recognition, Prediction
A model that allows controlling electronic applications using hand gesture recognition built from a database of 612 individuals, six gestures/person, 50 repetitions/gesture (183.600 entries). The AI model reported a 94% accuracy. This robust model was tested in an app with different bracelet-type sensors obtaining similar results, allowing the adaptability of multiple bracelet sensors without needing to re-train the model. The tested music app responded satisfactorily, and the system can control new devices remotely with minimal changes.
Quito, Ecuador
3 (good health and well-being)
4 (quality education)
(decent work and economic growth)
(industry, innovation, and infrastructure)
(reduced inequalities)
Internally
Alan Turing AI and Vision Research Laboratory
Alan Turing AI and Vision Research Laboratory
30%
Este documento fue diseñado como insumo de una hoja de ruta que permite crear un marco para el uso ético, responsable y seguro de la IA en Costa Rica.
El presente documento presenta un estudio de caso de un proyecto piloto en el estado de Jalisco que utilizó inteligencia artificial (IA) para el tamizaje de la Retinopatía Diabética.
Este documento forma parte de la serie “Camino hacia la inclusión educativa: 4 pasos para la construcción de sistemas de protección de trayectorias”.