SpeAKN (Speak for ALS with Korean NLP) is an innovative AI-powered communication system designed specifically for ALS patients who have lost their ability to speak. The system combines eye-tracking technology with advanced Korean natural language processing to provide contextually appropriate response suggestions, enabling meaningful communication for patients with progressive motor function decline.
ALS (Amyotrophic Lateral Sclerosis) is a neurodegenerative disease in which motor nerve cells in the brain and spinal cord progressively deteriorate. Over time, ALS patients lose their ability to communicate using natural language, with 80-95% of patients requiring alternative communication methods known as AAC (Augmentative and Alternative Communication).
Existing solutions like "Look to Speak" require users to make numerous selections before reaching their intended response, and users cannot pre-input their desired responses. Our AI-TRACKING system addresses these limitations by combining eye-tracking technology with artificial intelligence to provide contextually relevant response options.
The SpeAKN system operates through a two-stage AI pipeline:
This project utilized the National Institute of Korean Language's datasets:
We experimentally tested various optimization approaches:
The AdamW Optimizer demonstrated superior performance with faster learning progression and better stabilization during training compared to the Sophia Optimizer.
Our SpeAKN model employs several key innovations:
To handle varying audio lengths efficiently, we analyzed the distribution of audio data lengths:
We set the standardized audio length to 100,000 samples to balance computational efficiency with data preservation, avoiding the curse of dimensionality while maintaining essential information.
| Model Component | Train MSE | Validation MSE | Input Type | Output Quality |
|---|---|---|---|---|
| Speech2Text | 10.039 | 12.569 | Voice/Question | Moderate accuracy |
| Text2Text | 11.234 | 13.788 | Text/Answer | Requires improvement |
To verify model learning effectiveness, we visualized attention patterns:
Our analysis revealed important insights about the training data:
We identified specific challenges related to Korean language processing:
The final system integrates eye-tracking technology for user interaction. When a question like "아픈 곳은 없어요?" (Do you have any pain?) is processed, the system generates contextually appropriate response options that patients can select through eye movements.
Development Team
Research Team
SpeAKN represents a significant advancement in assistive technology for ALS patients by:
The eye-tracking implementation is available as an open-source project: https://github.com/junhyk-lee/Look_to_Speak
This work contributes to the growing body of research in assistive technology and demonstrates the potential of AI-powered solutions for improving quality of life for patients with neurodegenerative diseases.