Developing non-invasive BCI and computer vision technologies to empower individuals with disabilities — enabling control, communication, and independence through the power of thought.
Train an ML model to read electrical signals from a non-invasive BCI. This allows a user to control a motorized wheelchair with their thoughts — enabling highly disabled patients with little or no limb movement to navigate independently.
View Wheelchair ProgressDevelop a computer vision system that translates American Sign Language into text or speech, facilitating communication for those who are hard of hearing or deaf.
Utilize EEG signals to train a machine learning model that converts brain activity into text, potentially enabling non-verbal communication for millions with severe motor disabilities.
View Thought to Text ProgressThrough these cross-disciplinary projects, we aim to design, build, and test brain-computer interface (BCI) and computer vision technologies that empower individuals with disabilities. By enabling intuitive wheelchair control, translating American Sign Language to text or speech, and converting brain signals into text, we aspire to improve mobility, communication, and independence for those with physical and communicative challenges — ultimately advancing their quality of life and contributing to impactful healthcare solutions.