Healthcare
❯
AI-powered wearables
❯Patient Assistance
❯Facilitating language learning of deaf people
Dialogue-based social care services for people with mental illness, dementia and the elderly living alone
For:
Government or public institutions
Scope:
Daily life support AI services that provide an interaction with humans using natural language
Goal:
Improved Customer Experience
AI situation explanation service for people with visual impairments
For:
Personal services
Scope:
A real-time situation explanation service through voice for people with visual impairments
Goal:
Improved Customer Experience
AI sign language interpretation system for people with hearing impairment
For:
Government or public institutions
Scope:
Increase the convenience of public services to people with hearing impairment
by providing a service to translate sign language image information into natural
language
Goal:
Improved Customer Experience
Facilitating language learning of deaf people
Scope:
Use of advanced and multimodal sensing ability to facilitate a complex task.
Goal:
Improved Customer Experience
AI based text to speech services with personal voices for people with speech impairments
For:
People with speech impairments
Scope:
All people who have some sort of speech impairment including but not limited to three basic types: articulation disorders, fluency disorders, and voice disorders.
Goal:
Improved Customer Experience
Facilitating language learning of deaf people
Goal:
Improved Customer ExperienceProblem addressed
An avatar and social robot interact with deaf babies for facilitating language
learning.
Scope of use case
Use of advanced and multimodal sensing ability to facilitate a complex task.
Description
The RAVE system is designed as a dual agent that uses a
physical robot and a virtual human to engage 6 m to 12 m old
deaf infants in linguistic interactions. The system was
bolstered by a perception system capable of estimating
infant attention and engagement through thermal imaging
and eye tracking. RAVE has been designed for and
experienced by a unique population (deaf infants) during a
three years of observation and the development of three case
studies.
This system has been successful at soliciting infant attention,
directing attention to the linguistic content, and keeping the
infant engaged for developmentally appropriate lengths of
time. Instances have also been observed of infants copying
robot behaviour, infants producing signs displayed by the
avatar, and infants producing signs that they had observed
the virtual human perform to the non-signing robot agent.
These initial experiences give hope that longer-term
exposure to a system based on this work may be able to
impact long-term learning in this unique population.
Live Video
Audio
AI: Perceive
pattern recognition
AI: Understand