HANDS-FREE VOICE-ENABLED INTERFACE TO WEB APPLICATIONS FOR SMART HOME ENVIRONMENTS
Project Description

Nowadays, it is becoming increasingly affordable to enhance the home environment with several automation schemes, allowing remote control of e.g., heating/cooling, communication, lighting, media, etc. Such smart home functionalities are essential for people with disabilities and the elderly, as they not only provide assistive control of important everyday functionalities, but may prove to be life-saving in case of emergency. However, smart home functionalities may become useless for people who need those most, if they cannot be accessed via a natural, easy to use, interface.

The central objective of LISTEN is to design and implement a complete system, including both the software and hardware components, enabling robust hands-free large-vocabulary voice-based access to Internet applications in smart homes. This would allow the users to have natural control (i.e., using their voice) of the smart-home web-enabled functionalities (e.g., turning on/off web-enabled “smart” appliances), but also to access specific Internet applications (e.g., web search, email dictation, access to social networks). A truly hands-free system operation of the voice interface is equally important: users will not have to turn towards a microphone or other device, or wear a headset.

Therefore, LISTEN will develop (a) a robust hands-free speech capture system operating as a wireless acoustic sensor network (WASN), specifically designed for the smart home, and (b) a large-vocabulary automatic speech recognition system optimised for accessing web applications and controlling web-enabled smart home automation functionalities. LISTEN pushes the boundaries of current state-of-the-art by bridging the gap between the acoustic front-end and automatic speech recognition research communities, with the common goal of developing a smart-home-specific natural voice interface to web services.

FORTH - Institute of Computer Science