AVISPIRE

Audio-VIsual Speech Processing for Interaction in Realistic Environments

 Coordinatore "NATIONAL CENTER FOR SCIENTIFIC RESEARCH ""DEMOKRITOS""" 

 Organization address address: Patriarchou Gregoriou Str.
city: AGHIA PARASKEVI
postcode: 15310

contact info
Titolo: Ms.
Nome: Marina
Cognome: Fontara
Email: send email
Telefono: -6503276
Fax: -6503319

 Nazionalità Coordinatore Greece [EL]
 Totale costo 87˙500 €
 EC contributo 87˙500 €
 Programma FP7-PEOPLE
Specific programme "People" implementing the Seventh Framework Programme of the European Community for research, technological development and demonstration activities (2007 to 2013)
 Code Call FP7-PEOPLE-2009-RG
 Funding Scheme MC-IRG
 Anno di inizio 2009
 Periodo (anno-mese-giorno) 2009-10-01   -   2013-03-31

 Partecipanti

# participant  country  role  EC contrib. [€] 
1    "NATIONAL CENTER FOR SCIENTIFIC RESEARCH ""DEMOKRITOS"""

 Organization address address: Patriarchou Gregoriou Str.
city: AGHIA PARASKEVI
postcode: 15310

contact info
Titolo: Ms.
Nome: Marina
Cognome: Fontara
Email: send email
Telefono: -6503276
Fax: -6503319

EL (AGHIA PARASKEVI) coordinator 87˙500.00

Mappa


 Word cloud

Esplora la "nuvola delle parole (Word Cloud) per avere un'idea di massima del progetto.

life    recognition    speech    real    extracted    interaction    ideal    visual    environments    significant    voice    audio    human    acoustic    quality   

 Obiettivo del progetto (Objective)

'The topic of audio-visual speech processing has attracted significant interest over the past 15 years. Relevant research has been focusing on recruiting visual speech information, extracted from the speaker's mouth region, as a means to improve robustness of traditional, unimodal, acoustic-only based speech processing. Nevertheless, to-date, most work has been limited to ideal-case scenarios, where the visual data are of high-quality, typically of steady frontal head pose, high resolution, and uniform lighting, while the audio signal contains speech by a single subject, in most cases artificially contaminated by noise in order to demonstrate significant improvements in speech system performance. Obviously, these conditions remain far from unconstrained, multi-party human interaction, thus, not surprisingly, practical audio-visual speech systems have yet to be deployed in real life. In this proposal, we aim to work towards expanding the state-of-the-art from the ideal “toy” examples to realistic human-computer interaction in difficult environments like the office, the automobile, broadcast news, and during meetings. Successful audio-visual speech processing there requires progress beyond the state-of-the-art in processing and robust extraction of visual speech information, as well as its efficient fusion with the acoustic modality, due to the varying quality of the extracted stream information. We propose to study a number of speech technologies in such environments (e.g., speech recognition, activity detection, diarization, separation), which stand to benefit from multimodality. The envisaged work will span 42 months of activity, and is planned as a natural evolution of research efforts of the candidate, Dr. Gerasimos Potamianos, while at AT&T Labs and IBM Research in the US, to be conducted jointly with the host organization, the Institute of Informatics and Telecommunications at the National Center of Scientific Research, "Demokritos", in Athens, Greece.'

Introduzione (Teaser)

If computers could read lips just like humans doe, what techniques would then be required in order to effectively capture voice by using inexpensive equipment? This is the issue addressed by EU researchers working on improving speech recognition systems in distinguishing voice from multiple speakers under real-life conditions.

Altri progetti dello stesso programma (FP7-PEOPLE)

BUAFULL (2010)

Acoustics of friction under light loads

Read More  

GPCR EVOLUTION (2012)

Directed evolution of class B G-protein coupled receptors

Read More  

COGNITIONNET (2013)

Multilevel brain network analyses leading to improved therapeutics for cognitive impairments

Read More