Opendata, web and dolomites

MultisensoryIntegration

Multisensory Integration in Time and Space

Total Cost €

0

EC-Contrib. €

0

Partnership

0

Views

0

 MultisensoryIntegration project word cloud

Explore the words cloud of the MultisensoryIntegration project. It provides you a very rough idea of what is the project "MultisensoryIntegration" about.

combined    little    mechanisms    single    mouse    unprecedented    audiovisual    sync    multimodal    poorly    hears    operations    object    disorder    repertoire    footsteps    matteo    regions    techniques    cortical    record    underlying    move    majority    sees    neural    animals    medical    answer    responsible    time    environments    deficits    manipulate    recording    roles    sources    lip    natural    multisensory    constantly    correlations    bombarded    ethological    danger    localized    opportunity    watching    space    harris    technological    separate    fundamental    rabbit    schizophrenia    question    cortex    dubbed    believe    perturbing    significance    populations    physical    laboratory    integration    college    decision    despite    kenneth    movements    painfully    behaviors    autistic    types    neurons    movie    friend    humans    experiments    brain    words    cues    perform    temporal    spatial    spectrum    calcium    individual    provides    optogenetic    professors    behavioral    continually    led    university    becomes    psychological    apparent    carandini    blur    subconscious    london    imaging   

Project "MultisensoryIntegration" data sheet

The following table provides information about the project.

Coordinator
UNIVERSITY COLLEGE LONDON 

Organization address
address: GOWER STREET
city: LONDON
postcode: WC1E 6BT
website: n.a.

contact info
title: n.a.
name: n.a.
surname: n.a.
function: n.a.
email: n.a.
telephone: n.a.
fax: n.a.

 Coordinator Country United Kingdom [UK]
 Project website https://www.ucl.ac.uk/cortexlab
 Total cost 183˙454 €
 EC max contribution 183˙454 € (100%)
 Programme 1. H2020-EU.1.3.2. (Nurturing excellence by means of cross-border and cross-sector mobility)
 Code Call H2020-MSCA-IF-2015
 Funding Scheme MSCA-IF-EF-ST
 Starting year 2016
 Duration (year-month-day) from 2016-03-11   to  2018-03-10

 Partnership

Take a look of project's partnership.

# participants  country  role  EC contrib. [€] 
1    UNIVERSITY COLLEGE LONDON UK (LONDON) coordinator 183˙454.00

Map

 Project objective

In natural environments, the brain is typically bombarded with multisensory information. Animals must continually evaluate which of these multimodal cues should be associated with a single physical object, and which are separate. For example, when a rabbit hears footsteps and sees a blur, is the blur responsible for the footsteps, or are there two separate sources of danger? Animals use two types of information to make this decision: spatial correlations (are the footsteps and blur localized together) and temporal correlations (does the blur move in sync with the footsteps). Humans constantly perform these same operations. We use lip movements to help identify the words spoken by a friend, a subconscious process which becomes painfully apparent when watching a poorly dubbed movie. This may explain why psychological conditions such as autistic spectrum disorder and schizophrenia often present with deficits in audiovisual integration.

Despite the ethological and medical significance of multisensory integration, little is known about the underlying neural mechanisms. However, I believe the behavioral repertoire of the mouse, combined with new technological advances, provides an unprecedented opportunity to answer this fundamental research question. With recent developments in calcium imaging and optogenetic techniques, it is now possible to record and manipulate neural activity across the majority of mouse cortex. I will develop novel behavioral tasks for the mouse which require multisensory integration in space or time. By recording or perturbing cortical activity during these behaviors, I will determine the roles of individual neurons, neural populations, and cortical regions in multisensory integration.

I will perform these experiments in the Cortical Processing Laboratory at University College London, led by Professors Kenneth Harris and Matteo Carandini.

Are you the coordinator (or a participant) of this project? Plaese send me more information about the "MULTISENSORYINTEGRATION" project.

For instance: the website url (it has not provided by EU-opendata yet), the logo, a more detailed description of the project (in plain text as a rtf file or a word file), some pictures (as picture files, not embedded into any word file), twitter account, linkedin page, etc.

Send me an  email (fabio@fabiodisconzi.com) and I put them in your project's page as son as possible.

Thanks. And then put a link of this page into your project's website.

The information about "MULTISENSORYINTEGRATION" are provided by the European Opendata Portal: CORDIS opendata.

More projects from the same programme (H2020-EU.1.3.2.)

DEF2DEV (2019)

Identification of the mode of action of plant defensins during root development and plant defense responses.

Read More  

RipGEESE (2020)

Identifying the ripples of gene regulation evolution in the evolution of gene sequences to determine when animal nervous systems evolved

Read More  

GrowthDevStability (2020)

Characterization of the developmental mechanisms ensuring a robust symmetrical growth in the bilateral model organism Drosophila melanogaster

Read More