Explore the words cloud of the MECHIDENT project. It provides you a very rough idea of what is the project "MECHIDENT" about.
The following table provides information about the project.
Coordinator |
UNIVERSITY OF NEWCASTLE UPON TYNE
Organization address contact info |
Coordinator Country | United Kingdom [UK] |
Total cost | 1˙995˙677 € |
EC max contribution | 1˙995˙677 € (100%) |
Programme |
1. H2020-EU.1.1. (EXCELLENT SCIENCE - European Research Council (ERC)) |
Code Call | ERC-2016-COG |
Funding Scheme | ERC-COG |
Starting year | 2017 |
Duration (year-month-day) | from 2017-07-01 to 2022-06-30 |
Take a look of project's partnership.
# | ||||
---|---|---|---|---|
1 | UNIVERSITY OF NEWCASTLE UPON TYNE | UK (NEWCASTLE UPON TYNE) | coordinator | 1˙995˙677.00 |
Our social interactions and survival critically depend on identifying specific individuals to interact with or avoid (“who is that?”). Identifying individuals can be achieved by different sensory inputs, and by many accounts any sensory input elicits a representation of an individual that somehow becomes transmodal or independent of any sensory system. However, how the brain achieves transmodal integration facilitating individual recognition remains a mystery: Investigations in humans allowing direct access to site-specific neuronal processes are generally rare and have not focused on understanding neuronal multisensory integration for person recognition. Also, animal models to study the neuronal mechanisms of related processes have only recently become known. I propose to use direct recordings of neuronal activity in both humans and monkeys during face- and voice-identification tasks, combined with site-specific manipulation of the sensory input streams into the lateral anterior temporal lobe (ATL). The ATL brings together identity-specific content from the senses but the neuronal mechanisms for this convergence are entirely unknown. My core hypothesis is that auditory voice- or visual face-identity input into key ATL convergence sites elicits a sensory-modality invariant representation, which once elicited is robust to degradation or inactivation of neuronal input from the other sense. The central aim is to test this in human patients being monitored for surgery and to directly compare and link the results with those in monkeys where the neuronal circuit and mechanisms can be revealed using optogenetic control of neuronal responses. Analyses will assess neuronal dynamics and sensory integration frameworks. This proposal is poised to unravel how the brain combines multisensory input critical for identifying individuals and cognitive operations to act upon. The basic science insights gained may inform efforts to stratify patients with different types of ATL damage.
year | authors and title | journal | last update |
---|---|---|---|
2018 |
Oya, et al, Petkov, et al Neural phase locking predicts BOLD response in human auditory cortex published pages: , ISSN: 1053-8119, DOI: |
Neuroimage | 2019-05-16 |
2019 |
Wikman, Rinne & Petkov Reward cues readily direct monkeys’ auditory attention resulting in broad auditory cortex modulation and interaction with sites along cholinergic and dopaminergic pathways published pages: , ISSN: 2045-2322, DOI: |
Nature Scientific Reports | 2019-05-16 |
2018 |
Mitchell, Thiele, Petkov, Roberts, Robbins, Schultz and Lemon Continued need for non-human primate neuroscience research published pages: , ISSN: 0960-9822, DOI: |
Current Biology | 2019-05-16 |
2017 |
Oya, et al, Petkov, et al Mapping effective connectivity in the human brain with concurrent intracranial electrical stimulation and BOLD-fMRI published pages: , ISSN: 0165-0270, DOI: |
Journal of Neuroscience Methods | 2019-05-16 |
2018 |
Milham, et al, Petkov, Nacef, et al An Open Resource for Non-human Primate Imaging published pages: , ISSN: 0896-6273, DOI: |
Neuron | 2019-05-16 |
Are you the coordinator (or a participant) of this project? Plaese send me more information about the "MECHIDENT" project.
For instance: the website url (it has not provided by EU-opendata yet), the logo, a more detailed description of the project (in plain text as a rtf file or a word file), some pictures (as picture files, not embedded into any word file), twitter account, linkedin page, etc.
Send me an email (fabio@fabiodisconzi.com) and I put them in your project's page as son as possible.
Thanks. And then put a link of this page into your project's website.
The information about "MECHIDENT" are provided by the European Opendata Portal: CORDIS opendata.