Apr. 10, 2019
16:00
-17:00

Building 51 room 015

 

Dear members, students and guests of the Zlotowski center for Neuroscience,

 

Our next seminar will be held on Wednesday, April 10th, 16:00

 

Venue: Building 51 room 015

 

"Joint action and perception with humans and robots"

 

Presented by

Dr. Alessandra Sciutti

 

Istituto Italiano di Tecnologia – Center for Human Technologies

Genova, Italy

 

Abstract:

Humans show a great natural ability at interacting with each other. Such efficiency in joint actions depends on a synergy between planned collaboration and emergent coordination, a subconscious mechanism based on a tight link between action execution and perception. This link supports phenomena as mutual adaptation, synchronization, and anticipation, which cut drastically the delays in the interaction and the need of complex verbal instructions and result in the establishment of joint intentions, the backbone of social interaction. From a neurophysiological perspective, this is possible, because the same neural system supporting action execution is responsible of the understanding and the anticipation of the observed action of others. Defining which human motion features allow for such emergent coordination with another agent would be crucial to establish more natural and efficient interaction paradigms with artificial devices, ranging from assistive and rehabilitative technology to companion robots. However, investigating the behavioral and neural mechanisms supporting natural interaction poses substantial problems. In particular, the unconscious processes at the basis of emergent coordination (e.g., unintentional movements or gazing) are very difficult—if not impossible—to restrain or control in a quantitative way for a human agent. Moreover, during an interaction, participants influence each other continuously in a complex way, resulting in behaviors that go beyond experimental control. Alessandra proposes robotics technology as a potential solution to this methodological problem. Robots indeed can establish an interaction with a human partner, contingently reacting to his actions without losing the controllability of the experiment or the naturalness of the interactive scenario. A robot could represent an “interactive probe" to assess the sensory and motor mechanisms underlying human–human interaction, as demonstrated by her research with the humanoid robot iCub, where an interactive humanoid robot represents a key tool to serve the investigation of the psychological and neuroscientific bases of social interaction.

 

 

Please note that the talk will begin at 16:00 exactly.

 

See you all,

 

Hava Golan, Seminar coordinator