Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action

Detalhes bibliográficos
Autor(a) principal: Silva, Rui
Data de Publicação: 2016
Outros Autores: Louro, Luís, Malheiro, Tiago, Erlhagen, Wolfram, Bicho, Estela
Tipo de documento: Artigo
Idioma: eng
Título da fonte: Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
Texto Completo: http://hdl.handle.net/1822/45420
Resumo: We report on our approach towards creating socially intelligent robots, which is heavily inspired by recent experimental findings about the neurocognitive mechanisms underlying action and emotion understanding in humans. Our approach uses neuro-dynamics as a theoretical language to model cognition, emotional states, decision making and action. The control architecture is formalized by a coupled system of dynamic neural fields representing a distributed network of local but connected neural populations. Different pools of neurons encode relevant information in the form of self-sustained activation patterns, which are triggered by input from connected populations and evolve continuously in time. The architecture implements a dynamic and flexible context-dependent mapping from observed hand and facial actions of the human onto adequate complementary behaviors of the robot that take into account the inferred goal and inferred emotional state of the co-actor. The dynamic control architecture was validated in multiple scenarios in which an anthropomorphic robot and a human operator assemble a toy object from its components. The scenarios focus on the robot’s capacity to understand the human’s actions, and emotional states, detect errors and adapt its behavior accordingly by adjusting its decisions and movements during the execution of the task.
id RCAP_4378fcbe010d0813fa6699259bbb5ad5
oai_identifier_str oai:repositorium.sdum.uminho.pt:1822/45420
network_acronym_str RCAP
network_name_str Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
repository_id_str 7160
spelling Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint actionHuman-robot interactionsDyanmic field architectureGoal inferenceEmotional stateHuman-robot joint actionemotional state inferenceerror detectiondecision makingdynamic neural fieldsCiências Naturais::MatemáticasScience & TechnologySocial SciencesWe report on our approach towards creating socially intelligent robots, which is heavily inspired by recent experimental findings about the neurocognitive mechanisms underlying action and emotion understanding in humans. Our approach uses neuro-dynamics as a theoretical language to model cognition, emotional states, decision making and action. The control architecture is formalized by a coupled system of dynamic neural fields representing a distributed network of local but connected neural populations. Different pools of neurons encode relevant information in the form of self-sustained activation patterns, which are triggered by input from connected populations and evolve continuously in time. The architecture implements a dynamic and flexible context-dependent mapping from observed hand and facial actions of the human onto adequate complementary behaviors of the robot that take into account the inferred goal and inferred emotional state of the co-actor. The dynamic control architecture was validated in multiple scenarios in which an anthropomorphic robot and a human operator assemble a toy object from its components. The scenarios focus on the robot’s capacity to understand the human’s actions, and emotional states, detect errors and adapt its behavior accordingly by adjusting its decisions and movements during the execution of the task.The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work was possible in part by the funding of research grants from the Portuguese Foundation for Science and Technology (grant numbers SFRH/BD/48527/2008, SFRH/BPD/71874/2010, SFRH/BD/81334/2011), and with funding from FP6-IST2 EU-IP Project JAST (project number 003747) and FP7 Marie Curie ITN Neural Engineering Transformative Technologies NETT (project number 289146).info:eu-repo/semantics/publishedVersionSAGE PublicationsUniversidade do MinhoSilva, RuiLouro, LuísMalheiro, TiagoErlhagen, WolframBicho, Estela2016-10-102016-10-10T00:00:00Zinfo:eu-repo/semantics/publishedVersioninfo:eu-repo/semantics/articleapplication/pdfhttp://hdl.handle.net/1822/45420eng1059-71231741-263310.1177/1059712316665451info:eu-repo/semantics/openAccessreponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãoinstacron:RCAAP2023-07-21T12:42:29Zoai:repositorium.sdum.uminho.pt:1822/45420Portal AgregadorONGhttps://www.rcaap.pt/oai/openaireopendoar:71602024-03-19T19:39:44.600584Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informaçãofalse
dc.title.none.fl_str_mv Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
title Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
spellingShingle Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
Silva, Rui
Human-robot interactions
Dyanmic field architecture
Goal inference
Emotional state
Human-robot joint action
emotional state inference
error detection
decision making
dynamic neural fields
Ciências Naturais::Matemáticas
Science & Technology
Social Sciences
title_short Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
title_full Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
title_fullStr Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
title_full_unstemmed Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
title_sort Combining intention and emotional state inference in a dynamic neural field architecture for human-robot joint action
author Silva, Rui
author_facet Silva, Rui
Louro, Luís
Malheiro, Tiago
Erlhagen, Wolfram
Bicho, Estela
author_role author
author2 Louro, Luís
Malheiro, Tiago
Erlhagen, Wolfram
Bicho, Estela
author2_role author
author
author
author
dc.contributor.none.fl_str_mv Universidade do Minho
dc.contributor.author.fl_str_mv Silva, Rui
Louro, Luís
Malheiro, Tiago
Erlhagen, Wolfram
Bicho, Estela
dc.subject.por.fl_str_mv Human-robot interactions
Dyanmic field architecture
Goal inference
Emotional state
Human-robot joint action
emotional state inference
error detection
decision making
dynamic neural fields
Ciências Naturais::Matemáticas
Science & Technology
Social Sciences
topic Human-robot interactions
Dyanmic field architecture
Goal inference
Emotional state
Human-robot joint action
emotional state inference
error detection
decision making
dynamic neural fields
Ciências Naturais::Matemáticas
Science & Technology
Social Sciences
description We report on our approach towards creating socially intelligent robots, which is heavily inspired by recent experimental findings about the neurocognitive mechanisms underlying action and emotion understanding in humans. Our approach uses neuro-dynamics as a theoretical language to model cognition, emotional states, decision making and action. The control architecture is formalized by a coupled system of dynamic neural fields representing a distributed network of local but connected neural populations. Different pools of neurons encode relevant information in the form of self-sustained activation patterns, which are triggered by input from connected populations and evolve continuously in time. The architecture implements a dynamic and flexible context-dependent mapping from observed hand and facial actions of the human onto adequate complementary behaviors of the robot that take into account the inferred goal and inferred emotional state of the co-actor. The dynamic control architecture was validated in multiple scenarios in which an anthropomorphic robot and a human operator assemble a toy object from its components. The scenarios focus on the robot’s capacity to understand the human’s actions, and emotional states, detect errors and adapt its behavior accordingly by adjusting its decisions and movements during the execution of the task.
publishDate 2016
dc.date.none.fl_str_mv 2016-10-10
2016-10-10T00:00:00Z
dc.type.status.fl_str_mv info:eu-repo/semantics/publishedVersion
dc.type.driver.fl_str_mv info:eu-repo/semantics/article
format article
status_str publishedVersion
dc.identifier.uri.fl_str_mv http://hdl.handle.net/1822/45420
url http://hdl.handle.net/1822/45420
dc.language.iso.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv 1059-7123
1741-2633
10.1177/1059712316665451
dc.rights.driver.fl_str_mv info:eu-repo/semantics/openAccess
eu_rights_str_mv openAccess
dc.format.none.fl_str_mv application/pdf
dc.publisher.none.fl_str_mv SAGE Publications
publisher.none.fl_str_mv SAGE Publications
dc.source.none.fl_str_mv reponame:Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
instname:Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
instacron:RCAAP
instname_str Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
instacron_str RCAAP
institution RCAAP
reponame_str Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
collection Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos)
repository.name.fl_str_mv Repositório Científico de Acesso Aberto de Portugal (Repositórios Cientìficos) - Agência para a Sociedade do Conhecimento (UMIC) - FCT - Sociedade da Informação
repository.mail.fl_str_mv
_version_ 1799132940577275904