<?xml version="1.0" encoding="UTF-8"?><xml><records><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>47</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Angel Fernandez, Julian M.</style></author><author><style face="normal" font="default" size="100%">Bonarini, Andrea</style></author><author><style face="normal" font="default" size="100%">Lola Cañamero</style></author></authors><secondary-authors><author><style face="normal" font="default" size="100%">Tapus, Adriana</style></author><author><style face="normal" font="default" size="100%">André, Elisabeth</style></author><author><style face="normal" font="default" size="100%">Martin, Jean-Claude</style></author><author><style face="normal" font="default" size="100%">Ferland, François</style></author><author><style face="normal" font="default" size="100%">Ammi, Mehdi</style></author></secondary-authors></contributors><titles><title><style face="normal" font="default" size="100%">A Reactive Competitive Emotion Selection System</style></title><secondary-title><style face="normal" font="default" size="100%">Proc. 7th International Conference on Social Robotics (ICSR-2015)</style></secondary-title><tertiary-title><style face="normal" font="default" size="100%">Lecture Notes in Computer Science</style></tertiary-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">Emotion production</style></keyword><keyword><style  face="normal" font="default" size="100%">Emotional models</style></keyword><keyword><style  face="normal" font="default" size="100%">Human Robot Interaction</style></keyword><keyword><style  face="normal" font="default" size="100%">Social robotics</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2015</style></year></dates><urls><web-urls><url><style face="normal" font="default" size="100%">https://link.springer.com/chapter/10.1007%2F978-3-319-25554-5_4</style></url></web-urls></urls><publisher><style face="normal" font="default" size="100%">Springer International Publishing</style></publisher><pub-location><style face="normal" font="default" size="100%">Paris</style></pub-location><pages><style face="normal" font="default" size="100%">31–40</style></pages><isbn><style face="normal" font="default" size="100%">978-3-319-25553-8</style></isbn><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">We present a reactive emotion selection system designed to be used in a robot that needs to respond autonomously to relevant events. A variety of emotion selection models based on &quot;cognitive appraisal&quot; theories exist, but the complexity of the concepts used by most of these models limits their use in robotics. Robots have physical constrains that condition their understanding of the world and limit their capacity to built the complex concepts needed for such models. The system presented in this paper was conceived to respond to &quot;disturbances&quot; detected in the environment through a stream of images, and use this low-level information to update emotion intensities. They are increased when specific patterns, based on Tomkins’ affect theory, are detected or reduced when it is not. This system could also be used as part of (or as first step in the incremental design of) a more cognitively complex emotional system for autonomous robots.</style></abstract><notes><style face="normal" font="default" size="100%">&lt;a href=&quot;https://link.springer.com/chapter/10.1007%2F978-3-319-25554-5_4&quot;&gt;Download&lt;/a&gt;</style></notes></record><record><source-app name="Biblio" version="7.x">Drupal-Biblio</source-app><ref-type>17</ref-type><contributors><authors><author><style face="normal" font="default" size="100%">Aryel Beck</style></author><author><style face="normal" font="default" size="100%">Lola Cañamero</style></author><author><style face="normal" font="default" size="100%">Antoine Hiolle</style></author><author><style face="normal" font="default" size="100%">Luisa Damiano</style></author><author><style face="normal" font="default" size="100%">Cosi, Piero</style></author><author><style face="normal" font="default" size="100%">Tesser, Fabio</style></author><author><style face="normal" font="default" size="100%">Sommavilla, Giacomo</style></author></authors></contributors><titles><title><style face="normal" font="default" size="100%">Interpretation of Emotional Body Language Displayed by a Humanoid Robot: A Case Study with Children</style></title><secondary-title><style face="normal" font="default" size="100%">International Journal of Social Robotics</style></secondary-title></titles><keywords><keyword><style  face="normal" font="default" size="100%">emotion</style></keyword><keyword><style  face="normal" font="default" size="100%">emotional body language</style></keyword><keyword><style  face="normal" font="default" size="100%">perception</style></keyword><keyword><style  face="normal" font="default" size="100%">Social robotics</style></keyword></keywords><dates><year><style  face="normal" font="default" size="100%">2013</style></year></dates><urls><web-urls><url><style face="normal" font="default" size="100%">https://link.springer.com/article/10.1007/s12369-013-0193-z</style></url></web-urls></urls><volume><style face="normal" font="default" size="100%">5</style></volume><pages><style face="normal" font="default" size="100%">325–334</style></pages><language><style face="normal" font="default" size="100%">eng</style></language><abstract><style face="normal" font="default" size="100%">The work reported in this paper focuses on giving humanoid robots the capacity to express emotions with their body. Previous results show that adults are able to interpret different key poses displayed by a humanoid robot and also that changing the head position affects the expressiveness of the key poses in a consistent way. Moving the head down leads to decreased arousal (the level of energy) and valence (positive or negative emotion) whereas moving the head up produces an increase along these dimensions. Hence, changing the head position during an interaction should send intuitive signals. The study reported in this paper tested children’s ability to recognize the emotional body language displayed by a humanoid robot. The results suggest that body postures and head position can be used to convey emotions during child-robot interaction.</style></abstract><notes><style face="normal" font="default" size="100%">&lt;a href=&quot;https://link.springer.com/article/10.1007/s12369-013-0193-z&quot;&gt;Download&lt;/a&gt;</style></notes></record></records></xml>