Last semester I took Dr. Chernova‘s “Human-Robot Interaction”. And got the chance to beta test this cute little robot:
Cozmo has a very simple concept of an OLED display for his face that is able to express emotions from animation of his eyes. The guys at Anki gave us a basic API in python to be able to communicate with the robot and trigger some acctions, modify the behaviour and getting access to the camera and other sensors like accelerometer and gyro.
I wanted to explore an emotion based approach for driving the actions of the robot and how a robotic assistant could help to treat body focus repetitive behaviors. I defined a two axes temperament scale that defines the emotion of the robot according to the robot state in this x-y coordinate system.
I trained the robot to detect when a person is biting her nails and use that signal as the input to a system dynamics that drives the robot emotions. Then, using the emotions to trigger actions from the robot I studied how our cute friend Cozmo could help for treating the nail biting behaviour. You can check out how this set up work in this video:
You can find more about this in the paper: