In a study, 18 kids, between the ages of 4 and 12, five of whom had autism, interacted with two robots which expressed 20 emotional states, including boredom, excitement, and nervousness. As children heard, saw, smelled, tasted, and touched in different scenarios, the robots showed them appropriate responses. The results were increased engagement when the robots engaged with the participants in sensory stations.
Join ApplySci at the 12th Wearable Tech + Digital Health + Neurotech Boston conference on November 14, 2019 at Harvard Medical School featuring talks by Brad Ringeisen, DARPA – Joe Wang, UCSD – Carlos Pena, FDA – George Church, Harvard – Diane Chan, MIT – Giovanni Traverso, Harvard | Brigham & Womens – Anupam Goel, UnitedHealthcare – Nathan Intrator, Tel Aviv University | Neurosteer – Arto Nurmikko, Brown – Constance Lehman, Harvard | MGH – Mikael Eliasson, Roche – David Rhew, Samsung
Join ApplySci at the 13th Wearable Tech + Neurotech + Digital Health Silicon Valley conference on February 11-12, 2020 at Stanford University featuring talks by Zhenan Bao, Stanford – Rudy Tanzi, Harvard – David Rhew, Samsung – Carla Pugh, Stanford – Nathan Intrator, Tel Aviv University | Neurosteer