Research Goals

Identification of Activities

We investigate the types of activities that Haru can support in various learning environments (e.g., children distracted during class; an adolescent learning about healthy eating behaviours). We believe that some behaviours can be improved more readily than others with Haru.

Haru4Kids at home illustration

Investigation of Data Gathering

Haru4Kids at home illustration

We investigate how we can gather information about a person's behaviour. Which type of information could be gathered via Haru’s future functions (e.g., audio, visual, sentiment analysis). We aim to equip Haru with multisensory functions for deep human understanding.

Investigation of Data Output

Once we gather a person's information, our goal shifts to data output. Focusing on efficient interaction between Haru and a person, we can use the collected data to allow for exploration that could improve behviour. Questions such as "How do we deliver information to people so their behaviours can be improved effectively over time?" "Can we deliver data so anyone can benefit from it?" Our current plan is to a apply data video approach for this investigation. Our study confirmed that the way we deliver messages in a video can influence the user’s perception (Sallam et al., 2022). We plan to apply this approach to deliver information to our users.

Haru4Kids at home illustration

Deep Human Understanding for Education 4.0 

Haru4Kids at home illustration

For deeper human understanding in school settings, Haru will be gathering various types of information about children’s experiences. For example, EEG data will be captured to infer their potential focus, facial expressions will be captured to understand their emotion or social cues, and voice data will be also analyzed to understand their sentiment. We aim to deliver smooth communication between children and Haru.    

Haru Explored: Dialogues and Interactions

Haru focus areas

Haru has great potential, such as engaging in conversations, recognizing and expressing emotions, ensuring safety in interactions, detecting others' facial expressions, and presenting data visually. 

Video Insights 

Related Publications


Parent and Educator Concerns on the Pedagogical Use of AI-Equipped Social Robots

Francisco Perella-Holfeld, Samar Sallam, Julia Petrie, Randy Gomez, Pourang Irani, and Yumiko Sakamoto. (2024, February). Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies (In Submission). 

Exploring the Design of Social Robot User Interfaces for Presenting Data-Driven Stories 

Anuradha Herath, Samar Sallam, Yumiko Sakamoto, Randy Gomez, and Pourang Irani. (2023, December). Exploring the Design of Social Robot User Interfaces for Presenting Data-Driven Stories. In Proceedings of the 22nd International Conference on Mobile and Ubiquitous Multimedia (MUM '23). (pp. 321-339). Association for Computing Machinery. 

How Should a Social Robot Deliver Negative Feedback Without Creating Distance Between the Robot and Child Users? 

Yumiko Sakamoto, Anuradha Herath, Tanvi Vuradi, Samar Sallam, Randy Gomez and Pourang Irani. (2023, December). How Should a Social Robot Deliver Negative Feedback Without Creating Distance Between the Robot and Child Users? In Proceedings of the 11th International Conference on Human-Agent Interaction (HAI '23). (pp. 325-334). Association for Computing Machinery. 

Presenting Data with Social Robots: An Exploration into Conveying Data Videos using an Artificial Physical Narrator 

Anuradha Herath, Samar Sallam, Tanvi Vuradi, Yumiko Sakamoto, Randy Gomez and Pourang Irani. (2023, December). Presenting Data with Social Robots: An Exploration into Conveying Data Videos using an Artificial Physical Narrator (HAI '23). (pp. 476-478). Association for Computing Machinery. 

Design and Development of a Teleoperation System for Affective Tabletop Robot Haru 

Vasylkiv, Y., Ma, Z., Li, G., Brock, H., Nakamura, K., Pourang, I., & Gomez, R. (2021, August). Shaping Affective Robot Haru’s Reactive Response. In 2021 30th IEEE International Conference on Robot & Human Interactive Communication (RO-MAN) (pp. 989-996). IEEE. 

Shaping Affective Robot Haru’s Reactive Response  

Vasylkiv, Y., Ma, Z., Li, G., Brock, H., Nakamura, K., Pourang, I., & Gomez, R. (2021, August). Shaping Affective Robot Haru’s Reactive Response. In 2021 30th IEEE International Conference on Robot & Human Interactive Communication (RO-MAN) (pp. 989-996). IEEE.

Automating Behavior Selection for Affective Telepresence Robot

Vasylkiv, Y., Ma, Z., Li, G., Sandry, E., Brock, H., Nakamura, K., Pourang, I. and Gomez, R., 2021, May. Automating behavior selection for affective telepresence robot. In 2021 IEEE International Conference on Robotics and Automation (ICRA) (pp. 2026-2032). IEEE.

Persuasive Data Storytelling with a Data Video during Covid-19 Infodemic: Affective Pathway to Influence the Users’ Perception about Contact Tracing Apps in less than 6 Minutes   

Sakamoto Y, Sallam S, Leboe-McGowan J, Salo A, Irani P. Persuasive data storytelling with a data video during Covid-19 infodemic: Affective pathway to influence the users’ perception about contact tracing apps in less than 6 minutes. Proceedings of the 2022 Pacific Vis. 2022. 

Towards Design Guidelines for Effective Health-Related Data Videos: An Empirical Investigation of Affect, Personality, and Video Content

Sallam S, Sakamoto Y, Leboe-McGowan J, Latulipe C, Irani P. Towards Design Guidelines for Effective Health-Related Data Videos: An Empirical Investigation of Affect, Personality, and Video Content. Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems.

Persuasive Data Videos: Investigating Persuasive Self-Tracking Feedback with Augmented Data Videos 

Choe, E. K., Sakamoto, Y., Fatmi, Y., Lee, B., Hurter, C., Haghshenas, A., & Irani, P. (2019). Persuasive data videos: Investigating persuasive self-tracking feedback with augmented data videos. In AMIA Annual Symposium Proceedings (Vol. 2019, p. 295). American Medical Informatics Association.

Hooked on data videos: assessing the effect of animation and pictographs on viewer engagement 

Amini, F., Riche, N. H., Lee, B., Leboe-McGowan, J., & Irani, P. (2018, May). Hooked on data videos: assessing the effect of animation and pictographs on viewer engagement. In Proceedings of the 2018 International Conference on Advanced Visual Interfaces (pp. 1-9). 

Evaluating Data-Driven Stories and Storytelling Tools 

Amini, F., Brehmer, M., Bolduan, G., Elmer, C., & Wiederkehr, B. (2018). Evaluating data-driven stories and storytelling tools. In Data-driven storytelling (pp. 249-286). AK Peters/CRC Press  

Data Representations for In-Situ Exploration of Health and Fitness Data 

Amini, F., Hasan, K., Bunt, A., & Irani, P. (2017, May). Data representations for in-situ exploration of health and fitness data. In Proceedings of the 11th EAI international conference on pervasive computing technologies for healthcare (pp. 163-172). 

Authoring Data-Driven Videos with DataClips  

Amini, F., Riche, N. H., Lee, B., Monroy-Hernandez, A., & Irani, P. (2016). Authoring data-driven videos with dataclips. IEEE transactions on visualization and computer graphics, 23(1), 501-510. 

Understanding Data Videos: Looking at Narrative Visualization through the Cinematography Lens 

Amini, F., Henry Riche, N., Lee, B., Hurter, C., & Irani, P. (2015, April). Understanding data videos: Looking at narrative visualization through the cinematography lens. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (pp. 1459-1468). 

Collaborators

Hona Research Institute Mind Over Learning

Team Members

Pourang Irani

Pourang P. Irani

Professor
Principal's Research Chair in Ubiquitous Analytics
University of British Columbia (Okanagan campus)

Room No: 304
Charles E. Fipke Centre for Innovative Research
3247 University Way
Kelowna, BC V1V 1V7
Email: pourang.irani@ubc.ca

   Google Scholar dblp

I am a Professor in the Department of Computer Science at the University of British Columbia (Okanagan campus) and Principal's Research Chair in Ubiquitous Analytics. My research lies broadly in the areas of Human-Computer Interaction and Information Visualization. More specifically, our team is concentrating on designing and studying novel interactive systems for sensemaking "anywhere" and "anytime". For advancing this work we rely on mixed reality (MR) and wearable technologies for developing novel prototypes of visual interfaces and devices. I am also the Principal Investigator on an NSERC CREATE grant on Visual and Automated Disease Analytics. The aim of the VADA program is to train the next generation data scientists with a focus on health data analytics.

Yumiko Sakamoto
Dr. Yumiko Sakamoto
(Research Associate)
Eranga
Eranga De Saa
(MSc)
Tanvi
Tanvi Vuradi
(Co-op)
Francisco
Francisco Perella-Holfeld
(Co-op)
Nathania
Nathania Hendradjaja
(Honours)
Tiaan
Tiaan Boshoff
(Undergrad)
Julia
Julia Petrie
(Undergrad)
Vision