All Stories

  1. Expressive Furhat: Generating Real-Time Facial Expressions for Human-Robot Dialogue with LLMs
  2. Multimodal Large Language Models for Real-Time Situated Reasoning
  3. Concerns and Values in Human-Robot Interactions: A Focus on Social Robotics
  4. Do Robots Influence Our Decisions When They Disagree With Us?
  5. I Was Blind but Now I See: Implementing Vision-Enabled Dialogue in Social Robots
  6. Values in Social Robots: Implementing Inclusive, Value-Aware Human-Robot Interactions
  7. “Can You be my Mum?”: Manipulating Social Robots in the Large Language Models Era
  8. Vision Language Models as Values Detectors
  9. AwarePrompt: Using Diffusion Models to Create Methods for Measuring Value-Aware AI Architectures
  10. Towards a Definition of Awareness for Embodied AI
  11. Social Value Alignment in Large Language Models
  12. Assessing LLMs Responses in the Field of Domestic Sustainability: An Exploratory Study
  13. Enhancing Conversational Troubleshooting with Multi-modality: Design and Implementation
  14. MCTK: a Multi-modal Conversational Troubleshooting Kit for supporting users in web applications