Videos from the lab
User Modelling Using Multimodal Information for Dressing
Supplementary video for the Gao, Chang, and Demiris
Personalized Dressing using User Modeling in Latent Spaces
Supplementary video for the Zhang, Cully, and Demiris IROS2017 paper
Robots have the potential to provide tremendous support to disabled and elderly people in their everyday tasks, such as dressing. Many recent studies on robotic dressing assistance usually view dressing as a trajectory planning problem. However, the user movements during the dressing process are rarely taken into account, which often leads to the failures of the planned trajectory and may put the user at risk. The main difficulty of taking user movements into account is caused by severe occlusions created by the robot, the user, and the clothes during the dressing process, which prevent vision sensors from accurately detecting the postures of the user in real time. In this paper, we address this problem by introducing an approach that allows the robot to automatically adapt its motion according to the force applied on the robot's gripper caused by user movements. There are two main contributions introduced in this paper: 1) the use of a hierarchical multi-task control strategy to automatically adapt the robot motion and minimize the force applied between the user and the robot caused by user movements; 2) the online update of the dressing trajectory based on the user movement limitations modeled with the Gaussian Process Latent Variable Model in a latent space, and the density information extracted from such latent space. The combination of these two contributions leads to a personalized dressing assistance that can cope with unpredicted user movements during the dressing while constantly minimizing the force that the robot may apply on the user. The experimental results demonstrate that the proposed method allows the Baxter humanoid robot to provide personalized dressing assistance for human users with simulated upper-body impairments.
Authors: Fan Zhang, Antoine Cully, Yiannis Demiris
Attentional Network for Adaptive Visual Tracking
Supplementary video for the Choi et al. CVPR2017 paper
Title: Attentional Correlation Filter Network for Adaptive Visual Tracking
Authors: Jongwon Choi, Hyung Jin Chang, Sangdoo Yun, Tobias Fischer, Yiannis Demiris, and Jin Young Choi
Adaptive User Model in Car Racing Games
This video shows our framework for Adaptive User Modelling in Car Racing Games. It shows the sequent
This video shows our framework for Adaptive User Modelling in Car Racing Games. It shows the sequential steps of the model, the simulator as well as the steps carried out to implement the User Model.
Personalised Track Design in Car Racing Games
Video shows a short demo of the track changing algorithm that creates a personalised track according
Real-time adaptation of computer games’ content to the users’ skills and abilities can enhance the player’s engagement and immersion. Understanding of the user’s potential while playing is of high importance in order to allow the successful procedural generation of user-tailored content. We investigate how player models can be created in car racing games. Our user model uses a combination of data from unobtrusive sensors, while the user is playing a car racing simulator. It extracts features through machine learning techniques, which are then used to comprehend the user’s gameplay, by utilising the educational theoretical frameworks of the Concept of Flow and Zone of Proximal Development. The end result is to provide at a next stage a new track that fits to the user needs, which aids both the training of the driver and their engagement in the game. In order to validate that the system is designing personalised tracks, we associated the average performance from 41 users that played the game, with the difficulty factor of the generated track. In addition, the variation in paths of the implemented tracks between users provides a good indicator for the suitability of the system.
Conference: CIG 2016
Title: Personalised Track Design in Car Racing Games
Authors: Theodosis Georgiou and Yiannis Demiris
Supporting Article : https://spiral.imperial.ac.uk/handle/10044/1/39560
Multimodal Imitation Using Self-Learned Sensorimotor Repr.
Supplementary video for the Zambelli and Demiris IROS2016 paper
Although many tasks intrinsically involve multiple modalities, often only data from a single modality are used to improve complex robots acquisition of new skills. We present a method to equip robots with multimodal learning skills to achieve multimodal imitation on-the-fly on multiple concurrent task spaces, including vision, touch and proprioception, only using self-learned multimodal sensorimotor relations, without the need of solving inverse kinematic problems or explicit analytical models formulation. We evaluate the proposed method on a humanoid iCub robot learning to interact with a piano keyboard and imitating a human demonstration. Since no assumptions are made on the kinematic structure of the robot, the method can be also applied to different robotic platforms.
Authors: Martina Zambelli and Yiannis Demiris
Iterative Path Optimisation for Dressing Assistance
Supplementary video for the Gao, Chang, and Demiris IROS2016 paper
We propose an online iterative path optimisation method to enable a Baxter humanoid robot to assist human users to dress. The robot searches for the optimal personalised dressing path using vision and force sensor information: vision information is used to recognise the human pose and model the movement space of upper-body joints; force sensor information is used for the robot to detect external force resistance and to locally adjust its motion. We propose a new stochastic path optimisation method based on adaptive moment estimation. We first compare the proposed method with other path optimisation algorithms on synthetic data. Experimental results show that the performance of the method achieves the smallest error with fewer iterations and less computation time. We also evaluate real-world data by enabling the Baxter robot to assist real human users with their dressing.
Authors: Yixing Gao, Hyung Jin Chang, Yiannis Demiris
Kinematic Structure Correspondences via Hypergraph Matching
Supplementary video for the Chang, Fischer, Petit, Zambelli and Demiris CVPR2016 paper
In this paper, we present a novel framework for finding the kinematic structure correspondence between two objects in videos via hypergraph matching. In contrast to prior appearance and graph alignment based matching methods which have been applied among two similar static images, the proposed method finds correspondences between two dynamic kinematic structures of heterogeneous objects in videos.
Our main contributions can be summarised as follows:
(i) casting the kinematic structure correspondence problem into a hypergraph matching problem, incorporating multi-order similarities with normalising weights
(ii) structural topology similarity measure by a new topology constrained subgraph isomorphism aggregation
(iii) kinematic correlation measure between pairwise nodes
(iv) combinatorial local motion similarity measure using geodesic distance on the Riemannian manifold.
We demonstrate the robustness and accuracy of our method through a number of experiments on synthetic and real data, showing that various other methods are outperformed.
Authors: Hyung Jin Chang, Tobias Fischer, Maxime Petit, Martina Zambelli, Yiannis Demiris
Hierarchical Action Learning by Instruction
Supplementary video for the Petit and Demiris ICRA2016 paper
This video accompanies the paper titled "Hierarchical Action Learning by Instruction Through Interactive Grounding of Body Parts and Proto-actions" presented at IEEE International Conference on Robotics and Automation 2016.
Markerless Perspective Taking for Humanoid Robots
Supplementary video for the Fischer and Demiris ICRA2016 paper
Perspective taking enables humans to imagine the world from another viewpoint. This allows reasoning about the state of other agents, which in turn is used to more accurately predict their behavior. In this paper, we equip an iCub humanoid robot with the ability to perform visuospatial perspective taking (PT) using a single depth camera mounted above the robot. Our approach has the distinct benefit that the robot can be used in unconstrained environments, as opposed to previous works which employ marker-based motion capture systems. Prior to and during the PT, the iCub learns the environment, recognizes objects within the environment, and estimates the gaze of surrounding humans. We propose a new head pose estimation algorithm which shows a performance boost by normalizing the depth data to be aligned with the human head. Inspired by psychological studies, we employ two separate mechanisms for the two different types of PT. We implement line of sight tracing to determine whether an object is visible to the humans (level 1 PT). For more complex PT tasks (level 2 PT), the acquired point cloud is mentally rotated, which allows algorithms to reason as if the input data was acquired from an egocentric perspective. We show that this can be used to better judge where object are in relation to the humans. The multifaceted improvements to the PT pipeline advance the state of the art, and move PT in robots to markerless, unconstrained environments.
One-shot Learning of Assistance by Demonstration
Supplementary video for our ROMAN 2015 paper
An emerging research problem in the field of assis-
tive robotics is the design of methodologies that allow robots to
provide human-like assistance to the users. Especially within the
rehabilitation domain, a grand challenge is to program a robot
to mimic the operation of an occupational therapist, intervening
with the user when necessary so as to improve the therapeutic
power of the assistive robotic system. We propose a method
to estimate assistance policies from expert demonstrations to
present human-like intervention during navigation in a powered
wheelchair setup. For this purpose, we constructed a setting,
where a human offers assistance to the user over a haptic
shared control system. The robot learns from human assistance
demonstrations while the user is actively driving the wheelchair
in an unconstrained environment. We train a Gaussian process
regression model to learn assistance commands given past and
current actions of the user and the state of the environment. The
results indicate that the model can estimate human assistance
after only a single demonstration, i.e. in one-shot, so that the
robot can help the user by selecting the appropriate assistance
in a human-like fashion.
Personalised Dressing Assistance by Humanoid Robots
Supplementary video for our IROS 2015 paper
Lifelong Augmentation of Multi Modal Streaming Memories
We provide a principled framework for the cumulative organisation of streaming autobiographical data
Many robotics algorithms can benefit from storing and recalling large amounts of accumulated sensorimotor and interaction data. We provide a principled framework for the cumulative organisation of streaming autobiographical data so that data can be continuously processed and augmented as the processing and reasoning abilities of the agent develops and further interactions with humans take place. As an example, we show how a kinematic structure learning algorithm reasons a-posteriori about the skeleton of a human hand. A partner can be asked to provide feedback about the augmented memories, which can in turn be supplied to the reasoning processes in order to adapt their parameters. We employ active, multi- modal remembering, so the robot as well as humans can gain insights of both the original and augmented memories. Our framework is capable of storing discrete and continuous data in real-time, and thus creates a full memory. The data can cover multiple modalities and several layers of abstraction (e.g. from raw sound signals over sentences to extracted meanings). We show a typical interaction with a human partner using an iCub humanoid robot. The framework is implemented in a platform-independent manner. In particular, we validate multi platform capabilities using the iCub, Baxter and NAO robots. We also provide an interface to cloud based services, which allow automatic annotation of episodes. Our framework is geared towards the developmental robotics community, as it 1) provides a variety of interfaces for other modules, 2) unifies previous works on autobiographical memory, and 3) is licensed as open source software.
Online Heterogeneous Ensemble Learning
Online Heterogeneous Ensemble Learning of Sensorimotor Contingencies from Motor Babbling
Online Heterogeneous Ensemble Learning of Sensorimotor Contingencies from Motor Babbling.
Unsupervised Complex Kinematic Structure Learning
Supplementary video of our CVPR 2015 paper
Supplementary video of CVPR 2015 paper: 'Unsupervised Learning of Complex Articulated Kinematic Structures combining Motion and Skeleton Information'.
Musical Human-Robot Collaboration with Baxter
This video shows our framework for adaptive musical human-robot collaboration
This video shows our framework for adaptive musical human-robot collaboration. Baxter is in charge of the drum accompaniment and is learning the preferences of the user, who is in charge of the melody. For more information read: Sarabia M, Lee K, Demiris Y (2015) Towards a Synchronised Grammars Framework for Adaptive Musical Human-Robot Collaboration, RO-MAN [to appear].
Assistive Robotic Technology for Hospital Patients
Junior spent a week keeping company many patients at the Chelsea & Westminster hospital
A NAO humanoid robot, Junior, spent a week keeping company many patients at the Chelsea & Westminster Hospital in one of the largest trials of its kind in the world. Our results show that patients really enjoyed interacting with the robot.
The Online Echo State Gaussian Process (OESGP)
A video demonstrating the Online Echo State Gaussian Process (OESGP) for temporal learning
iCub Learning Human Dance Structures for Imitation
The iCub shows off its dance moves
Kyuhwa Lee, Tae-Kyun Kim and Yiannis Demiris, "Learning Reusable Task Representations using Hierarchical Activity Grammars with Uncertainties", IEEE International Conference on Robotics and Automation (ICRA), St. Paul, USA, 2012
iCub Grasping Demonstration
A demonstration of the iCub grasping mechanism
Yanyu Su, Yan Wu, Kyuhwa Lee, Zhijiang Du, Yiannis Demiris, "Robust Grasping Mechanism for an Under-actuated Anthropomorphic Hand under Object Position Uncertainty", IEEE-RAS International Conference on Humanoid Robots, Osaka, Japan, 2012.
iCub playing the Theremin
The iCub humanoid robot plays one of the most difficult musical instruments
The iCub humanoid robot plays the Theremin, one of the most difficult musical instrument, in real-time.
ARTY Smart Wheelchair
Helping young children safely use a wheelchair
The Assistive Robotic Transport for Youngsters (ARTY) is a smart wheelchair designed to help young children with disabilities who are unable to safely use a regular powered wheelchair. It is our hope that ARTY will give users an opportunity to independently explore, learn and play.