Main content block
Head of Group
Dr George Mylonas
About us
We use perceptual methods, AI, and frugal robotics innovation to deliver transformative diagnostic and treatment solutions.
- Youtube: HARMS Lab
Research lab info
What we do
The HARMS lab leverages perceptually enabled methodologies, artificial intelligence, and frugal innovation in robotics (such as soft surgical robots) to deliver transformative solutions for diagnosis and treatment. Our research is driven by both problem-solving and curiosity, aiming to build a comprehensive understanding of the actions, interactions, and reactions occurring in the operating room. We focus on using robotic technologies to facilitate procedures that are not yet widely adopted, particularly in endoluminal surgery, such as advanced treatments for gastrointestinal cancer.
Why it is important?
......
How can it benefit patients?
....
Meet the team
Results
- Showing results for:
- Reset all filters
Search results
-
Conference paperKogkas A, Darzi A, Mylonas GP, 2016,
Gaze-Driven Human-Robot Interaction in the Operating Theatre
, 6th Joint Workshop on New Technologies for Computer/Robot Assisted Surgery (CRAS 2016) -
Conference paperKhan DZ, Oude Vrielink TJC, Marcus H, et al., 2016,
NeuroCYCLOPS: development and preclinical validation of a robotic platform for endoscopic neurosurgery
, European Association of Neurosurgical Societies (EANS 2016), Publisher: European Association of Neurosurgical Societies -
Conference paperOude Vrielink TJC, Khan DZ, Marcus H, et al., 2016,
NeuroCYCLOPS: a novel system for endoscopic neurosurgery
, London, The Hamlyn Symposium on Medical Robotics, Publisher: Imperial College London, Pages: 36-37 -
Conference paperKogkas AA, Sodergren M, Darzi A, et al., 2016,
Macro- and micro-scale 3D gaze tracking in the operating theatre
, The Hamlyn Symposium on Medical Robotics 2016, Publisher: Imperial College London, Pages: 100-101 -
Journal articleLeff DR, James D, Orihuela-Espina F, et al., 2015,
The impact of expert visual guidance on trainee visual search strategy, visual attention and motor skills
, Frontiers in Human Neuroscience, Vol: 9, ISSN: 1662-5161Minimally invasive and robotic surgery changes the capacity for surgical mentors to guide their trainees with the control customary to open surgery. This neuroergonomic study aims to assess a “Collaborative Gaze Channel” (CGC); which detects trainer gaze-behaviour and displays the point of regard to the trainee. A randomised crossover study was conducted in which twenty subjects performed a simulated robotic surgical task necessitating collaboration either with verbal (control condition) or visual guidance with CGC (study condition). Trainee occipito-parietal (O-P) cortical function was assessed with optical topography (OT) and gaze-behaviour was evaluated using video-oculography. Performance during gaze-assistance was significantly superior [biopsy number: (mean ± SD): control=5·6±1·8 vs. CGC=6·6±2·0; p< 0.05] and was associated with significantly lower O-P cortical activity [∆HbO2 mMol x cm [median (IQR)] control = 2.5 (12.0) vs. CGC 0.63 (11.2), p < 0.001]. A random effect model confirmed the association between guidance mode and O-P excitation. Network cost and global efficiency and global efficiency were not significantly influenced by guidance mode. A gaze channel enhances performance, modulates visual search, and alleviates the burden in brain centres subserving visual attention and does not induce changes in the trainee's O-P functional network observable with the current OT technique. The results imply that through visual guidance, attentional resources may be liberated, potentially improving the capability trainees to attend to other safety critical events during the procedure.
-
PatentAvila Rencoret FB, Elson DS, Mylonas G, 2015,
Probe Deployment Device
-
Conference paperAvila-Rencoret FB, Elson DS, Mylonas G, 2015,
Towards a robotic-assisted cartography of the colon: a proof of concept
, Publisher: IEEE COMPUTER SOC, Pages: 1757-1763 -
Journal articlePaggetti G, Leff DR, Orihuela-Espina F, et al., 2014,
The role of the posterior parietal cortex in stereopsis and hand-eye coordination during motor task behaviours
, Cognitive Processing, Vol: 16, Pages: 177-190, ISSN: 1612-4790 -
Conference paperMylonas GP, Vitiello V, Cundy TP, et al., 2014,
CYCLOPS: A versatile robotic tool for bimanual single-access and natural-orifice endoscopic surgery
, IEEE International Conference on Robotics and Automation (ICRA), Publisher: IEEE, Pages: 2436-2442This paper introduces the CYCLOPS, a novel robotic tool for single-access and natural-orifice endoscopic surgery. Based on the concept of tendon-driven parallel robots, this highly original design gives the system some of its unique capabilities. Just to name a few, unparalleled force exertion capabilities of up to 65N, large and adjustable workspace, bimanual instrument triangulation. Due to the simplicity and nature of the design, the system could be adapted to an existing laparoscope or flexible endoscope. This promises a more immediate and accelerated route to clinical translation not only through endearing low-cost and adaptive features, but also by directly addressing several major barriers of existing designs.
-
Conference paperZhang L, Lee S-L, Yang G-Z, et al., 2014,
Semi-autonomous navigation for robot assisted tele-echography using generalized shape models and co-registered RGB-D cameras
, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2014), Publisher: IEEE, Pages: 3496-3502— This paper proposes a semi-autonomous navigatedmaster-slave system, for robot assisted remote echography forearly trauma assessment. Two RGB-D sensors are used tocapture real-time 3D information of the scene at the slave sidewhere the patient is located. A 3D statistical shape model isbuilt and used to generate a customized patient model basedon the point cloud generated by the RGB-D sensors. Thecustomized patient model can be updated and adaptively fittedto the patient. The model is also used to generate a trajectoryto navigate a KUKA robotic arm and safely conduct theultrasound examination. Extensive validation of the proposedsystem shows promising results in terms of accuracy androbustness.
This data is extracted from the Web of Science and reproduced under a licence from Thomson Reuters. You may not copy or re-distribute this data in whole or in part without the written consent of the Science business of Thomson Reuters.
Contact Us
General enquiries
hamlyn@imperial.ac.uk
Facility enquiries
hamlyn.facility@imperial.ac.uk
The Hamlyn Centre
Bessemer Building
South Kensington Campus
Imperial College
London, SW7 2AZ
Map location