Search or filter publications

Filter by type:

Filter by publication type

Filter by year:

to

Results

  • Showing results for:
  • Reset all filters

Search results

  • Journal article
    Ma J, Jiang M, Fang X, Chen J, Wang Y, Yang Get al., 2026,

    Hybrid aggregation strategy with double inverted residual blocks for lightweight salient object detection

    , NEURAL NETWORKS, Vol: 194, ISSN: 0893-6080
  • Journal article
    Zhang S, Nan Y, Fang Y, Wang S, Liu Y, Papanastasiou G, Gao Z, Li S, Walsh S, Yang Get al., 2026,

    Dynamical multi-order responses and global semantic-infused adversarial learning: A robust airway segmentation method.

    , Med Image Anal, Vol: 108

    Automated airway segmentation in computerized tomography (CT) images is crucial for the accurate diagnosis of lung diseases. However, the scarcity of manual annotations hinders the efficacy of supervised learning, while unconstrained intensities and sample imbalance lead to discontinuity and false-negative issues. To address these challenges, we propose a novel airway segmentation model named Dynamical Multi-order responses and Global Semantic-infused Adversarial network (DMGSA), integrating the unsupervised and supervised learning in parallel to alleviate the label scarcity of airway. In the unsupervised branch, (1) we propose several novel strategies of Dynamic Mask-Ratio (DMR) to empower the model to perceive context information of varying sizes, mimicking the laws of human learning vividly; (2) we present a novel target of Multi-Order Normalized Responses (MONR), exploiting the distinct order exponential operation of raw images and oriented gradients to enhance the textural representations of bronchioles; (3) we introduce the Adversarial Learning (AL) on the top of MONR module to discern nuances between real and fake images, focusing on capturing the textural features of terminal bronchioles. For the supervised branch, we propose an innovative Generalized Mean pooling based Global Semantic-infused (GMGS) module to ulteriorly improve the robustness. Ultimately, we have verified the method performance and robustness by training on normal lung disease datasets, while testing on lung cancer, COVID-19 and Lung fibrosis datasets. All experimental results have proved that our method exceeds state-of-the-art methods significantly.

  • Conference paper
    Hasan MK, Yang G, Yap CH, 2026,

    Motion-Enhanced Cardiac Anatomy Segmentation via an Insertable Temporal Attention Module

    , Pages: 143-153, ISSN: 0302-9743

    Cardiac anatomy segmentation is useful for clinical assessment of cardiac morphology to inform diagnosis and intervention. Deep learning (DL), especially with motion information, has improved segmentation accuracy. However, existing techniques for motion enhancement are not yet optimal, and they have high computational costs due to increased dimensionality or reduced robustness due to suboptimal approaches that use non-DL motion registration, non-attention models, or single-headed attention. They further have limited adaptability and are inconvenient for incorporation into existing networks where motion awareness is desired. Here, we propose a novel, computationally efficient Temporal Attention Module (TAM) that offers robust motion enhancement, modeled as a small, multi-headed, cross-temporal attention module. TAM’s uniqueness is that it is a lightweight, plug-and-play module that can be inserted into a broad range of segmentation networks (CNN-based, Transformer-based, or hybrid) for motion enhancement without requiring substantial changes in the network’s backbone. This feature enables high adaptability and ease of integration for enhancing both existing and future networks. Extensive experiments on multiple 2D and 3D cardiac ultrasound and MRI datasets confirm that TAM consistently improves segmentation across a range of networks while maintaining computational efficiency and improving on currently reported performance. The evidence demonstrates that it is a robust, generalizable solution for motion-awareness enhancement that is scalable (such as from 2D to 3D). The code is available at https://github.com/kamruleee51/TAM.

  • Journal article
    Khalique Z, Scott AD, Ferreira PF, Molto M, Nielles-Vallespin S, Pennell DJet al., 2025,

    Diffusion Tensor CMR Assessment of the Microstructural Response to Dobutamine Stress in Health and Comparison With Patients With Recovered Dilated Cardiomyopathy.

    , Circ Cardiovasc Imaging

    BACKGROUND: Contractile reserve assessment assesses myocardial performance and prognosis. The microstructural mechanisms that facilitate increased cardiac function have not been described, but can be studied using diffusion tensor cardiovascular magnetic resonance. Resting microstructural contractile function is characterized by reorientation of aggregated cardiomyocytes (sheetlets) from wall-parallel in diastole to a more wall-perpendicular configuration in systole, with the diffusion tensor cardiovascular magnetic resonance parameter E2A defining their orientation, and sheetlet mobility defining the angle through which they rotate. We used diffusion tensor cardiovascular magnetic resonance to identify the microstructural response to dobutamine stress in healthy volunteers and then compared with patients with recovered dilated cardiomyopathy (rDCM). METHODS: In this first-of-its-kind prospective observational study, 20 healthy volunteers and 32 patients with rDCM underwent diffusion tensor cardiovascular magnetic resonance at rest, during dobutamine, and on recovery. RESULTS: In healthy volunteers, both diastolic and systolic E2A increased with dobutamine stress (13±3° to 17±5°; P<0.001 and 59±11° to 65±7°; P=0.002). Sheetlet mobility remained unchanged (45±11° to 49±10°; P=0.19), but biphasic mean E2A increased (36±6° to 41±4°; P<0.001). In rDCM, diastolic E2A at rest was higher than in healthy volunteers (20±8° versus 13±3°, P<0.001), and sheetlet mobility was reduced (34±12° versus 45±11°; P<0.001). During dobutamine stress, rDCM diastolic and systolic E2A increased compared with rest (20±8° to 24±10°; P=0.001 and 54±13° to 63±11°; P=0.005). However, sheetlet mobility in patients with rDCM failed to increase with dobutamine to healthy levels (39±13° versus 49±

  • Journal article
    Jing P, Lee K, Zhang Z, Zhou H, Yuan Z, Gao Z, Zhu L, Papanastasiou G, Fang Y, Yang Get al., 2025,

    Reason like a radiologist: Chain-of-thought and reinforcement learning for verifiable report generation.

    , Med Image Anal, Vol: 109

    Radiology report generation is critical for efficiency, but current models often lack the structured reasoning of experts and the ability to explicitly ground findings in anatomical evidence, which limits clinical trust and explainability. This paper introduces BoxMed-RL, a unified training framework to generate spatially verifiable and explainable chest X-ray reports. BoxMed-RL advances chest X-ray report generation through two integrated phases: (1) Pretraining Phase. BoxMed-RL learns radiologist-like reasoning through medical concept learning and enforces spatial grounding with reinforcement learning. (2) Downstream Adapter Phase. Pretrained weights are frozen while a lightweight adapter ensures fluency and clinical credibility. Experiments on two widely used public benchmarks (MIMIC-CXR and IU X-Ray) demonstrate that BoxMed-RL achieves an average 7 % improvement in both METEOR and ROUGE-L metrics compared to state-of-the-art methods. An average 5 % improvement in large language model-based metrics further underscores BoxMed-RL's robustness in generating high-quality reports. Related code and training templates are publicly available at https://github.com/ayanglab/BoxMed-RL.

  • Journal article
    Wang F, Wang Z, Li Y, Lyu J, Qin C, Wang S, Guo K, Sun M, Huang M, Zhang H, Tanzer M, Li Q, Chen X, Huang J, Wu Y, Zhang H, Hamedani KA, Lyu Y, Sun L, Li Q, He T, Lan L, Yao Q, Xu Z, Xin B, Metaxas DN, Razizadeh N, Nabavi S, Yiasemis G, Teuwen J, Zhang Z, Wang S, Zhang C, Ennis DB, Xue Z, Hu C, Xu R, Oksuz I, Lyu D, Huang Y, Guo X, Hao R, Patel JH, Cai G, Chen B, Zhang Y, Hua S, Chen Z, Dou Q, Zhuang X, Tao Q, Bai W, Qin J, Wang H, Prieto C, Markl M, Young A, Li H, Hu X, Wu L, Qu X, Yang G, Wang Cet al., 2025,

    Towards Modality- and Sampling-Universal Learning Strategies for Accelerating Cardiovascular Imaging: Summary of the CMRxRecon2024 Challenge.

    , IEEE Trans Med Imaging, Vol: PP

    Cardiovascular health is vital to human well-being, and cardiac magnetic resonance (CMR) imaging is considered the clinical reference standard for diagnosing cardiovascular disease. However, its adoption is hindered by long scan times, complex contrasts, and inconsistent quality. While deep learning methods perform well on specific CMR imaging sequences, they often fail to generalize across modalities and sampling schemes. The lack of benchmarks for high-quality, fast CMR image reconstruction further limits technology comparison and adoption. The CMRxRecon2024 challenge, attracting over 200 teams from 18 countries, addressed these issues with two tasks: generalization to unseen modalities and robustness to diverse undersampling patterns. We introduced the largest public multi-modality CMR raw dataset, an open benchmarking platform, and shared code. Analysis of the best-performing solutions revealed that prompt-based adaptation and enhanced physics-driven consistency enabled strong cross-scenario performance. These findings establish principles for generalizable reconstruction models and advance clinically translatable AI in cardiovascular imaging.

  • Journal article
    Ma X, Tao Y, Zhang Z, Zhang Y, Wang X, Zhang S, Ji Z, Zhang Y, Chen Q, Yang Get al., 2025,

    Test-time generative augmentation for medical image segmentation.

    , Med Image Anal, Vol: 109

    Medical image segmentation is critical for clinical diagnosis, treatment planning, and monitoring, yet segmentation models often struggle with uncertainties stemming from occlusions, ambiguous boundaries, and variations in imaging devices. Traditional test-time augmentation (TTA) techniques typically rely on predefined geometric and photometric transformations, limiting their adaptability and effectiveness in complex medical scenarios. In this study, we introduced Test-Time Generative Augmentation (TTGA), a novel augmentation strategy specifically tailored for medical image segmentation at inference time. Different from conventional augmentation strategies that suffer from excessive randomness or limited flexibility, TTGA leverages a domain-fine-tuned generative model to produce contextually relevant and diverse augmentations tailored to the characteristics of each test image. Built upon diffusion model inversion, a masked null-text inversion method is proposed to enable region-specific augmentations during sampling. Furthermore, a dual denoising pathway is designed to balance precise identity preservation with controlled variability. We demonstrate the efficacy of our TTGA through extensive experiments across three distinct segmentation tasks spanning nine datasets. Our results consistently demonstrate that TTGA not only improves segmentation accuracy (with DSC gains ranging from 0.1 % to 2.3 % over the baseline) but also offers pixel-wise error estimation (with DSC gains ranging from 1.1 % to 29.0 % over the baseline). The source code and demonstration are available at: https://github.com/maxiao0234/TTGA.

  • Journal article
    Jin W, Tian X, Wang N, Wu B, Shi B, Zhao B, Yang Get al., 2025,

    Representation-driven sampling and adaptive policy resetting for improving multi-Agent reinforcement learning

    , NEURAL NETWORKS, Vol: 192, ISSN: 0893-6080
  • Journal article
    Hao P, Wang H, Yang G, Zhu Let al., 2025,

    Enhancing Visual Reasoning With LLM-Powered Knowledge Graphs for Visual Question Localized-Answering in Robotic Surgery.

    , IEEE J Biomed Health Inform, Vol: 29, Pages: 9027-9040

    Expert surgeons often have heavy workloads and cannot promptly respond to queries from medical students and junior doctors about surgical procedures. Thus, research on Visual Question Localized-Answering in Surgery (Surgical-VQLA) is essential to assist medical students and junior doctors in understanding surgical scenarios. Surgical-VQLA aims to generate accurate answers and locate relevant areas in the surgical scene, requiring models to identify and understand surgical instruments, operative organs, and procedures. A key issue is the model's ability to accurately distinguish surgical instruments. Current Surgical-VQLA models rely primarily on sparse textual information, limiting their visual reasoning capabilities. To address this issue, we propose a framework called Enhancing Visual Reasoning with LLM-Powered Knowledge Graphs (EnVR-LPKG) for the Surgical-VQLA task. This framework enhances the model's understanding of the surgical scenario by utilizing knowledge graphs of surgical instruments constructed by the Large Language Model (LLM). Specifically, we design a Fine-grained Knowledge Extractor (FKE) to extract the most relevant information from knowledge graphs and perform contrastive learning with the extracted knowledge graphs and local image. Furthermore, we design a Multi-attention-based Surgical Instrument Enhancer (MSIE) module, which employs knowledge graphs to obtain an enhanced representation of the corresponding surgical instrument in the global scene. Through the MSIE module, the model can learn how to fuse visual features with knowledge graph text features, thereby strengthening the understanding of surgical instruments and further improving visual reasoning capabilities. Extensive experimental results on the EndoVis-17-VQLA and EndoVis-18-VQLA datasets demonstrate that our proposed method outperforms other state-of-the-art methods. We will release our code for future research.

  • Journal article
    Wang Z, Xiao M, Zhou Y, Wang C, Wu N, Li Y, Gong Y, Chang S, Chen Y, Zhu L, Zhou J, Cai C, Wang H, Jiang X, Guo D, Yang G, Qu Xet al., 2025,

    Deep Separable Spatiotemporal Learning for Fast Dynamic Cardiac MRI.

    , IEEE Trans Biomed Eng, Vol: 72, Pages: 3642-3654

    OBJECTIVE: Dynamic magnetic resonance imaging (MRI) plays an indispensable role in cardiac diagnosis. To enable fast imaging, the k-space data can be undersampled but the image reconstruction poses a great challenge of high-dimensional processing. This challenge necessitates extensive training data in deep learning reconstruction methods. In this work, we propose a novel and efficient approach, leveraging a dimension-reduced separable learning scheme that can perform exceptionally well even with highly limited training data. METHODS: We design this new approach by incorporating spatiotemporal priors into the development of a Deep Separable Spatiotemporal Learning network (DeepSSL), which unrolls an iteration process of a 2D spatiotemporal reconstruction model with both temporal low-rankness and spatial sparsity. Intermediate outputs can also be visualized to provide insights into the network behavior and enhance interpretability. RESULTS: Extensive results on cardiac cine datasets demonstrate that the proposed DeepSSL surpasses state-of-the-art methods both visually and quantitatively, while reducing the demand for training cases by up to 75%. Additionally, its preliminary adaptability to unseen cardiac patients has been verified through a blind reader study conducted by experienced radiologists and cardiologists. Furthermore, DeepSSL enhances the accuracy of the downstream task of cardiac segmentation and exhibits robustness in prospectively undersampled real-time cardiac MRI. CONCLUSION: DeepSSL is efficient under highly limited training data and adaptive to patients and prospective undersampling. SIGNIFICANCE: This approach holds promise in addressing the escalating demand for high-dimensional data reconstruction in MRI applications.

This data is extracted from the Web of Science and reproduced under a licence from Thomson Reuters. You may not copy or re-distribute this data in whole or in part without the written consent of the Science business of Thomson Reuters.

Request URL: http://www.imperial.ac.uk:80/respub/WEB-INF/jsp/search-t4-html.jsp Request URI: /respub/WEB-INF/jsp/search-t4-html.jsp Query String: id=1107&limit=10&resgrpMemberPubs=true&respub-action=search.html Current Millis: 1766959042368 Current Time: Sun Dec 28 21:57:22 GMT 2025

Contact


For enquiries about the MRI Physics Collective, please contact:

Mary Finnegan
Senior MR Physicist at the Imperial College Healthcare NHS Trust

Pete Lally
Assistant Professor in Magnetic Resonance (MR) Physics at Imperial College

Jan Sedlacik
MR Physicist at the Robert Steiner MR Unit, Hammersmith Hospital Campus