Imperial College London

Dr Ben Glocker

Faculty of EngineeringDepartment of Computing

Professor in Machine Learning for Imaging
 
 
 
//

Contact

 

+44 (0)20 7594 8334b.glocker Website CV

 
 
//

Location

 

377Huxley BuildingSouth Kensington Campus

//

Summary

 

Publications

Publication Type
Year
to

352 results found

Lee M, Oktay O, Schuh A, Schaap M, Glocker Bet al., 2019, Image-and-spatial transformer networks for structure-guided image registration, International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Publisher: Springer Verlag, Pages: 337-345, ISSN: 0302-9743

mage registration with deep neural networks has become anactive field of research and exciting avenue for a long standing problem inmedical imaging. The goal is to learn a complex function that maps theappearance of input image pairs to parameters of a spatial transforma-tion in order to align corresponding anatomical structures. We argue andshow that the current direct, non-iterative approaches are sub-optimal,in particular if we seek accurate alignment of Structures-of-Interest (SoI).Information about SoI is often available at training time, for example,in form of segmentations or landmarks. We introduce a novel, genericframework, Image-and-Spatial Transformer Networks (ISTNs), to lever-age SoI information allowing us to learn new image representations thatare optimised for the downstream registration task. Thanks to these rep-resentations we can employ a test-specific, iterative refinement over thetransformation parameters which yields highly accurate registration evenwith very limited training data. Performance is demonstrated on pairwise3D brain registration and illustrative synthetic data.

Conference paper

Li Z, Kamnitsas K, Glocker B, 2019, Overfitting of neural nets under class imbalance: Analysis and improvements for segmentation, International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Publisher: Springer Verlag, Pages: 402-410, ISSN: 0302-9743

Overfitting in deep learning has been the focus of a num-ber of recent works, yet its exact impact on the behaviour of neuralnetworks is not well understood. This study analyzes overfitting by ex-amining how the distribution of logits alters in relation to how muchthe model overfits. Specifically, we find that when training with few datasamples, the distribution of logit activations when processing unseen testsamples of an under-represented class tends to shift towards and evenacross the decision boundary, while the over-represented class seems un-affected. In image segmentation, foreground samples are often heavilyunder-represented. We observe that sensitivity of the model drops asa result of overfitting, while precision remains mostly stable. Based onour analysis, we derive asymmetric modifications of existing loss func-tions and regularizers including a large margin loss, focal loss, adver-sarial training and mixup, which specifically aim at reducing the shiftobserved when embedding unseen samples of the under-represented class.We study the case of binary segmentation of brain tumor core and showthat our proposed simple modifications lead to significantly improvedsegmentation performance over the symmetric variants.

Conference paper

Castro DC, Tan J, Kainz B, Konukoglu E, Glocker Bet al., 2019, Morpho-MNIST: quantitative assessment and diagnostics for representation learning, Journal of Machine Learning Research, Vol: 20, Pages: 1-29, ISSN: 1532-4435

Revealing latent structure in data is an active field of research, havingintroduced exciting technologies such as variational autoencoders andadversarial networks, and is essential to push machine learning towardsunsupervised knowledge discovery. However, a major challenge is the lack ofsuitable benchmarks for an objective and quantitative evaluation of learnedrepresentations. To address this issue we introduce Morpho-MNIST, a frameworkthat aims to answer: "to what extent has my model learned to represent specificfactors of variation in the data?" We extend the popular MNIST dataset byadding a morphometric analysis enabling quantitative comparison of trainedmodels, identification of the roles of latent variables, and characterisationof sample diversity. We further propose a set of quantifiable perturbations toassess the performance of unsupervised and supervised methods on challengingtasks such as outlier detection and domain adaptation. Data and code areavailable at https://github.com/dccastro/Morpho-MNIST.

Journal article

Pawlowski N, Bhooshan S, Ballas N, Ciompi F, Glocker B, Drozdzal Met al., 2019, Needles in haystacks: On classifying tiny objects in large images, Publisher: arXiv

In some computer vision domains, such as medical or hyperspectral imaging, wecare about the classification of tiny objects in large images. However, mostConvolutional Neural Networks (CNNs) for image classification were developedand analyzed using biased datasets that contain large objects, most often, incentral image positions. To assess whether classical CNN architectures workwell for tiny object classification we build a comprehensive testbed containingtwo datasets: one derived from MNIST digits and other from histopathologyimages. This testbed allows us to perform controlled experiments to stress-testCNN architectures using a broad spectrum of signal-to-noise ratios. Ourobservations suggest that: (1) There exists a limit to signal-to-noise belowwhich CNNs fail to generalize and that this limit is affected by dataset size -more data leading to better performances; however, the amount of training datarequired for the model to generalize scales rapidly with the inverse of theobject-to-image ratio (2) in general, higher capacity models exhibit bettergeneralization; (3) when knowing the approximate object sizes, adaptingreceptive field is beneficial; and (4) for very small signal-to-noise ratio thechoice of global pooling operation affects optimization, whereas for relativelylarge signal-to-noise values, all tested global pooling operations exhibitsimilar performance.

Working paper

Sokooti H, Saygili G, Glocker B, Lelieveldt BPF, Staring Met al., 2019, Quantitative error prediction of medical image registration using regression forests, Medical Image Analysis, Vol: 56, Pages: 110-121, ISSN: 1361-8415

Predicting registration error can be useful for evaluation of registration procedures, which is important for the adoption of registration techniques in the clinic. In addition, quantitative error prediction can be helpful in improving the registration quality. The task of predicting registration error is demanding due to the lack of a ground truth in medical images. This paper proposes a new automatic method to predict the registration error in a quantitative manner, and is applied to chest CT scans. A random regression forest is utilized to predict the registration error locally. The forest is built with features related to the transformation model and features related to the dissimilarity after registration. The forest is trained and tested using manually annotated corresponding points between pairs of chest CT scans in two experiments: SPREAD (trained and tested on SPREAD) and inter-database (including three databases SPREAD, DIR-Lab-4DCT and DIR-Lab-COPDgene). The results show that the mean absolute errors of regression are 1.07  ±  1.86 and 1.76  ±  2.59 mm for the SPREAD and inter-database experiment, respectively. The overall accuracy of classification in three classes (correct, poor and wrong registration) is 90.7% and 75.4%, for SPREAD and inter-database respectively. The good performance of the proposed method enables important applications such as automatic quality control in large-scale image analysis.

Journal article

Pawlowski N, Glocker B, 2019, Is texture predictive for age and sex in brain MRI?, Publisher: arXiv

Deep learning builds the foundation for many medical image analysis taskswhere neuralnetworks are often designed to have a large receptive field toincorporate long spatialdependencies. Recent work has shown that largereceptive fields are not always necessaryfor computer vision tasks on naturalimages. We explore whether this translates to certainmedical imaging tasks suchas age and sex prediction from a T1-weighted brain MRI scans.

Working paper

Dou Q, Ouyang C, Chen C, Chen H, Glocker B, Zhuang X, Heng P-Aet al., 2019, PnP-AdaNet: plug-and-play adversarial domain adaptation network at unpaired cross-modality cardiac segmentation, IEEE Access, ISSN: 2169-3536

Deep convolutional networks have demonstrated the state-of-the-art performance on variouschallenging medical image processing tasks. Leveraging images from different modalities for the sameanalysis task holds large clinical benefits. However, the generalization capability of deep networks ontest data sampled from different distribution remains as a major challenge. In this paper, we propose aPnP-AdaNet(plug-and-play adversarial domain adaptation network) for adapting segmentation networksbetween different modalities of medical images, e.g., MRI and CT. We tackle the significant domain shift byaligning the feature spaces of source and target domains at multiple scales in an unsupervised manner. Withadversarial loss, we learn a domain adaptation module which flexibly replaces the early encoder layers of thesource network, and the higher layers are shared between two domains. We validate our domain adaptationmethod on cardiac segmentation in unpaired MRI and CT, with four different anatomical structures. Theaverage Dice achieved 63.9%, which is a significant recover from the complete failure (Dice score of13.2%) if we directly test a MRI segmentation network on CT data. In addition, our proposedPnP-AdaNetoutperforms many state-of-the-art unsupervised domain adaptation approaches on the same dataset. Theexperimental results with comprehensive ablation studies have demonstrated the excellent efficacy of ourproposed method for unsupervised cross-modality domain adaptation. Our code is publically available at:https://github.com/carrenD/Medical-Cross-Modality-Domain-Adaptation

Journal article

Zlocha M, Dou Q, Glocker B, 2019, Improving retinanet for CT lesion detection with dense masks from weak recist labels., International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Publisher: arXiv, Pages: 402-410, ISSN: 0302-9743

Accurate, automated lesion detection in Computed Tomography (CT) is animportant yet challenging task due to the large variation of lesion types,sizes, locations and appearances. Recent work on CT lesion detection employstwo-stage region proposal based methods trained with centroid or bounding-boxannotations. We propose a highly accurate and efficient one-stage lesiondetector, by re-designing a RetinaNet to meet the particular challenges inmedical imaging. Specifically, we optimize the anchor configurations using adifferential evolution search algorithm. For training, we leverage the responseevaluation criteria in solid tumors (RECIST) annotation which are measured inclinical routine. We incorporate dense masks from weak RECIST labels, obtainedautomatically using GrabCut, into the training objective, which in combinationwith other advancements yields new state-of-the-art performance. We evaluateour method on the public DeepLesion benchmark, consisting of 32,735 lesionsacross the body. Our one-stage detector achieves a sensitivity of 90.77% at 4false positives per image, significantly outperforming the best reportedmethods by over 5%.

Conference paper

Winzeck S, Mocking SJT, Bezerra R, Bouts MJRJ, McIntosh EC, Diwan I, Garg P, Chutinet A, Kimberly WT, Copen WA, Schaefer PW, Ay H, Singhal AB, Kamnitsas K, Glocker B, Sorensen AG, Wu Oet al., 2019, Ensemble of Convolutional Neural Networks Improves Automated Segmentation of Acute Ischemic Lesions Using Multiparametric Diffusion-Weighted MRI, AMERICAN JOURNAL OF NEURORADIOLOGY, Vol: 40, Pages: 938-945, ISSN: 0195-6108

Journal article

Walker I, Glocker B, 2019, Graph convolutional Gaussian processes, International Conference on Machine Learning (ICML), Publisher: PMLR, Pages: 6495-6504, ISSN: 2640-3498

We propose a novel Bayesian nonparametricmethod to learn translation-invariant relationshipson non-Euclidean domains. The resulting graphconvolutional Gaussian processes can be appliedto problems in machine learning for which theinput observations are functions with domains ongeneral graphs. The structure of these models al-lows for high dimensional inputs while retainingexpressibility, as is the case with convolutionalneural networks. We present applications of graphconvolutional Gaussian processes to images andtriangular meshes, demonstrating their versatilityand effectiveness, comparing favorably to existingmethods, despite being relatively simple models.

Conference paper

Folgoc LL, Castro DC, Tan J, Khanal B, Kamnitsas K, Walker I, Alansary A, Glocker Bet al., 2019, Controlling meshes via curvature: spin transformations for pose-invariant shape processing, International Conference on Information Processing in Medical Imaging (IPMI 2019), Publisher: Springer Verlag, Pages: 221-234, ISSN: 0302-9743

We investigate discrete spin transformations, a geometric framework tomanipulate surface meshes by controlling mean curvature. Applications includesurface fairing -- flowing a mesh onto say, a reference sphere -- and meshextrusion -- e.g., rebuilding a complex shape from a reference sphere andcurvature specification. Because they operate in curvature space, theseoperations can be conducted very stably across large deformations with no needfor remeshing. Spin transformations add to the algorithmic toolbox forpose-invariant shape analysis. Mathematically speaking, mean curvature is ashape invariant and in general fully characterizes closed shapes (together withthe metric). Computationally speaking, spin transformations make thatrelationship explicit. Our work expands on a discrete formulation of spintransformations. Like their smooth counterpart, discrete spin transformationsare naturally close to conformal (angle-preserving). This quasi-conformalitycan nevertheless be relaxed to satisfy the desired trade-off between areadistortion and angle preservation. We derive such constraints and propose aformulation in which they can be efficiently incorporated. The approach isshowcased on subcortical structures.

Conference paper

Tarroni G, Oktay O, Bai W, Schuh A, Suzuki H, Passerat-Palmbach J, de Marvao A, O'Regan D, Cook S, Glocker B, Matthews P, Rueckert Det al., 2019, Learning-based quality control for cardiac MR images, IEEE Transactions on Medical Imaging, Vol: 38, Pages: 1127-1138, ISSN: 0278-0062

The effectiveness of a cardiovascular magnetic resonance (CMR) scan depends on the ability of the operator to correctly tune the acquisition parameters to the subject being scanned and on the potential occurrence of imaging artefacts such as cardiac and respiratory motion. In the clinical practice, a quality control step is performed by visual assessment of the acquired images: however, this procedure is strongly operatordependent, cumbersome and sometimes incompatible with the time constraints in clinical settings and large-scale studies. We propose a fast, fully-automated, learning-based quality control pipeline for CMR images, specifically for short-axis image stacks. Our pipeline performs three important quality checks: 1) heart coverage estimation, 2) inter-slice motion detection, 3) image contrast estimation in the cardiac region. The pipeline uses a hybrid decision forest method - integrating both regression and structured classification models - to extract landmarks as well as probabilistic segmentation maps from both long- and short-axis images as a basis to perform the quality checks. The technique was tested on up to 3000 cases from the UK Biobank as well as on 100 cases from the UK Digital Heart Project, and validated against manual annotations and visual inspections performed by expert interpreters. The results show the capability of the proposed pipeline to correctly detect incomplete or corrupted scans (e.g. on UK Biobank, sensitivity and specificity respectively 88% and 99% for heart coverage estimation, 85% and 95% for motion detection), allowing their exclusion from the analysed dataset or the triggering of a new acquisition.

Journal article

Lavdas I, Glocker B, Rueckert D, Taylor SA, Aboagye EO, Rockall AGet al., 2019, Machine learning in whole-body MRI: experiences and challenges from an applied study using multicentre data, Clinical Radiology, Vol: 74, Pages: 346-356, ISSN: 0009-9260

Machine learning is now being increasingly employed in radiology to assist with tasks such as automatic lesion detection, segmentation, and characterisation. We are currently involved in an National Institute of Health Research (NIHR)-funded project, which aims to develop machine learning methods to improve the diagnostic performance and reduce the radiology reading time of whole-body magnetic resonance imaging (MRI) scans, in patients being staged for cancer (MALIBO study). We describe here the main challenges we have encountered during the course of this project. Data quality and uniformity are the two most important data traits to be considered in clinical trials incorporating machine learning. Robust data pre-processing and machine learning pipelines have been employed in MALIBO, a task facilitated by the now freely available machine learning libraries and toolboxes. Another important consideration for achieving the desired clinical outcome in MALIBO, was to effectively host the resulting machine learning output, along with the clinical images, for reading in a clinical environment. Finally, a range of legal, ethical, and clinical acceptance issues should be considered when attempting to incorporate computer-assisting tools into clinical practice. The road from translating computational methods into potentially useful clinical tools involves an analytical, stepwise adaptation approach, as well as engagement of a multidisciplinary team.

Journal article

Alansary A, Oktay O, Li Y, Folgoc LL, Hou B, Vaillant G, Kamnitsas K, Vlontzos A, Glocker B, Kainz B, Rueckert Det al., 2019, Evaluating reinforcement learning agents for anatomical landmark detection, Medical Image Analysis, Vol: 53, Pages: 156-164, ISSN: 1361-8415

Automatic detection of anatomical landmarks is an important step for a wide range of applications in medical image analysis. Manual annotation of landmarks is a tedious task and prone to observer errors. In this paper, we evaluate novel deep reinforcement learning (RL) strategies to train agents that can precisely and robustly localize target landmarks in medical scans. An artificial RL agent learns to identify the optimal path to the landmark by interacting with an environment, in our case 3D images. Furthermore, we investigate the use of fixed- and multi-scale search strategies with novel hierarchical action steps in a coarse-to-fine manner. Several deep Q-network (DQN) architectures are evaluated for detecting multiple landmarks using three different medical imaging datasets: fetal head ultrasound (US), adult brain and cardiac magnetic resonance imaging (MRI). The performance of our agents surpasses state-of-the-art supervised and RL methods. Our experiments also show that multi-scale search strategies perform significantly better than fixed-scale agents in images with large field of view and noisy background such as in cardiac MRI. Moreover, the novel hierarchical steps can significantly speed up the searching process by a factor of 4-5 times.

Journal article

Schlemper J, Oktay O, Schaap M, Heinrich M, Kainz B, Glocker B, Rueckert Det al., 2019, Attention gated networks: Learning to leverage salient regions in medical images., Med Image Anal, Vol: 53, Pages: 197-207

We propose a novel attention gate (AG) model for medical image analysis that automatically learns to focus on target structures of varying shapes and sizes. Models trained with AGs implicitly learn to suppress irrelevant regions in an input image while highlighting salient features useful for a specific task. This enables us to eliminate the necessity of using explicit external tissue/organ localisation modules when using convolutional neural networks (CNNs). AGs can be easily integrated into standard CNN models such as VGG or U-Net architectures with minimal computational overhead while increasing the model sensitivity and prediction accuracy. The proposed AG models are evaluated on a variety of tasks, including medical image classification and segmentation. For classification, we demonstrate the use case of AGs in scan plane detection for fetal ultrasound screening. We show that the proposed attention mechanism can provide efficient object localisation while improving the overall prediction performance by reducing false positives. For segmentation, the proposed architecture is evaluated on two large 3D CT abdominal datasets with manual annotations for multiple organs. Experimental results show that AG models consistently improve the prediction performance of the base architectures across different datasets and training sizes while preserving computational efficiency. Moreover, AGs guide the model activations to be focused around salient regions, which provides better insights into how model predictions are made. The source code for the proposed AG models is publicly available.

Journal article

Bakas S, Reyes M, Jakab A, Bauer S, Rempfler M, Crimi A, Shinohara RT, Berger C, Ha SM, Rozycki M, Prastawa M, Alberts E, Lipkova J, Freymann J, Kirby J, Bilello M, Fathallah-Shaykh H, Wiest R, Kirschke J, Wiestler B, Colen R, Kotrotsou A, Lamontagne P, Marcus D, Milchenko M, Nazeri A, Weber M-A, Mahajan A, Baid U, Gerstner E, Kwon D, Acharya G, Agarwal M, Alam M, Albiol A, Albiol A, Albiol FJ, Alex V, Allinson N, Amorim PHA, Amrutkar A, Anand G, Andermatt S, Arbel T, Arbelaez P, Avery A, Azmat M, Pranjal B, Bai W, Banerjee S, Barth B, Batchelder T, Batmanghelich K, Battistella E, Beers A, Belyaev M, Bendszus M, Benson E, Bernal J, Bharath HN, Biros G, Bisdas S, Brown J, Cabezas M, Cao S, Cardoso JM, Carver EN, Casamitjana A, Castillo LS, Catà M, Cattin P, Cerigues A, Chagas VS, Chandra S, Chang Y-J, Chang S, Chang K, Chazalon J, Chen S, Chen W, Chen JW, Chen Z, Cheng K, Choudhury AR, Chylla R, Clérigues A, Colleman S, Colmeiro RGR, Combalia M, Costa A, Cui X, Dai Z, Dai L, Daza LA, Deutsch E, Ding C, Dong C, Dong S, Dudzik W, Eaton-Rosen Z, Egan G, Escudero G, Estienne T, Everson R, Fabrizio J, Fan Y, Fang L, Feng X, Ferrante E, Fidon L, Fischer M, French AP, Fridman N, Fu H, Fuentes D, Gao Y, Gates E, Gering D, Gholami A, Gierke W, Glocker B, Gong M, González-Villá S, Grosges T, Guan Y, Guo S, Gupta S, Han W-S, Han IS, Harmuth K, He H, Hernández-Sabaté A, Herrmann E, Himthani N, Hsu W, Hsu C, Hu X, Hu X, Hu Y, Hu Y, Hua R, Huang T-Y, Huang W, Huffel SV, Huo Q, Vivek HV, Iftekharuddin KM, Isensee F, Islam M, Jackson AS, Jambawalikar SR, Jesson A, Jian W, Jin P, Jose VJM, Jungo A, Kainz B, Kamnitsas K, Kao P-Y, Karnawat A, Kellermeier T, Kermi A, Keutzer K, Khadir MT, Khened M, Kickingereder P, Kim G, King N, Knapp H, Knecht U, Kohli L, Kong D, Kong X, Koppers S, Kori A, Krishnamurthi G, Krivov E, Kumar P, Kushibar K, Lachinov D, Lambrou T, Lee J, Lee C, Lee Y, Lee M, Lefkovits S, Lefkovits L, Levitt J, Li T, Li H, Li W, Li H, Li X, Li Y, Li H, Li Z, Li X, Li Z, Liet al., 2019, Identifying the best machine learning algorithms for brain tumorsegmentation, progression assessment, and overall survival prediction in the BRATS challenge

Gliomas are the most common primary brain malignancies, with differentdegrees of aggressiveness, variable prognosis and various heterogeneoushistologic sub-regions, i.e., peritumoral edematous/invaded tissue, necroticcore, active and non-enhancing core. This intrinsic heterogeneity is alsoportrayed in their radio-phenotype, as their sub-regions are depicted byvarying intensity profiles disseminated across multi-parametric magneticresonance imaging (mpMRI) scans, reflecting varying biological properties.Their heterogeneous shape, extent, and location are some of the factors thatmake these tumors difficult to resect, and in some cases inoperable. The amountof resected tumor is a factor also considered in longitudinal scans, whenevaluating the apparent tumor for potential diagnosis of progression.Furthermore, there is mounting evidence that accurate segmentation of thevarious tumor sub-regions can offer the basis for quantitative image analysistowards prediction of patient overall survival. This study assesses thestate-of-the-art machine learning (ML) methods used for brain tumor imageanalysis in mpMRI scans, during the last seven instances of the InternationalBrain Tumor Segmentation (BraTS) challenge, i.e., 2012-2018. Specifically, wefocus on i) evaluating segmentations of the various glioma sub-regions inpre-operative mpMRI scans, ii) assessing potential tumor progression by virtueof longitudinal growth of tumor sub-regions, beyond use of the RECIST/RANOcriteria, and iii) predicting the overall survival from pre-operative mpMRIscans of patients that underwent gross total resection. Finally, we investigatethe challenge of identifying the best ML algorithms for each of these tasks,considering that apart from being diverse on each instance of the challenge,the multi-institutional mpMRI BraTS dataset has also been a continuouslyevolving/growing dataset.

Working paper

Robinson R, Valindria VV, Bai W, Oktay O, Kainz B, Suzuki H, Sanghvi MM, Aung N, Paiva JÉM, Zemrak F, Fung K, Lukaschuk E, Lee AM, Carapella V, Kim YJ, Piechnik SK, Neubauer S, Petersen SE, Page C, Matthews PM, Rueckert D, Glocker Bet al., 2019, Automated quality control in image segmentation: application to the UK Biobank cardiac MR imaging study, Journal of Cardiovascular Magnetic Resonance, Vol: 21, ISSN: 1097-6647

Background: The trend towards large-scale studies including population imaging poses new challenges in terms of quality control (QC). This is a particular issue when automatic processing tools, e.g. image segmentation methods, are employed to derive quantitative measures or biomarkers for later analyses. Manual inspection and visual QC of each segmentation isn't feasible at large scale. However, it's important to be able to automatically detect when a segmentation method fails so as to avoid inclusion of wrong measurements into subsequent analyses which could lead to incorrect conclusions. Methods: To overcome this challenge, we explore an approach for predicting segmentation quality based on Reverse Classification Accuracy, which enables us to discriminate between successful and failed segmentations on a per-cases basis. We validate this approach on a new, large-scale manually-annotated set of 4,800 cardiac magnetic resonance scans. We then apply our method to a large cohort of 7,250 cardiac MRI on which we have performed manual QC. Results: We report results used for predicting segmentation quality metrics including Dice Similarity Coefficient (DSC) and surface-distance measures. As initial validation, we present data for 400 scans demonstrating 99% accuracy for classifying low and high quality segmentations using predicted DSC scores. As further validation we show high correlation between real and predicted scores and 95% classification accuracy on 4,800 scans for which manual segmentations were available. We mimic real-world application of the method on 7,250 cardiac MRI where we show good agreement between predicted quality metrics and manual visual QC scores. Conclusions: We show that RCA has the potential for accurate and fully automatic segmentation QC on a per-case basis in the context of large-scale population imaging as in the UK Biobank Imaging Study.

Journal article

Harvey H, Glocker B, 2019, A standardised approach for preparing imaging data for machine learning tasks in radiology, Artificial Intelligence in Medical Imaging: Opportunities, Applications and Risks, Pages: 61-72, ISBN: 9783319948775

Book chapter

Reyes M, Konukoglu E, Glocker B, Wiest Ret al., 2019, Preface IMIMIC 2019, ISBN: 9783030338497

Book

Pawlowski N, Jaques M, Glocker B, 2019, Efficient variational Bayesian neural network ensembles for outlier detection

In this work we perform outlier detection using ensembles of neural networks obtained by variational approximation of the posterior in a Bayesian neural network setting. The variational parameters are obtained by sampling from the true posterior by gradient descent. We show our outlier detection results are comparable to those obtained using other efficient ensembling methods.

Conference paper

Konukoglu E, Glocker B, 2019, Random forests in medical image computing, Handbook of Medical Image Computing and Computer Assisted Intervention, Pages: 457-480, ISBN: 9780128161760

The Random Forests algorithm had a substantial impact on medical image computing over the last decade. This chapter presents basic algorithmic details, some variations proposed in the recent years and applications in medical image computing. Arguably, Random Forests’ main impact was on the analysis tasks that required understanding spatial context within the images. We take a specific angle and view Random Forests as a machine learning tool that can integrate contextual information. We position the algorithm and its contributions within the larger field from this respect. Lastly, we briefly discuss how Random Forests and deep learning methods relate to each other and how they differ.

Book chapter

Cardoso MJ, Feragen A, Glocker B, Konukoglu E, Oguz I, Unal G, Vercauteren Tet al., 2019, Preface, Proceedings of Machine Learning Research, Vol: 102, Pages: 1-3

Journal article

, 2019, Preface., Publisher: PMLR, Pages: 1-3

Conference paper

Chen X, Pawlowski N, Glocker B, Konukoglu Eet al., 2019, Unsupervised Lesion Detection with Locally Gaussian Approximation, 10th International Workshop on Machine Learning in Medical Imaging (MLMI) / 22nd International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Publisher: SPRINGER INTERNATIONAL PUBLISHING AG, Pages: 355-363, ISSN: 0302-9743

Conference paper

Baweja C, Glocker B, Kamnitsas K, 2018, Towards continual learning in medical imaging, Medical imaging meets NIPS

This work investigates continual learning of two segmentation tasks in brain MRIwith neural networks. To explore in this context the capabilities of current methodsfor countering catastrophic forgetting of the first task when a new one is learned,we investigateelastic weight consolidation[1], a recently proposed method basedon Fisher information, originally evaluated on reinforcement learning of Atarigames. We use it to sequentially learn segmentation of normal brain structures andthen segmentation of white matter lesions. Our findings show this recent methodreduces catastrophic forgetting, while large room for improvement exists in thesechallenging settings for continual learning.

Conference paper

Mitchell J, Kamnitsas K, Singleton K, Whitmire S, Clark-Swanson K, Rickertsen C, Glocker B, Hu L, Swanson Ket al., 2018, DEEP LEARNING FOR ACCURATE, RAPID, FULLY AUTOMATIC MEASUREMENT OF BRAIN TUMOR-ASSOCIATED ABNORMALITY SEEN ON MRI, 23rd Annual Scientific Meeting and Education Day of the Society-for-Neuro-Oncology (SNO) / 3rd CNS Anticancer Drug Discovery and Development Conference, Publisher: OXFORD UNIV PRESS INC, Pages: 180-181, ISSN: 1522-8517

Conference paper

Singleton K, Mitchell J, Ranjbar S, Kamnitsas K, Whitmire S, Clark-Swanson K, Rickertsen C, Rubin J, Glocker B, Hu L, Swanson Ket al., 2018, DEEP LEARNING DETECTS DIFFERENCES IN THE MRIs OF MALE AND FEMALE GLIOMAS, 23rd Annual Scientific Meeting and Education Day of the Society-for-Neuro-Oncology (SNO) / 3rd CNS Anticancer Drug Discovery and Development Conference, Publisher: OXFORD UNIV PRESS INC, Pages: 177-177, ISSN: 1522-8517

Conference paper

Hou B, Miolane N, Khanal B, Lee M, Alansary A, McDonagh SG, Hajnal JV, Rueckert D, Glocker B, Kainz Bet al., 2018, Computing CNN loss and gradients for pose estimation with Riemannian geometry, International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Publisher: Springer Verlag, Pages: 756-764, ISSN: 0302-9743

Pose estimation, i.e. predicting a 3D rigid transformation with respect to a fixed co-ordinate frame in, SE(3), is an omnipresent problem in medical image analysis. Deep learning methods often parameterise poses with a representation that separates rotation and translation.As commonly available frameworks do not provide means to calculate loss on a manifold, regression is usually performed using the L2-norm independently on the rotation’s and the translation’s parameterisations. This is a metric for linear spaces that does not take into account the Lie group structure of SE(3). In this paper, we propose a general Riemannian formulation of the pose estimation problem, and train CNNs directly on SE(3) equipped with a left-invariant Riemannian metric. The loss between the ground truth and predicted pose (elements of the manifold) is calculated as the Riemannian geodesic distance, which couples together the translation and rotation components. Network weights are updated by back-propagating the gradient with respect to the predicted pose on the tangent space of the manifold SE(3). We thoroughly evaluate the effectiveness of our loss function by comparing its performance with popular and most commonly used existing methods, on tasks such as image-based localisation and intensity-based 2D/3D registration. We also show that hyper-parameters, used in our loss function to weight the contribution between rotations andtranslations, can be intrinsically calculated from the dataset to achievegreater performance margins.

Conference paper

Alansary A, Le Folgoc L, Vaillant G, Oktay O, Li Y, Bai W, Passerat-Palmbach J, Guerrero R, Kamnitsas K, Hou B, McDonagh S, Glocker B, Kainz B, Rueckert Det al., 2018, Automatic view planning with multi-scale deep reinforcement learning agents, International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Publisher: Springer Verlag, Pages: 277-285, ISSN: 0302-9743

We propose a fully automatic method to find standardizedview planes in 3D image acquisitions. Standard view images are impor-tant in clinical practice as they provide a means to perform biometricmeasurements from similar anatomical regions. These views are often constrained to the native orientation of a 3D image acquisition. Navigating through target anatomy to find the required view plane is tedious and operator-dependent. For this task, we employ a multi-scale reinforcement learning (RL) agent framework and extensively evaluate several DeepQ-Network (DQN) based strategies. RL enables a natural learning paradigm by interaction with the environment, which can be used to mimic experienced operators. We evaluate our results using the distance between the anatomical landmarks and detected planes, and the angles between their normal vector and target. The proposed algorithm is assessed on the mid-sagittal and anterior-posterior commissure planes of brain MRI, and the 4-chamber long-axis plane commonly used in cardiac MRI, achieving accuracy of 1.53mm, 1.98mm and 4.84mm, respectively.

Conference paper

Castro DC, Glocker B, 2018, Nonparametric density flows for MRI intensity normalisation, International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), Publisher: Springer, Cham, Pages: 206-214, ISSN: 0302-9743

With the adoption of powerful machine learning methods in medical image analysis, it is becoming increasingly desirable to aggregate data that is acquired across multiple sites. However, the underlying assumption of many analysis techniques that corresponding tissues have consistent intensities in all images is often violated in multi-centre databases. We introduce a novel intensity normalisation scheme based on density matching, wherein the histograms are modelled as Dirichlet process Gaussian mixtures. The source mixture model is transformed to minimise its L2 divergence towards a target model, then the voxel intensities are transported through a mass-conserving flow to maintain agreement with the moving density. In a multi-centre study with brain MRI data, we show that the proposed technique produces excellent correspondence between the matched densities and histograms. We further demonstrate that our method makes tissue intensity statistics substantially more compatible between images than a baseline affine transformation and is comparable to state-of-the-art while providing considerably smoother transformations. Finally, we validate that nonlinear intensity normalisation is a step toward effective imaging data harmonisation.

Conference paper

This data is extracted from the Web of Science and reproduced under a licence from Thomson Reuters. You may not copy or re-distribute this data in whole or in part without the written consent of the Science business of Thomson Reuters.

Request URL: http://wlsprd.imperial.ac.uk:80/respub/WEB-INF/jsp/search-html.jsp Request URI: /respub/WEB-INF/jsp/search-html.jsp Query String: limit=30&id=00795421&person=true&page=7&respub-action=search.html