## Research

Much of what we do in science is inference, which means we have some experimental data, from which we infer something about the Universe. For example we may wish to determine how fast the Universe is expanding, or what the properties of Dark Energy are. We can also ask broader questions such as whether the Big Bang model is the preferred theoretical framework, or whether Einstein's General Relativity is favoured over another gravity theory. These are the sorts of questions I and colleagues at the ICIC address, through statistical analysis of cosmological data such as obtained from gravitational lensing, galaxy or microwave background surveys. My specific interests centre on developing and applying the best methods for extracting information from data, such as: using the full

3D information from weak lensing surveys, and using size as well as shape information; transforming the data so that theory can be applied more effectively; compressing data in an optimised way so statistical analysis can be done very fast; looking for non-Gaussianity in the Cosmic Microwave Background and in the galaxy distribution. The scientific questions I am interested in include whether or not the acceleration of the Universe is due to Einstein's cosmological constant, Dark Energy, or alternative gravity models.

## Bayesian Evidence from Planck MCMC chains

In Heavens et al. 2017 we compute Bayesian Evidence (or Marginal Likelihood for cosmological models from the Planck Monte Carlo Markov Chains. The Python code to do this is public here. The full grids of models and datasets analysed are: chains analysed separately, chains concatenated, and a folder of txt and csv files.

If the latest Riess et al (2016) Hubble constant result is added as an additional Gaussian likelihood, then the wCDM model is moderately preferred over standard LCDM, but does not exclude standard LCDM.

## Baryon Acoustic Oscillation scale

Raul Jimenez, Licia Verde and I have measured the BAO ruler length - a key observable quantity in cosmology theories, in an almost model-independent way from supernova and galaxy clustering data. The length, 101.9 +/- 1.9 Mpc/h is measured independently of General Relativity. Published in December 2014 in PRL as Editors' Suggestion, there is an accessible account in their physics highlights. The paper itself is here. Picture courtesy of Chris Blake and Sam Moorfield.

## MEDICAL IMAGING

Massive data compression techniques allow very rapid analysis of large data sets, such as come from medical scanners. I am a founding director of Blackford Analysis, a spin-out company from the University of Edinburgh, which specialises in making radiology tasks more efficient.

## BAYESIAN HIERARCHICAL MODELS

The outcome of Bayesian inference is the posterior probability:

PROBABILITY OF WHAT YOU WANT TO KNOW, GIVEN THE DATA AND ANY PRIOR INFORMATION THAT YOU HAVE.

Computing this is usually an interesting challenge, but is now possible for cosmological datasets. In our work, it involves exploring very high-dimensional parameter spaces, of the order of a million dimensions. We do this successfully with Gibbs sampling or Hamiltonian Monte Carlo sampling. Examples are here, in work done with then PhD student Justin Alsing, Andrew Jaffe and colleagues.