Home | Research | Software | CV | Publications | Talks | Funding and Awards | Teaching and Supervision | Collaborators |

I am interested in the development of **fast numerical methods for the solution of computationally challenging physical problems**, such as numerical weather- and climate prediction models, solution of stochastic differential equations in dispersion modelling or simulations of fundamental particle interactions. Usually this requires the implementation of complex algorithms on **massively parallel computers** and a careful analysis of model performance to achieve optimal efficiency. On modern chip architectures, such as GPUs, access to global memory should be reduced as much as possible, and this creates additional challenges for the implementation of memory bound applications such as solvers for very large sparse systems of equations.

The following examples demonstrate my strong interest in the application of these methods in 'real-life' applications such as operational atmospheric forecast models like the Met Office's Unified Model and the NAME Atmospheric Dispersion Model.

One of the main topics of my current research is the efficient implementation of I have a strong interest in emerging computer architectures and the application of I demonstrated the scalability of our solvers for problems with more than 10 Multigrid solver have also been implemented for more challenging mimetic (mixed) finite element discretisations in the Firedrake framework [Mitchell, Müller (2016)]. Performance measurements on the ARCHER supercomputer have demonstrated that the performance of a bespoke tensor-product geometric multigrid solver is comparable to that of the Boomer AMG implementation. I also implemented an optimised |

Higher order Discontinuous Galerkin (DG) methods can be used to discretise a wide range of partial differential equations in science an engineering. Applications include subsurface slow phenomena (e.g. oil- and gas- reservoir modelling, aquifers and transport in radioactive waste disposal) and numerical weather prediction. Compared to low-order discretisations, higher-order DG methods have the advantage of locality and very high computational intensity, in particular if a sum-factorisation approach is used. This makes optimal use of modern computer architectures, which can execute a very large number of floating point operations, but are slow in loading data from memory. However, the solution of elliptic PDEs requires the inversion of small dense matrices in the block-smoother, and traditionally those matrices need to be assembled and inverted, which counteracts the advantages of the matrix-free operator application. To overcome this issue, I work with colleagues at Heidelberg University to invert the dense matrix blocks iteratively, thus recovering the advantages of the matrix-free code. Tests demonstrated the superior performance of this approach even for relatively low polynomial degrees. All code has been implemented in the EXADUNE fork of the DUNE C++ library. |

The simulation of large numbers of particles with Molecular Dynamics (MD) algorithms is a key tool in computational chemistry and physics. To extract physically meaningful data requires implementations on massively parallel computer architectures. In a diversifying hardware landscape (manycore CPUs, GPUs, Xeon Phi, ...) it is increasingly difficult for the domain specialist (computational physicist/chemist) to be an expert in the low-level optimisation of their codes. This makes the implementation of efficient yet portable MD codes for modern parallel computers very challenging. To overcome this issue, we developed a Python code generation framework [Saunders, Grant, Müller (2017b)], [Saunders, Grant, Müller (2017a)] inspired by the PyOP2 ``Separation-of-Concerns`` approach for related problems in the solution of grid-based PDEs. The domain specialist describes the fundamental interaction between particle pairs as a short C kernel, and this is then executed over all particles pairs. Since the allowed kernels are very general, this also allows the easy implementation of structure analysis codes and with increasing data volumes this aspect becomes more important. |

Image source: Árni Friðriksson (Own work) [CC-BY-SA-3.0 or GFDL], via Wikimedia Commons |
As a scientist in the Met Office Atmospheric Dispersion group (November 2009 - August 2011) I contributed to the development of the NAME dispersion model, which is used for research applications and operational prediction of the spread and transport of atmospheric pollutants. The model was used for the operational forecast of the spread of volcanic ash in the European airspace following the 2010 and 2011 eruptions of Eyjafjallajökull and Grimsvotn. In particular I was responsible for the During my time at the Met Office I was also involved in the development of a new operational emergency response system for accidents such as volcanic eruptions or nuclear accidents. Together with Prof. Rob Scheichl and Dr. Tony Shardlow I |

During my PhD at the University of Edinburgh (September 2006 - August 2009) I investigated the interactions of particles on a much smaller scale: I used and extended a Python/Fortran MPI code for calculating the gluonic corrections to a heavy meson decay. Due to the heavy mass of the decaying b-quark, an effective theory is used and the In my diploma thesis at the University of Bonn (Jul 2005 - Jul 2006) I used |

© Eike Müller, University of Bath. Last updated: Mon 13 Nov 2017 16:54:55h