Fridays at 12.15 at Wolfson 4W 1.7. All talks will be broadcast on Zoom.

Link to Zoom meeting

Everyone is welcome at these talks.

Date Speaker Title
10 Feb 2023 Matthias Sachs (Birmingham)
Learning of tensor-valued quantities with the atomic cluster expansion - from (hyper-)active learning of interatomic forcefields to tensor-valued atomic cluster expansion for learning of dynamical systems


17 Feb 2023 Changpeng Shao (Bristol)
Quantum algorithms for linear regression

Quantum computers could solve linear algebra problems much faster than classical computers. In this talk, I will talk about some recent quantum algorithms for linear regression with the goal of outputting a vector solution. First, I will introduce a quantum algorithm that accelerates the technique of leverage score sampling (which is a useful technique in randomised numerical linear algebra). Then I will show how to use it to propose efficient quantum algorithms for solving linear regressions. The talk is mainly based on arXiv 2301.06107.

24 Feb 2023 Rihuan Ke (Bristol)
Hybrid learning for label-efficient solvers for inverse problems and beyond

Recent deep learning techniques enable the learning of high-quality inverse problem solvers from data. These techniques handle the ill-posedness of the problems with data, by looking at the examples, and do not require explicit regularisation modelling. Typically, a large set of paired measurements and ground truth solutions is used in order to learn a well-performing model. The learning task can be challenging without such a high-quality dataset, and it is an ill-posed problem itself. In this talk, I will introduce some hybrid learning methods for solving inverse problems in a label-efficient manner.

3 Mar 2023 CANCELLED


10 Mar 2023 Francisco de Lima Andrade (ENS Paris)
Distributed Banach-Picard Iteration for Distributed Inference - Theory and Applications

Many inference problems can be mathematically formulated as finding a fixed point of a contractive operator/map. In modern distributed scenarios (e.g., distributed machine learning or sensor networks), this map can be naturally written as the average of individual maps held locally and privately (i.e., the agents don’t want to share their local data with the others) by a set of agents linked by a (maybe sparse) communication network. Starting with the classical Banach-Picard iteration (BPI), which is is a widely used natural choice to find fixed points of locally contractive maps, this talk shows how to extend the BPI to these distributed settings. We do not assume that the locally contractive map comes from an underlying optimization problem, which precludes exploiting strong global properties such as convexity, coercivity, or Lipschitzianity. Yet, we present a distributed algorithm (called distributed Banach-Picard iteration DBPI) that keeps the linear convergence rate of the standard BPI for the average locally contractive map. As an application, we derive and prove linear convergence of two distributed algorithms for two classical data analysis problems - the expectation-maximization algorithm for parameter estimation from noisy and faulty distributed sensors; principal component analysis with distributed data (equivalently finding the top m eigenvectors of a positive semidefinite matrix, which is the average of local matrices held by the network agents).

17 Mar 2023 Eike Mueller (Bath)
Equivariant and invariant neural networks

Equivariance and invariance play a very important role in physics since they strongly constrain the dynamics of a system. For example, Einstein’s famous theory of Special Relativity and Maxwell’s electrodynamics follow naturally if we assume that the equations of motion are equivariant under Lorentz transformations. Recently, there has been a lot of work on constructing equivariant neural networks. A well-known example are Convolutional Neural Networks (CNNs) for image classification, in which the output is invariant under translations - the network doesn’t care whether a cat appears in the upper left or lower right corner of an image. Here translational invariance provides a strong prior, which allows the network to generalise from limited training data. Equivariance under other symmetry groups such as rotations can be incorporated into the architecture of neural networks in similar ways. In this session I will try to summarise some of the main ideas, based on some papers I came across recently. I’m by no means an expert in this area, so this will be a very informal session, hopefully with lively and interesting discussions.

24 Mar 2023 Philip J Herbert (Heriot-Watt)
Shape optimisation with Lipschitz functions

In this talk, we discuss a novel method in PDE constrained shape optimisation. We begin by introducing the concept of PDE constrained shape optimisation. While it is known that many shape optimisation problems have a solution, their approximation in a meaningful way is non-trivial. To find a minimiser, it is typical to use first order methods. The novel method we propose is to deform the shape with fields which are a direction of steepest descent in the topology of $W^{1,\infty}$. We present an analysis of this in a discrete setting along with the existence of directions of steepest descent. Several numerical experiments will be considered which compare a classical Hilbertian approach to this novel approach.

31 Mar 2023 Arieh Iserles (Cambridge)
An overarching framework for spectral methods and dispersive equations

Many invariants of time-evolving PDEs, e.g. mass and some Hamiltonians, can be formulated as a bilinear form. In this talk we are concerned with formal orthogonal systems on the real line (and, by tensorisation, in $\mathbb{R}^d$) with respect to bilinear forms and with a tridiagonal differentiation matrix. The theory is virtually complete for $L_2$ and Sobolev inner products - using a Fourier transform we show that such systems are in a one-to-one relationship with determinate Borel measures and that their closure is a Paley–Wiener space. We provide several examples, commencing from the familiar Hermite functions. We also characterise all such systems that can be computed fast – for $L_2$ orthogonality using Fast Fourier/Cosine/Sine Transform, for Sobolev orthogonality the above in tandem with a narrow-banded matrix of connection coefficients. We conclude with preliminary results on systems that are orthogonal with respect to a bilinear form generated by the Hamiltonian of a linear Schrödinger equation. This is joint work with Marcus Webb.

21 Apr 2023 Kostas Papafitsoros (QMUL)


28 Apr 2023 Jemima M. Tabeart (Oxford)


Subscribe to seminar calendar

You can subscribe to the NA calendar directly from your calendar client, including Outlook, Apple’s iCalendar or Google calendar. The web address of the calendar is this ICS link which you will need to copy.

To subscribe to a calendar in Outlook:

  1. In Calendar view, select “Add Calendar” (large green +)
  2. Select “From Internet”
  3. Copy paste the ICS link, click OK, and click Yes to subscribe.

To subscribe to a calendar in iCalendar, please follow these instructions. Copy paste the ICS link in “web address”.

To subscribe to a calendar in Google Calendar:

  1. Go to link.
  2. On the left side go to "Other Calendars" and click on the dropdown.
  3. Choose "Add by URL".
  4. Copy paste the ICS link in the URL of the calendar.
  5. Click on "Add Calendar" and wait for Google to import your events. This creates a calendar with a somewhat unreadable name.
  6. To give a readable name to the calendar, click on the three vertical dots sign next to the newly created calendar and select Settings.
  7. Choose a name for the calendar, eg. Numerical Analysis @ Bath, and click back button on top left.

How to get to Bath See here for instructions how to get to Bath. Please email Pranav Singh ( if you intend to come by car and require a parking permit for Bath University Campus for the day.
Tips for giving talks

Tips for new students on giving talks

Since the audience of the NA seminar contains both PhD students and staff with quite wide interests and backgrounds, the following are some guidelines/hints to make sure people don't give you evil looks at lunch afterwards.

Before too much time passes in your talk, ideally the audience should know the answers to the following 4 questions:

  • What is the problem you're considering?
  • Why do you find this interesting?
  • What has been done before on this problem/what's the background?
  • What is your approach/what are you going to talk about?

There are lots of different ways to communicate this information. One way, if you're doing a slide show, could be for the first 4 slides to cover these 4 questions; although in this case you may want to revisit these points later on in the talk (e.g. to give more detail).


  • "vertebrate style" (structure hidden inside - like the skeleton of a vertebrate) = good for detective stories, bad for maths talks.
  • "crustacean style" (structure visible from outside - like the skeleton of a crustacean) = bad for detective stories, good for maths talks.