Surrogates and Dimension Reduction in Scientific Machine Learning

9th-10th September, 2025

Hybrid Event in the University of Manchester, United Kingdom

Register

About DRSciML-2025

Scientific machine learning has increasingly focused on surrogates for solving partial differential equations, modeling dynamical systems, and learning physical phenomena from data. These surrogates offer efficient and scalable alternatives to traditional methods, with significant potential across diverse applications. However, challenges persist in ensuring stability and providing robust theoretical guarantees. This workshop will explore recent advances in surrogate modeling, emphasizing dimension reduction, effective training strategies, and methods that enhance accuracy and theoretical rigor.

DRSciML-2025 Programme

9th September 2025 - Tuesday

08:45 - 09:00 BST Registration
09:10 - 10:00 BST
Plenary Talk - George Stepaniants
Title: Learning Memory and Material Dependent Constitutive Laws

The simulation of multiscale viscoelastic materials poses a significant challenge in computational materials science, requiring expensive numerical solvers that can resolve dynamics of material deformations at the microscopic scale. The theory of homogenization offers an alternative approach to modeling, by locally averaging the strains and stresses of multiscale materials. This procedure eliminates the smaller scale dynamics but introduces a history dependence between strain and stress that proves very challenging to characterize analytically. In the one-dimensional setting, we give the first full characterization of the memory-dependent constitutive laws that arise in multiscale viscoelastic materials. Using this theory, we develop a neural operator architecture, that simultaneously across a wide range of material microstructures, accurately predicts their homogenized constitutive laws, thus enabling us to simulate their deformations under forcing. We use the approximation theory of neural operators to provide guarantees on the generalization of our approach to unseen material samples.

10:00 - 10:30 BST Coffee Break
10:30 - 11:00 BST Talk - Fernando Henríquez
11:00 - 11:30 BST
Title: Operator Learning for PDEs

In this talk, I will present a series of recent works that explore the intersection of deep learning and scientific computing, with a particular focus on operator learning and generative modeling for PDEs. I will begin with our work on Convolutional Neural Operators (CNOs), which reinterprets operator learning through the lens of classical CNNs. I will then discuss Representation Equivalent Neural Operators (ReNO), where we address representational ambiguities and aliasing in operator learning. Building on these foundations, I will introduce Poseidon, a foundation model for time-dependent PDEs that leverages multiscale attention and temporal conditioning. Finally, I will present GenCFD, a diffusion-based generative model for statistical CFD, which captures uncertainty and high-order statistics in turbulent flows. Each of these works contributes a step toward building accurate, robust, and generalizable models for physical systems.

11:30 - 12:00 BST
Title: Bayesian Operator Learning

We discuss a Bayesian framework for learning non-linear mappings in infinite-dimensional spaces. Given a map $g_0:\mathcal{X}\to\mathcal{Y}$ between two separable Hilbert spaces, we study recovery of $g_0$ from $n\in\mathbb{N}$ noisy input-output pairs $(\pmb{x},\pmb{y})=(X_i,Y_i)_{i=1}^n$ with $Y_i = g_0 (X_i ) + E_i$ ; here the $X_i\in\mathcal{X}$ represent randomly drawn 'design' points, and the $E_i$ are assumed to be i.i.d. draws from a Gaussian white noise process indexed by $\mathcal{Y}$. Choosing a suitable 'operator-valued' prior $\Pi_0$, we show well-posedness of the posterior $G|(\pmb{X},\pmb{Y})$ and establish convergence rates for the posterior mean towards the ground truth in terms of the data size $n$.

12:00 - 14:00 BST Lunch
14:00 - 14:30 BST Talk - Dibyakanti Kumar
14:30 - 15:00 BST
Title: Optimization-Ready Surrogates with Derivative-Informed Training

Modern decision-making for complex physical and engineered systems increasingly requires the ability to quantify high-dimensional uncertainties and make decisions by solving risk-averse optimization problems under uncertainty all in near real-time. Operator learning has emerged as a promising framework for enabling scalable surrogate modeling in this context. However, such approaches inevitably introduce approximation errors that can degrade the quality of downstream decisions. In this talk, we explore how formulating operator learning tasks through the lens of decision-making objectives such as inverse problems and optimization under uncertainty can help mitigate these limitations. We derive a priori error bounds for these optimization tasks, which motivate operator learning formulations that explicitly penalize errors in the derivatives of input-output maps. We demonstrate the effectiveness of these methods on problems in structural health monitoring, shape optimization, and fluid flow control. Our approaches lead to (i) improved accuracy in surrogate-based optimization and (ii) empirically enhanced performance in statistical learning tasks.

15:00 - 15:30 BST
Title: Scalable, adaptive, and explainable geometric deep learning with applications to computational physics

This work introduces an interpretable graph neural network (GNN) autoencoding framework for surrogate modeling of unstructured mesh fluid dynamics. The approach addresses two key challenges: latent space interpretability and compatibility with unstructured meshes. Interpretability is achieved through an adaptive graph reduction procedure that performs flowfield-conditioned node sampling, yielding masked fields that (i) localize latent activity in physical space, (ii) track the evolution of unsteady flow features such as recirculation zones and shear layers, and (iii) identify sub-graphs most relevant to forecasting. Compatibility with unstructured meshes is ensured through multi-scale message passing (MMP) layers, which extend standard message passing with learnable coarsening operations to efficiently reconstruct flowfields across multiple lengthscales. A regularization procedure further enhances interpretability by enabling error-tagging, i.e., the identification of nodes most associated with forecasting uncertainty. Demonstrations are conducted on large-eddy simulation data of backward-facing step flows at high Reynolds numbers, with extrapolations to ramp and wall-mounted cube configurations. The result is a new class of GNN autoencoders that combine predictive accuracy with physics-informed insight into both flow structures and model reliability.

15:30 - 16:00 BST Coffee Break
16:00 - 16:50 BST
Plenary Talk - Olga Fink
Title: From Physics to Machine Learning and Back: Applications in Dynamical Systems

Deep learning has become an essential tool in many engineering applications. However, its effectiveness is often limited by its reliance on large, representative, and well-labeled datasets. In contrast, condition monitoring data from complex systems is typically sparse, unlabeled, and unrepresentative, making it difficult to apply purely data-driven methods effectively. Moreover, deep learning models often perform poorly in extrapolation scenarios—common in engineering systems with long service lives and evolving operational regimes. To address these limitations, integrating physical laws and domain knowledge into deep learning frameworks has shown significant potential. This presentation will explore a range of approaches that integrate physics-based principles into machine learning models. Particular attention will be given to the use of structural inductive biases—such as those introduced by physics-informed graph neural networks—to improve model robustness, generalization and extrapolation. Finally, the talk will examine emerging methods in symbolic regression that aim to close the loop between data-driven learning and physical understanding, enabling the discovery of interpretable, physics-consistent models from data.

18:00 - 20:00 BST Conference Dinner

10th September 2025 - Wednesday

09:10 - 10:00 BST
Plenary Talk - Olga Mula
Title: Stable Nonlinear Dynamical Approximation with Dynamical Sampling

In this talk, I present a strategy to solve time-dependent PDEs with nonlinear dynamical approximation schemes which are provably stable. The stability, and numerical complexity of the approach rely on the ability to evaluate the numerical solution (or some moments of it) at specific locations that evolve in time, and which are chosen in such a way to maximize the stability of the method. I will explain the concept of stability that is involved as well as a possible approach for the dynamical sampling. I will illustrate the behavior of the method in several numerical examples.

10:00 - 10:30 BST Coffee Break
10:30 - 11:00 BST
Talk - Anshima Singh
Title: Wavelet-based Physics-Informed Neural Networks

Physics-informed neural networks (PINNs) face significant challenges when solving differential equations with rapid oscillations, steep gradients, or singular behavior. Considering these challenges, we designed an efficient wavelet-based PINNs (W-PINNs) model to address this class of differential equations. We represent the solution in wavelet space using a family of smooth-compactly supported wavelets. This approach captures the dynamics of complex physical phenomena with significantly fewer degrees of freedom, enabling faster and more accurate training by searching for solutions within the wavelet space. Developed model eliminates the need for automatic differentiation of derivatives in differential equations and requires no prior knowledge about solution behavior, such as the location of abrupt features. In this talk, I will demonstrate how W-PINNs excel at capturing localized nonlinear information through a strategic fusion of wavelets with PINNs, making them particularly effective for problems exhibiting abrupt behavior in specific regions. Our experimental findings show that W-PINNs significantly outperform traditional PINNs, PINNs with wavelets as an activation function, and other state-of-the-art methods, offering a promising approach for tackling challenging differential equations in scientific and engineering applications.

11:00 - 11:30 BST
Talk - Benno Huber
Title: Low-rank surrogates for parametric PDEs

Mathematically, operator surrogates provide approximations of mappings between infinite dimensional spaces. We analyze an encoder-decoder framework, where the input and output function spaces are parametrized using coefficient sequences of admissable representation systems. The goal then becomes to approximate a coefficient-to-coefficient map, which can be realized using various approximation tools. In the present work we focus on low-rank tensor representation using the tensor-train (TT) format. This format allows for efficient storage, evaluation and basic linear algebra as long as the tensor ranks remain low. We show approximation rates with regards to the storage complexity of such TT surrogates for certain holomorphic maps by providing rank bounds. Furthermore, we draw comparisons to other approximation tools such as neural networks and sparse polynomial interpolation.

11:30 - 12:00 BST
Title: The mathematics behind deep autoencoders: what we know and what is next

Deep autoencoders have become a fundamental tool in various machine learning applications, ranging from dimensionality reduction and reduced order modeling of partial differential equations to anomaly detection and neural machine translation. However, despite their empirical success, the theoretical understanding of these architectures remains relatively underdeveloped, particularly in contrast to classical linear approaches such as Principal Component Analysis. In this talk, I will overview some of the theory concerning deep autoencoders, addressing different aspects through different mathematical tools. More precisely, I will discuss: (i) the relationship between the latent dimension of deep autoencoders and some results in topology and dimension theory, such as the Borsuk-Ulam theorem and the Menger-Nöbeling theorem; (ii) the extension of such results to the context of stochastic systems, where Gaussian processes enter into play; (iii) some results about the expressivity of deep autoencoders, specifically focusing on convolutional architectures and, finally, (iv) some recent advancements concerning the design and optimization of symmetric architectures inspired by the Eckart-Young theorem. Theoretical results will be accompanied by suitable numerical experiments, illustrating the interplay between theory and practice.

12:00 - 13:30 BST Lunch
13:30 - 14:20 BST
Plenary Talk - Ricardo Vinuesa (online)
Title: Improving turbulence control through explainable deep learning

In this work we first use explainable deep learning based on Shapley explanations to identify the most important regions for predicting the future states of a turbulent channel flow. The explainability framework (based on gradient SHAP) is applied to each grid point in the domain, and through percolation analysis we identify coherent flow regions of high importance. These regions have around 70% overlap with the intense Reynolds-stress (Q) events in two-dimensional vertical planes. Interestingly, these importance-based structures have high overlap with classical turbulence structures (Q events, streaks and vortex clusters) in different wall-normal locations, suggesting that this new framework provides a more comprehensive way to study turbulence. We also discuss the application of deep reinforcement learning (DRL) to discover active-flow-control strategies for turbulent flows, including turbulent channels, three-dimensional cylinders and turbulent separation bubbles. In all the cases, the discovered DRL-based strategies significantly outperform classical flow-control approaches. We conclude that DRL has tremendous potential for drag reduction in a wide range of complex turbulent-flow configurations.

14:20 - 14:50 BST Coffee Break
14:50 - 15:20 BST
Title: Introduction to Fractional Dynamics and Learning Approaches

Many natural and engineered systems remember their past, for example, materials that slowly deform, fluids with long-lasting vortices, or biological populations affected by historical conditions. Traditional models often struggle to capture this memory effect, but fractional differential equations provide a powerful mathematical framework. In this talk, I will introduce fractional dynamics using simple examples to show how past states influence present behaviour. Using visualizations and simulations, I will demonstrate how learning-based methods, such as neural networks, can efficiently approximate these systems. While the focus is on conceptual understanding, I will also briefly highlight how these methods connect to advanced research areas, such as operator learning, neural PDE solvers, and physics-informed learning. The goal is to provide a clear understanding of memory-dependent systems while presenting a path to more advanced modelling techniques.

15:20 - 15:50 BST
Title: Learning probabilistic filters for data assimilation

Filtering - the task of estimating the conditional distribution for states of a dynamical system given partial and noisy observations - is important in many areas of science and engineering, including weather and climate prediction. However, the filtering distribution is generally intractable to obtain for high-dimensional, nonlinear systems. Filters used in practice, such as the ensemble Kalman filter (EnKF), provide biased probabilistic estimates for nonlinear systems and have numerous tuning parameters. I will present a framework for learning a parameterised analysis map - the transformation that takes samples from a forecast distribution, and combines with an observation, to update the approximate filtering distribution - using variational inference. In principle this can lead to a better approximation of the filtering distribution, and hence smaller bias. We show that this methodology can be used to learn the gain matrix, in an affine analysis map, for filtering linear and nonlinear dynamical systems; we also study the learning of inflation and localisation parameters for an EnKF. The framework developed here can also be used to learn new filtering algorithms with more general forms for the analysis map. I will also present some recent work on learning corrections to the EnKF using permutation-invariant neural architectures, leading to superior performance compared to leading methods in filtering chaotic systems.

Plenary Speakers

EPFL, Switzerland
University of Vienna, Austria
California Institute of Technology, US
University of Michigan, US

Talks

Organizers

Anirbit Mukherjee
University of Manchester, United Kingdom
Jakob Zech
Heidelberg University, Germany


Co-organizers

Mauricio A Álvarez
University of Manchester, United Kingdom
Tobias Buck
Heidelberg University, Germany


Student Volunteer

Dibyakanti Kumar
University of Manchester, United Kingdom

Location

Room 2B.026, Engineering Building B, University of Manchester, United Kingdom
This will be a hybrid event. Zoom links will be available on the page for joining during the event.

Contact Us

Please reach out to the organizers for any questions via this email: anirbit.mukherjee@manchester.ac.uk