Invited Speakers Talks
Steve Awodey

Steve Awodey
Carnegie Mellon University &
Royal Society Wolfson Visiting Fellow

What is HoTT?

Homotopy Type Theory (HoTT) is a new field of research combining constructive type theory, homotopy theory, and higher category theory, with applications to the formalization of mathematics. This survey talk will explain the surprising connection between the foundations of mathematics, algebraic topology, and formalization of mathematics, and report on a case study in formal theorem proving in the years since its discovery. The methodology of synthetic mathematics provides a rich area for potential applications of AI to mathematics.
Dominique Beaini

Dominique Beaini (online)
Valence Discovery & Mila

How to learn molecules?

How can we build powerful representation of molecules for drug discovery? We'll dive in the different algorithms, the challenges that they face in terms of expressivity, and how to overcome them. We'll discuss building positional and structural encodings, and how they can make GNNs quite expressive without algorithmic changes. Then, we'll see how to build graph Transformers for pre-training on large molecular datasets, and scale them to infinity. Finally, we'll step into the world of multi-modality to learn how molecules impact human cells from a morphological perspective.
Tony Cohn

Tony Cohn
Alan Turing Institute &
University of Leeds

Can Large Language Models Solve Spatial Puzzles?

Spatial reasoning is a core component of an agent’s ability to operate in or reason about the physical world. LLMs are widely promoted as having abilities to reason about a wide variety of domains, but much of our human ability to reason about spatial and temporal information is grounded in our acquired knowledge of the physical world as an embodied agent. Can a disembodied LLM reason about problems involving spatial and temporal information? In this talk I will discuss the ability of state-of-the-art LLMs to reason about qualitative spatial and temporal information. Across a wide range of LLMs, although they usually show abilities rather better than chance, they still struggle with many questions and tasks, for example when reasoning about directions, topological relations, or temporal points and intervals.
Cristina Cornelio

Cristina Cornelio
Samsung AI in Cambridge

Derivable Scientific Discovery

Scientific progress has long relied on discovering new laws through domain expertise and experimental validation. Modern AI can now generate candidate hypotheses at unprecedented scale and speed. Yet this creates a critical bottleneck: without rigorous, scalable verification, the volume of AI-generated hypotheses risks overwhelming discovery rather than accelerating it. Verification is not an afterthought but the foundation of meaningful AI-assisted science. This talk presents a vision for scientific discovery in which models are derivable from explicit axioms while using minimal experimental data. This allows understanding why a law holds, when to trust it, and what must change when it fails. I will present three complementary systems we developed to achieve this: 1) AI-Descartes: uses symbolic regression to propose candidate models from data, then applies logical reasoning to select those most consistent with established axioms; 2) AI-Hilbert: integrates polynomial optimization with logical constraints, enforcing theoretical consistency and empirical validity simultaneously; and 3) AI-Noether: when current theory cannot derive a hypothesis, it proposes a minimal set of new axioms that make the hypothesis derivable. Together, these methods establish a new paradigm of "derivable scientific discovery", where integrating data and logic transforms AI from a mere hypothesis generator into a system that produces meaningful, verifiable laws.
Thomas Fink

Thomas Fink
London Institute for Mathematical Science

Deep-layered machines have a built-in Occam's razor

Input-output maps are prevalent throughout science and technology. They are empirically observed to be biased towards simple outputs, but we don’t understand why. To address this puzzle, we study the archetypal input-output map: a deep-layered machine in which every node is a Boolean function of all the nodes below it. We give an exact theory for the distribution of outputs, and we confirm our predictions through extensive computer experiments. As the network depth increases, the distribution becomes exponentially biased towards simple outputs. This suggests that deep-layered machines and other learning methodologies may be inherently biased towards simplicity in the models that they generate.
Stefania Fresca

Stefania Fresca
University of Washington

Embedding mathematical structure in neural reduced order models for parametrized PDEs

Solving differential problems using full order models (FOMs), such as the finite element method, may result in prohibitive computational costs, particularly in real-time simulations and multi-query routines. Neural reduced order models (ROMs) provide an efficient surrogate framework for parametrized PDEs, but purely data-driven approaches may fail to preserve key physical, mathematical and numerical properties of the underlying physical system. In this talk, mathematical structure is embedded into neural ROMs to move beyond purely data-driven approximation capabilities. In particular, inductive biases derived from the FOM, reflecting physical principles, numerical properties, and geometric variability, are incorporated directly into the model architecture and training procedure.
Moshe Eliasof

Moshe Eliasof
University of Cambridge

A Graph Learning Perspective on Quadratic Binary Optimization

Quadratic unconstrained binary optimization (QUBO) arises in many real-world decision and optimization problems, including portfolio optimization, manufacturing and supply chain optimization, telecommunications network optimization, and resource allocation. In this talk, I will discuss a link between graph neural networks (GNNs) and QUBO. Viewing the QUBO matrix as a graph and the observed vector as node features leads to a graph learning perspective on these computationally challenging problems. I will show how the behavior of the QUBO solution with respect to the observed vector motivates treating QUBO as a heterophilic node classification problem. Building on this perspective, I will present QUBO-GNN, a physics-inspired graph neural architecture equipped with QUBO-aware features, together with a self-supervised data-generation mechanism that enables scalable training. 
Ling Guo

Ling Guo
Shanghai Normal University

Uncertainty Quantification and Model Discrepancy in Scientific Machine Learning

Uncertainty quantification (UQ) is essential for reliable scientific machine learning, especially under sparse and noisy observations and in the presence of model misspecification. This talk presents a coherent progression of methods toward trustworthy learning of physical systems. We first develop an information-bottleneck UQ approach for operator learning, using a confidence-aware encoder and a Gaussian decoder to produce calibrated predictive means and variances without costly posterior sampling. We then extend the confidence-aware latent representation to physics-informed learning for PDEs under sparse and noisy data via a latent variable model coupled with Gaussian processes, enabling uncertainty-aware forward and inverse modeling with soft physics constraints. Finally, we introduce a latent-space model-correction strategy with dual decoders for the solution and a discrepancy term, providing joint UQ of both within a single training pipeline. Numerical experiments demonstrate improved robustness and efficiency across representative operator-learning and PDE benchmarks.
Anders C. Hansen

Anders C. Hansen
Cambridge

Necessary mechanisms for super AI and stopping hallucinations--The consistent reasoning paradox and the indeterminacy function

Creating Artificial Super Intelligence (ASI) (AI that surpasses human intelligence) is the ultimate challenge in AI research. This is, as we will discuss, fundamentally linked to the problem of avoiding hallucinations (wrong, yet plausible answers) in AI. We will describe a key mechanism that must be present in any ASI. This mechanism is not present in any modern chatbot and we will discuss how, without it, ASI will never be achievable. Moreover, we reveal that AI missing this mechanism will always hallucinate. Specifically, this mechanism is the computation of what we call an indeterminacy function. An indeterminacy function determines when an AI is correct and when it will not be able to answer with 100% confidence. The root to these findings is the Consistent Reasoning Paradox (CRP), which is a new paradox in logical reasoning that we will describe in the talk. The CRP shows that the above mechanism must be present as – surprisingly – an ASI that is ‘pretty sure’ (more than 50%) can rewrite itself to become 100% certain. It will compute an indeterminacy function and either be correct with 100% confidence, or it will not be more than 50% sure. The CRP addresses a long-standing issue that stems from Turing’s famous statement that infallible AI cannot be intelligent, where he questions how much intelligence may be displayed if an AI makes no pretence at infallibility. The CRP answers this – consistent reasoning requires fallibility – and thus marks a necessary fundamental shift in AI design if ASI is to ever be achieved and hallucinations to be stopped.
Yang-Hui He

Yang-Hui He (online)
London Institute for Mathematical Science

AI and the Future of Mathematics (online)

We argue how AI can assist mathematics in three ways: theorem-proving, conjecture formulation, and language processing. Inspired by initial experiments in geometry and string theory in 2017, we summarize how this emerging field has grown over the past years, and show how various machine-learning algorithms can help with pattern detection across disciplines ranging from algebraic geometry to representation theory, to combinatorics, and to number theory. At the heart of the programme is the question how does AI help with theoretical discovery, and the implications for the future of mathematics.

FEAT: Free energy Estimators with Adaptive Transport

We present Free energy Estimators with Adaptive Transport (FEAT), a novel framework for free energy estimation -- a critical challenge across scientific domains. FEAT leverages learned transports implemented via stochastic interpolants and provides consistent, minimum-variance estimators based on escorted Jarzynski equality and controlled Crooks theorem, alongside variational upper and lower bounds on free energy differences. Unifying equilibrium and non-equilibrium methods under a single theoretical framework, FEAT establishes a principled foundation for neural free energy calculations. Experimental validation on toy examples, molecular simulations, and quantum field theory demonstrates improvements over existing learning-based methods.
Asim Munawar

Asim Munawar (online)
IBM Watson Center

Small Language Models for Enterprise Agentic Workflows

Large language models dominate current discussions on agentic AI, but their cost and complexity often limit enterprise adoption. Our research explores how small language models can be equipped with function calling, reasoning, and planning capabilities to perform effectively in enterprise workflows. Agentic AI enables systems to reason over complex states, call tools, and plan multi-step actions. I will present methods for aligning small models with these tasks, leveraging synthetic data for adaptation, and evaluating their reliability in real-world automation settings. The goal is to show how efficient models can deliver trustworthy, enterprise-grade agentic AI solutions. While our design principles are broadly applicable, I will focus on IT automation—drawing on IT Bench from IBM, BFCL v4, and other agentic benchmarks. IT automation is critical because enterprises depend on reliable, scalable, and secure IT services, and maintaining this reliability requires continuous monitoring, rapid incident response, and efficient remediation.
Alessandro Sperduti

Alessandro Sperduti
Università di Padova

Learning neuro-symbolic convergent term rewriting systems

Building neural systems that can learn to execute symbolic algorithms is a challenging open problem in artificial intelligence, especially when aiming for strong generalization and out-of-distribution performance. In this talk, I introduce a general framework for learning convergent term rewriting systems using a neuro-symbolic architecture inspired by the rewriting algorithm itself. I present two modular implementations of such architecture: the Neural Rewriting System (NRS) and the Fast Neural Rewriting System (FastNRS). As a result of algorithmic-inspired design and key architectural elements, both models can generalize to out-of-distribution instances, with FastNRS offering significant improvements in terms of memory efficiency, training speed, and inference time. We evaluate both architectures on four tasks involving the simplification of mathematical formulas and further demonstrate their versatility in a multi-domain learning scenario, where a single model is trained to solve multiple types of problems simultaneously.
Barbara Tversky

Barbara Tversky
Stanford University &
Columbia Teachers College

Mind in Motion: How Action Shapes Thought

I will argue that spatial thinking is the foundation of thought, not the entire edifice, but the foundation. I will bring support from neuroscience, language, gesture, and visualizations and bring them together with the notion of spraction, actions in space create abstractions. I will also put forth that these findings about human thought and creativity present challenges to current GenAI.
Petar Veličković

Petar Veličković
Google DeepMind

Please maximise signal... Do you copy? ...

We are heading toward a world where large language model (LLM)-based systems drive general-purpose computation. It is hence important to assess the extent to which such models robustly perform this computation. We will focus on basic tasks that often form part of a larger-scale system invocation—such as predicting maxima and input copying—to demonstrate that contemporary decoder-only Transformers cannot robustly perform these tasks, or even always know when they're wrong.
Pei Wang

Pei Wang
Temple University

A Model of Reasoning that is Both Normative and Realistic

This talk presents NARS (Non‑Axiomatic Reasoning System), an AGI framework built upon the Assumption of Insufficient Knowledge and Resources (AIKR). Within this paradigm, intelligence is defined not as the pursuit of absolute truths or optima, but as the capacity to utilize available knowledge and resources to achieve goals in a changing environment. NARS employs an experience‑grounded semantics in which concepts are abstractions of experience and truth‑values quantify evidential support. Knowledge is organized as a continually evolving concept graph, where a unified term logic drives reasoning, learning, and other cognitive processes across the structure. The system integrates multimodal experience, including spatiotemporal perception, linguistic input, and introspective signals. By constructing inference processes under real-time pressure, NARS is guided by a dynamic priority distribution over tasks, beliefs, and concepts. The result is a rigorous and flexible reasoning engine that stands in contrast to the statistical nature of Large Language Models (LLMs), while still allowing LLMs to serve as complementary tools when appropriate.
Yuguang Wang

Yu-Guang Wang
Shanghai Jiaotong University

AI Antibody Design Superintelligent Agent: An All-Atom Modeling and Automated Laboratory Closed-Loop Framework for Multi-Objective Developability Optimization

Antibody drug discovery is constrained by the vastness of the design space, the long cycles of wet-lab experimentation, and the difficulty of jointly optimizing multiple competing objectives. We developed an AI antibody design superintelligent agent that systematically integrates all-atom protein modeling, diffusion- and flow-matching–based sequence editing, cross-modal text-to-sequence retrieval, and zero-code training and deployment capabilities into an automated laboratory workflow, forming a closed-loop dry–wet data flywheel. In each round, the AI designs 96 candidate sequences, which undergo automated gene synthesis, expression, and multidimensional developability assessment before being fed back for model fine-tuning. Within at most three iterative cycles, the system achieves or surpasses state-of-the-art benchmarks across affinity, stability, expression yield, specificity, and immunogenicity. The platform demonstrates high hit rates and strong generalization across three tumor and immune targets, and further enhances clinical translatability through organoid-based toxicity evaluation. These results show that the deep coupling of AI and automated laboratories can elevate antibody design from single-point optimization to multi-objective coordinated optimization, providing a systematic and auditable framework for iterative biologics R&D.
Kelin Xia

Kelin Xia
NTU Singapore

Mathematical AI: from topological data analysis to topological deep learning

A central challenge in artificial intelligence (AI)-driven molecular science lies in efficiently representing molecular data and developing learning architectures that capture intrinsic structure-function relationships. In this work, we introduce advanced mathematics-based molecular representations and learning frameworks. Molecular structures and interactions are encoded using high-order topological and algebraic representations, including Rips complexes, Alpha complexes, Neighborhood complexes, Dowker complexes, Hom-complexes, Tor-algebras, Rhomboid tilings, Sheaves, Categories, etc. Building on these foundations, we design physics-informed geometric and topological deep learning models that systematically integrate high-order, multiscale, and periodic information of molecular systems. These models have been successfully applied to diverse molecular datasets across chemistry, biology, and materials science, demonstrating their versatility and effectiveness in uncovering complex structural-functional relationships.