string data 2025
from
Monday 8 December 2025 (08:50)
to
Wednesday 10 December 2025 (20:45)
Monday 8 December 2025
08:50
Welcome
-
Sven Krippendorf
(
University of Cambridge
)
Welcome
Sven Krippendorf
(
University of Cambridge
)
08:50 - 09:00
09:00
Structure-Induced Interpretability in Kolmogorov–Arnold Networks
-
Fabian Ruhle
Structure-Induced Interpretability in Kolmogorov–Arnold Networks
Fabian Ruhle
09:00 - 09:45
Kolmogorov-Arnold Networks exhibit several properties beneficial to scientific discovery. I will outline these properties and show how they can be leveraged to build more interpretable AI models, which I will illustrate with an example from representation theory. I will also discuss how the intrinsic structure of KANs facilitates symbolic regression by employing ideas similar to Google's fun-search. Incorporating LLMs and vision transformers into the regression workflow allows for priming the NN with domain-specific symbolic regression targets. Finally, I will describe how KAN's intrinsic structure can be thought of as arising from wide, dense neural networks that have been sparsified in a specific way, and I will speculate on how this structural perspective relates to desirable empirical properties of KANs.
09:45
AdS-GNN: A conformally equivariant neural network
-
NABIL IQBAL
AdS-GNN: A conformally equivariant neural network
NABIL IQBAL
09:45 - 10:30
One important organizing principle for building such neural networks is that of symmetry, i.e. the idea that the symmetries of the problem should be encoded in the architecture of the network. I will provide an introduction to the resulting field of “geometric deep learning”. I will then discuss our construction of a neural network that transforms nicely under the conformal group, i.e. the set of transformations that preserve angles. I will describe how our construction leverages some of the kinematics of the AdS/CFT correspondence of quantum gravity, with potential applications to problems in computer vision and critical phenomena. Based on 2505.12880 with Maksim Zhdanov, Erik Bekkers, and Patrick Forré.
10:30
Coffee Break
Coffee Break
10:30 - 11:00
11:00
Building a Calabi-Yau Generator
-
Gary Shiu
(
University of Wisconsin-Madison
)
Building a Calabi-Yau Generator
Gary Shiu
(
University of Wisconsin-Madison
)
11:00 - 11:45
Fine, regular, and star triangulations (FRSTs) of four-dimensional reflexive polytopes give rise to toric varieties, within which generic anticanonical hypersurfaces yield smooth Calabi-Yau threefolds. We introduce CYTransformer, a deep learning model based on the transformer architecture, to automate the generation of FRSTs. We demonstrate that CYTransformer efficiently and unbiasedly samples FRSTs for polytopes across a range of sizes, and can self-improve through retraining on its own output. These results lay the foundation for AICY: a community-driven platform designed to combine self-improving machine learning models with a continuously expanding database to explore and catalog the Calabi-Yau landscape.
11:45
Deep Finite Temperature Bootstrap
-
Costis Papageorgakis
(
Queen Mary University of London
)
Deep Finite Temperature Bootstrap
Costis Papageorgakis
(
Queen Mary University of London
)
11:45 - 12:30
We introduce a neural network-based method to bootstrap crossing equations in Conformal Field Theory at finite temperature. Traditional approaches relying on positivity constraints or truncation schemes that discard infinite towers of operators are not applicable to this problem. Instead, we use MLPs to model spin-dependent tail functions that capture the combined contribution of infinitely many operators. The method formulates the Kubo-Martin-Schwinger (KMS) condition as a non-convex optimisation problem, combining thermal dispersion relations with neural network representations of the unknown data. We demonstrate the approach on Generalized Free Fields and apply it to extract double-twist thermal data in holographic CFTs.
12:30
Lunch break
Lunch break
12:30 - 13:30
13:30
Recent Progress on Axions in String Theory
-
David Marsh
(
King's College London
)
Recent Progress on Axions in String Theory
David Marsh
(
King's College London
)
13:30 - 14:00
The past few years have seen major advances in understanding the properties of axions in string theory. This progress is thanks to new computational tools that allow for fast and automated calculations with Calabi-Yau manifolds. I will describe the predictions string theory makes for axion masses, decay constants, and axion-photon couplings, and how these depend precisely on the topology of the Calabi-Yau. I will describe explicit constructions of fuzzy dark matter, and detailed calculations relating to decaying heavy axions, both of which seem to point to a preference for low reheating temperatures in cosmology. Lastly, I will describe the correlation between QCD axion mass and topology, and how this makes it possible for axion haloscopes to experimentally test the string theory landscape.
14:00
AInstein, atlas architecture vs embedding architecture
-
Tancredi Schettini Gherardini
(
University of Bonn
)
AInstein, atlas architecture vs embedding architecture
Tancredi Schettini Gherardini
(
University of Bonn
)
14:00 - 14:30
A numerical scheme based on semi-supervised machine learning, "AInstein", was recently introduced (see https://iopscience.iop.org/article/10.1088/3050-287X/ae1117) to approximate generic Riemannian Einstein metrics on a given manifold. Its versatility stems from encoding the differentiable structure directly in the loss function, making the method applicable to manifolds constructed in a "bottom-up" fashion that admit no natural embedding in R^n. After a brief review of the original AInstein model, we focus on a new architecture, adapted to the special case of real (n-1)-dimensional manifolds that can be embedded in R^n; this has the advantage that the neural-network ansatz is automatically globally defined. We present novel preliminary results obtained with the new architecture, concerning two open problems: the Kazdan–Warner (prescribed curvature) problem on S^2 and the existence of negative-curvature metrics on S^4, S^5. Finally, if time permits, we will briefly comment on a further ongoing extension of AInstein to Lorentzian metrics and black holes.
14:30
Solving inverse problems of Type IIB flux vacua with conditional generative models
-
Zhimei Liu
(
University of Cambridge
)
Solving inverse problems of Type IIB flux vacua with conditional generative models
Zhimei Liu
(
University of Cambridge
)
14:30 - 15:00
We address the inverse problem in Type IIB flux compactifications of identifying flux vacua with targeted phenomenological properties such as specific superpotential values or tadpole constraints using conditional generative models. These machine learning techniques overcome computational bottlenecks in traditional approaches such as rejection sampling and Markov Chain Monte Carlo (MCMC), which struggle to generate rare, finely-tuned vacua. As a proof of concept, we demonstrate that conditional generative models provide a more efficient alternative, specifically using conditional variational autoencoders (CVAEs). We introduce a CVAE framework tailored to flux compactifications, incorporating physical constraints directly into the loss function — enabling the generation of physically consistent vacua beyond the training set. Our experiments on conifold and symmetric torus background geometries show that the CVAE achieves a speedup of about O(10^3) compared to Metropolis sampling, particularly in narrow target ranges for superpotential values. Additionally, the CVAE generates novel, distinct flux configurations beyond the training data, highlighting its potential for probing computationally challenging regions of the string landscape. Our results establish conditional generative models as a powerful and scalable tool for targeted flux vacua generation, opening new pathways for model building in regions of the landscape previously inaccessible by traditional model building techniques.
15:00
Coffee break with Posters
Coffee break with Posters
15:00 - 15:30
15:30
BART-L: Towards a Foundation Model for Theoretical High-Energy Physics
-
Yong Sheng Koay
BART-L: Towards a Foundation Model for Theoretical High-Energy Physics
Yong Sheng Koay
15:30 - 16:00
Transformers excel at natural language processing, raising the question of whether they can also learn the mathematical structures underlying particle physics. We introduce BART-L, a transformer-based model trained to generate particle physics Lagrangians from field content and symmetry information. Trained on Lagrangians consistent with the Standard Model gauge group SU(3)×SU(2)×U(1), BART-L achieves over 90% accuracy for expressions involving up to six matter fields. Embedding analyses indicate that the model internalizes concepts such as group representations and conjugation operations, despite not being explicitly trained for them. We further study its out-of-distribution performance to identify architectural factors limiting generalization. This framework provides an early indication of what a foundation model for theoretical high-energy physics might look like, along with its potential capabilities and inherent limitations.
16:00
Searching for new Special Lagrangians with Quality-Diversity Optimization
-
Yidi Qi
(
Northeastern University
)
Searching for new Special Lagrangians with Quality-Diversity Optimization
Yidi Qi
(
Northeastern University
)
16:00 - 16:30
Special Lagrangian (sLags) submanifolds are crucial objects for string phenomenology and the SYZ conjecture, yet their explicit construction remains a significant challenge in geometry. In this talk, we introduce a novel computational framework to tackle this problem. Our approach leverages a Quality-Diversity (QD) search algorithm to navigate families of parametrized geometries. This method is engineered to simultaneously reward approximations of the sLag condition while maximizing geometric diversity. Our preliminary findings showcase the potential of this QD framework as a powerful new tool in the ongoing search for novel sLag constructions.
16:30
Chasing the Standard Model within heterotic line bundle constructions
-
Luca Nutricati
Chasing the Standard Model within heterotic line bundle constructions
Luca Nutricati
16:30 - 17:00
In this talk, we explore the phenomenological potential of heterotic line bundle models as a promising framework for deriving the Standard Model from string theory. We present a systematic approach to constrain the low-energy effective theories derived from such compactifications using remnants of anomalous U(1) symmetries to retrieve realistic quarks and leptons masses, mixing patterns and light Higgs. We show that these features emerge in specific regions of moduli space, where vacuum expectation values of geometric moduli and bundle singlets eventually align to suppress unwanted $R$-parity violating couplings. This work underscores the viability of heterotic line bundle models as a pathway to connecting string theory with observable particle physics.
Tuesday 9 December 2025
09:00
PhysLean: Digitalizing physics into Lean
-
Joseph Tooby-Smith
PhysLean: Digitalizing physics into Lean
Joseph Tooby-Smith
09:00 - 09:45
This talk will discuss an open-source, community run, project called PhysLean, which aims to write physics into the interactive theorem prover Lean 4. That is, a computer programming language where mathematical correctness is guaranteed. There are numerous benefits to this endeavor including correctness, searchability of results, and the ability to interact with AI in a way safe from (certain) hallucinations. In this talk, I will give an overview of this project, and briefly discuss its use in String theory.
09:45
Ricci-flat Calabi-Yau metrics and Yukawa couplings
-
Andre Lukas
(
U
)
Ricci-flat Calabi-Yau metrics and Yukawa couplings
Andre Lukas
(
U
)
09:45 - 10:30
10:30
Coffee break
Coffee break
10:30 - 11:00
11:00
Sampling IIB Flux Vacua Using Generative Models
-
Magdalena Larfors
(
Uppsala university
)
Sampling IIB Flux Vacua Using Generative Models
Magdalena Larfors
(
Uppsala university
)
11:00 - 11:45
We discus the application of two types of generative models are applied to the construction of ISD flux vacua of type IIB string theory. The models under study are Bayesian Flow Networks (BFNs) and Transformers. We find that both models demonstrate good performance, in particular on interpolation and conditional sampling. Some extrapolation capabilities are observed, which could be leveraged in future studies to generate flux vacua in less well explored settings.
11:45
Deep Reinforcement Learning for Fano Hypersurfaces
-
Marc Truter
Deep Reinforcement Learning for Fano Hypersurfaces
Marc Truter
11:45 - 12:30
In this talk, we show how deep reinforcement learning can be used to overcome the computational challenges faced by conventional search algorithms in the discovery of new Fano hypersurfaces. By the Minimal Model Program, varieties can be reduced to three building blocks: Fano, Calabi-Yau, and general type varieties. The building blocks have singular points, which have a property known as terminality. By a result of Birkar there are only finitely many Fano varieties with terminal singularities per dimension, and so one can construct `periodic tables'. Tables are complete in dimensions 1, 2 and partially in 3. Little is known in dimension 4. My work aims to classify terminal Fano 4-fold hypersurfaces. There are two obstacles, however; determining terminality is difficult and computationally expensive, and the occurrence of terminal examples is sparse in a vast search space. Together, this makes the discovery of new examples difficult. I will discuss how creating a deep reinforcement learning search architecture allowed me to find 70,000 examples in under an hour from the 10,000 I was able to produce in days using conventional methods. Interestingly, from nearly all of the examples we have discovered so far, we are able to construct associated nontoric Calabi-Yau 3-fold hypersurfaces with canonical singularities. Their ambient weighted projective spaces are nonreflexive polytopes, and thus are not necessarily in the famous Kreuzer-Skarke list of 473,800,776.
12:30
Lunch
Lunch
12:30 - 14:00
14:00
Introduction: Outward facing afternoon
Introduction: Outward facing afternoon
14:00 - 14:05
14:05
Is AI like hadron physics?
-
James Halverson
Is AI like hadron physics?
James Halverson
14:05 - 15:05
Machine learning (ML) systems fueled by neural networks have entered our daily lives and led to scientific breakthroughs, but many open questions remain. After a nod toward the question of rigor with ML and recent progress, I'll turn to the theory of neural networks. I will argue that understanding neural networks inevitably leads to ideas from quantum field theory, the theoretical framework underlying much of modern physics. This connection was realized in the simplest case in the 1990s. I will then propose that the connection might be more general, an NN-FT correspondence of sorts, with neural networks providing a way to define a field theory. The apparent non-sequitur in the title will be used as a rhetorical device to explore where we are and where we'd like to go.
15:05
Coffee break
Coffee break
15:05 - 15:40
15:40
Panel Discussion
Panel Discussion
15:40 - 16:40
Wednesday 10 December 2025
09:00
Learning the Landscape: Differentiable Frameworks for String Theory
-
Andreas Schachner
(
Cornell University
)
Learning the Landscape: Differentiable Frameworks for String Theory
Andreas Schachner
(
Cornell University
)
09:00 - 09:45
String theory naturally gives rise to vast, high-dimensional datasets, yet systematic investigations of this landscape has long been impeded by the complexity of moduli spaces, quantum corrections, and the vastness of flux configurations. In this talk, I present new differentiable frameworks for string compactifications that combine automatic differentiation, and ML-based inference to construct and analyse four-dimensional effective theories. These pipelines enable controlled moduli stabilisation, incorporate strong coupling corrections, and allow efficient searches for phenomenologically relevant vacua, including regimes with small cosmological constant and rich axion sectors. By integrating analytic insights with scalable computational tools, this approach transforms the study of the string landscape into a quantitative and data-driven enterprise, opening new avenues for connecting quantum gravity to observable physics.
09:45
The First Large-Scale Computation of the Gukov–Manolescu Series
-
Lara San Martin Suarez
(
Caltech
)
The First Large-Scale Computation of the Gukov–Manolescu Series
Lara San Martin Suarez
(
Caltech
)
09:45 - 10:30
The Z-hat invariant defined by Gukov–Pei–Putrov–Vafa is a BPS-counting q-series built from Calabi–Yau geometry, whose values at roots of unity recover the Witten–Reshetikhin–Turaev invariants of 3-manifolds. In this talk, I will review a construction for its knot-theoretic counterpart, the Gukov–Manolescu series F_K, and present the first large-scale computation of F_K for 1,246 knots, obtained via an extensive search in the braid space and a fast C++ implementation of the F_K state-sum model.
10:30
Coffee Break
Coffee Break
10:30 - 11:00
11:00
Learning some Physics and Maths
-
Miranda Cheng
Learning some Physics and Maths
Miranda Cheng
11:00 - 11:45
In this talk, I present new results on (machine) learning of physics and mathematics. First, we report on advancements in training flow-based models to simulate quantum field theories (QFT). We advocate adopting physics-inspired generation paths and exploiting the physics information gained by learning such flows. The latter can be applied to simulate continuous families of theories in one go, and leads to novel and efficient methods to compute connected correlation functions. Second, in the domain of mathematical physics, we analyze datasets consisting of infinite q-series, which appear as QFT partition functions, topological invariants, or quantum modular forms. We introduce a data-analysis pipeline designed to represent, manipulate, and interpret this data. As concrete examples, we analyse how a classifier learns the homology groups of three-manifolds, and present evidence of unexpected relations between specific three-manifold topological invariants and cobordism invariants. The first part will be based on work done with Tobias Göbel and Mathis Gerdes, while the second based on work with Brandon Robinson, Shimal Harichurn, Fabian Ruehle, Sergei Gukov, and Rak-Kyeong Seong.
11:45
Deep Learning based discovery of Integrable Systems
-
Evgeny Sobko
Deep Learning based discovery of Integrable Systems
Evgeny Sobko
11:45 - 12:30
Integrable systems are exactly solvable models that play a central role in QFT, string theory and statistical physics offering an ideal setting for understanding complex physical phenomena and developing novel analytical methods. However, the discovery of new integrable systems remains a major open challenge due to the nonlinearity of the Yang–Baxter equation (YBE) that defines them, and the vastness of its solution space. Here we present the first AI-based framework that enables the discovery of new quantum integrable systems in exact analytical form. Our method combines an ensemble of neural networks, trained to identify high-precision numerical solutions to the YBE, with an algebraic extraction procedure based on the Reshetikhin integrability condition, which reconstructs the corresponding Hamiltonian families analytically. When applied to spin chains with three- and four-dimensional site spaces, we discover hundreds of previously unknown integrable Hamiltonians. Remarkably, these Hamiltonians organize into rational algebraic varieties, and we conjecture that this rationality holds universally — revealing a deep and previously unexplored connection between quantum integrability and algebraic geometry. By unlocking integrable systems far beyond the reach of traditional methods, this AI-driven approach substantially expands the landscape of exactly solvable models and opens a scalable path to further discoveries.
12:30
Lunch
Lunch
12:30 - 13:30
13:30
QM, SUSY, and Outlook in NN-FT
-
James Halverson
QM, SUSY, and Outlook in NN-FT
James Halverson
13:30 - 14:15
I'll review the essentials of a neural network approach to defining and studying field theories, listing previous results related to symmetries, conformal symmetry, locality, interactions, etc. The focus of the talk will be on two 2025 papers, one related to QM, and the other related to fermions and SUSY. The QM results include a universality theorem, a new construction of Euclidean QM theories with deep neural networks, and the recovery of cherished properties from QM, such as Heisenberg uncertainty and the spectrum of the harmonic oscillator. The fermion and SUSY results include the definition of Grassmann-valued neural networks and their use in constructing fermionic NN-FTs, including the free Dirac spinor in an infinite width limit. With fermions in hand, I will also present a new construction of interacting supersymmetric theories in 1d and 4d. I'll try to devote a significant portion at the end for outlook and also getting feedback from friends.
14:15
String Data for Maths and Physics
-
Andrei Constantin
String Data for Maths and Physics
Andrei Constantin
14:15 - 15:00
15:00
Coffee Break
Coffee Break
15:00 - 15:30
15:30
Machine Learning Aspects of Gravity
-
Vishnu Jejjala
Machine Learning Aspects of Gravity
Vishnu Jejjala
15:30 - 16:15
From holography, we know that gravitational systems have codimension one degrees of freedom. Through a number of experiments, we use machine learning to study physical and mathematical aspects of black hole entropy.
16:15
Wrap up
-
Sven Krippendorf
(
University of Cambridge
)
Wrap up
Sven Krippendorf
(
University of Cambridge
)
16:15 - 16:25