We're happy to report that our Physics-based Flow Matching framework got an accept for ICLR'26! Physics-Based Flow Matching (PBFM) is a principled framework that explicitly targets Pareto-optimal solutions between physics-constraints and data-driven objectives.
09.02.2026 12:48
π 1
π 0
π¬ 0
π 0
Fast rotational equivariance for physics GNNs β the source code is now available: github.com/tum-pbs/stra...
Please also check out the full Physics-of-Fluids paper here: pubs.aip.org/aip/pof/arti...
07.02.2026 14:57
π 1
π 0
π¬ 0
π 0
We're very excited to report that our P3D Transformer was accepted at ICLR openreview.net/forum?id=8Ud...
We introduce a scalable hybrid CNNβTransformer architecture that pushes neural surrogate modeling into the regime of truly high-resolution 3D simulations.
03.02.2026 14:58
π 8
π 0
π¬ 0
π 0
Benchmarking Autoregressive Conditional Diffusion Models for Turbulent Flow Simulation
Simulating turbulent flows is crucial for a wide range of applications, and machine learning-based solvers are gaining increasing relevance. However, β¦
I'm very happy to report that our autoregressive predictions with generative diffusion models is _finally_ accepted π Congratulations Georg!
It's been a long journey, this paper was first submitted to NeurIPS'23, and now, almost 3y later, finally got accepted www.sciencedirect.com/science/arti...
30.01.2026 08:46
π 2
π 0
π¬ 0
π 0
Guiding diffusion models to reconstruct flow fields from sparse data
The reconstruction of unsteady flow fields from limited measurements is a challenging and crucial task for many engineering applications. Machine learning model
Great to see our paper on physics-constrained reconstruction / super-res with generative models posted online now at doi.org/10.1063/5.03... π
- PDE Transformer as backbone architecture
- differentiable physics constraints to guide
- and ConFIG as optimizer to resolve conflicts in the gradients
09.01.2026 11:25
π 5
π 0
π¬ 0
π 0
The SuperWing dataset is a large-scale, open dataset of transonic swept-wing aerodynamics, combining thousands of richly parameterized 3D wing geometries with high-fidelity RANS simulations across the operational flight envelope: arxiv.org/abs/2512.14397
17.12.2025 18:48
π 2
π 0
π¬ 0
π 0
Please join our mini symposium on "AI for Computational Fluid Dynamics - Opportunities and Challenges" MS279 , wccm-eccomas2026.org/event/area/8... at WCCM ECCOMAS in Munich next year in July (July 2026, wccm-eccomas2026.org). Inspiring discussions, and a proper "Mass" at the beergarden π»π
10.12.2025 15:41
π 2
π 0
π¬ 0
π 0
Our full course "advanced deep learning for physics" (ADL4P) is online now at tum-pbs.github.io/ADL4P/ π The course covers AI and neural network techniques for physics simulations & combinations with numerical methods. All recordings, slides and exercises are freely available!
09.12.2025 12:03
π 4
π 0
π¬ 0
π 0
Can AI surrogates outperform their training data? Turns out the answer is yes - with a few caveats π tum-pbs.github.io/emulator-sup... #neurips This surprising behavior leads to interesting and fundamental questions about the role of training data, and about how NN surrogates should be evaluated.
26.11.2025 19:58
π 5
π 0
π¬ 0
π 0
Can your AI surpass the simulator that taught it? What if the key to more accurate PDE modeling lies in questioning your training data's origins? π€
Excited to share my #NeurIPS 2025 paper with @thuereygroup.bsky.social: "Neural Emulator Superiority"!
14.11.2025 08:44
π 9
π 3
π¬ 1
π 1
Congratulations to Hao, Aleksandra and Bjoern for their NeurIPS paper tum-pbs.github.io/inc-paper/ π It analyzes how hybrid PDE solvers fundamentally and provably benefit from "indirect" (force-based) corrections rather than direct ones. Baking the corrections via INC reduces error growth!
25.11.2025 12:10
π 2
π 0
π¬ 0
π 0
I wanted to highlight that source code and data for our physics-based flow matching (PBFM) algorithm are online now at: github.com/tum-pbs/PBFM/ feel free to give it a try, and we'd be curious to hear how it works for you!
06.11.2025 09:49
π 5
π 0
π¬ 0
π 0
I'm happy to report that our collaborative project on 3D sparse-reconstruction and super-resolution with diffusion models, physics constraints and PDE Transformers is online now as preprint arxiv.org/abs/2510.19971 and source code github.com/tum-pbs/spar.... Great work Marc, Luis, Qiang and Luca π
29.10.2025 14:37
π 4
π 0
π¬ 0
π 0
I'm very excited to introduce P3D: our PDE-Transformer architecture in 3 dimensions by . Demonstrated for unprecedented 512^3 resolutions! That means the Transformer produces over 400 million degrees of freedom in one go π a regime that was previously out of reach: arxiv.org/abs/2509.10186
16.09.2025 07:40
π 10
π 2
π¬ 0
π 0
Congratulations to Bjoern for his accepted PoF paper on equivariant GraphNets π doi.org/10.1063/5.02...
the core idea is a very generic and powerful one: we compute a local Eigenbasis from flow features for equivariance. Mathematically it's identical to previous approaches, but faster and simpler π
09.09.2025 07:15
π 2
π 0
π¬ 0
π 0
I also wanted to mention that our paper detailing the differentiable SPH solver by Rene is online now on arxiv: arxiv.org/abs/2507.21684 If you're interested in fast and efficient neighborhoods, differentiable SPH operators and neat first optimization and learning tasks, please take a look!
01.08.2025 07:37
π 1
π 0
π¬ 0
π 0
Get ready for the PDE-Transformer: our new NN architecture tailored to scientific tasks π It combines hierarchical processing (UDiT), scalability (SWin) and flexible conditioning mechanisms. Code and paper available at tum-pbs.github.io/pde-transfor...
30.06.2025 19:05
π 17
π 1
π¬ 0
π 0
I'm really excited to share our latest work combining physics priors with probabilistic models: Flow Matching Meets PDEs - A Unified Framework for Physics-Constrained Generation , arxiv.org/abs/2506.08604 , great work by Giacomo and Qiang!
17.06.2025 14:36
π 6
π 0
π¬ 0
π 0
Have you faced challenges like SPH-based inverse problems, or learning Lagrangian closure models?
For these weβre excited to announce the first public release of DiffSPH , our differentiable Smoothed Particle Hydrodynamics solver.
Code: diffsph.fluids.dev
Short demo: lnkd.in/dYABSeKG
13.06.2025 14:08
π 3
π 1
π¬ 0
π 0
Congratulations to Bernhard for his first #SIGGRAPH paper! Great work π His two-phase Navier-Stokes solver is even more impressive given the fact that it's all done on a regular workstation, and without a GPU. Enjoy the sims in full screen & hi-quality here: youtu.be/nt9BohngvoE
04.06.2025 19:14
π 10
π 1
π¬ 0
π 0
Introducing PICT: the differentiable Fluid Solver for AI & machine learning in PyTorch
YouTube video by Nils Thuerey
I also just recorded a quick overview video for our new PICT solver: youtu.be/GGLidL0oT3s , enjoy! In case you missed it: PICT provides a new fully-differentiable multi-block Navier-Stokes solver for AI and learning tasks in PyTorch, e.g. learning turbulence closure in 3D
02.06.2025 12:38
π 5
π 1
π¬ 0
π 0
I'd like to highlight PICT, our new differentiable Fluid Solver built for AI & learning: github.com/tum-pbs/PICT
Simulating fluids is hard, and learning 3D closure models even harder: This is where PICT comes in β a GPU-accelerated, fully differentiable fluid solver for PyTorch π₯³
28.05.2025 15:41
π 6
π 1
π¬ 0
π 0
Learning Distributions of Complex Fluid Simulations with Diffusion Graph Networks
YouTube video by Mario Lino
I can highly recommend checking out Mario's talk about our Diffusion Graph Net paper from ICLR'25: www.youtube.com/watch?v=4Vx_... , enjoy!
21.05.2025 12:58
π 2
π 0
π¬ 0
π 0
I wanted to highlight PBDL's brand-new sections on diffusion models with code and derivations! Great work by Benjamin Holzschuh, with neat Jupyter notebooks π All the way from normalizing flow basics over score matching to denoising & flow matching. E.g., colab.research.google.com/github/tum-p...
13.05.2025 07:19
π 5
π 1
π¬ 0
π 0
If you're at #ICLR 2025 in Singapore, please check out our posters π€ I'm sure it's going to be a great conference! Have fun everyone...
23.04.2025 08:53
π 1
π 0
π¬ 0
π 0
I wanted to highlight that our project website (with code!) for our progressively-refined training with physics simulations is up now at: kanishkbh.github.io/prdp-paper/ #ICLR25 , the main ideas are: match network approximation and physics accuracy, refine the physics over the course of training.
11.04.2025 14:21
π 4
π 0
π¬ 0
π 0
The full PBDL book is available in a single PDF now arxiv.org/pdf/2109.05237, and has grown to 451 pages π³ Enjoy all the new highlights on generative models, simulation-based constraints and long term stability with diffusion models π
28.03.2025 08:23
π 5
π 0
π¬ 0
π 0
I'm very excited to highlight PBDL v0.3 www.physicsbaseddeeplearning.org, the latest version of our physics-based deep learning "book" π₯³ This version features a huge new chapter on generative AI, covering topics ranging from the derivation, over graph-based inference to physics-based constraints!
20.03.2025 19:36
π 6
π 1
π¬ 0
π 0
Congratulations to Kanishk and Felix π for their #ICLR'25 paper "Progressively Refined Differentiable Physics" kanishkbh.github.io/prdp-paper/ , the key insight is that training can be accelerated substantially by using fast approximates of the gradient (especially in early phases of training)
25.02.2025 07:06
π 2
π 0
π¬ 0
π 0
Congratulations to Youssef and Benjamin π for their #ICLR'25 paper on Truncated Diffusion Sampling openreview.net/forum?id=0Fb... It investigates several key questions of generative AI and diffusion for physics simulations to improve accuracy via Tweedie's formula
19.02.2025 01:29
π 6
π 0
π¬ 0
π 0