Tuesday, 21 April, 2026
- Share
- Share on Facebook
- Share on LinkedIn
High Performance Online Deep Neural Network Training from Synthetic Data with Active Learning for Scientific Computing (or: "How to train neural networks to model physical systems accurately — through smarter data, not more compute?")
Lay summary
Scientific simulations are expensive. Deep learning surrogates promise to replace them — but training one still requires running hundreds or thousands of simulations, sampled blindly before training even begins. This creates a compounding problem: costly data, poor coverage, and a workflow that cannot adapt to what the model actually needs. The field has focused on building better models; this thesis argues the bottleneck is in the data.
We build on an online training framework in which simulation data is streamed directly into the training process, removing the need for data storage. Within this setting, we develop active learning methods that monitor training progress and steer data creation toward the most informative configurations — spending the compute budget where it matters most. The proposed methods are lightweight, model-agnostic, and show consistent gains in surrogate accuracy and reliability across diverse physical systems and model architectures.
Date and place
Tuesday, 21 April at 15:00
IMAG Building, ground floor, Seminar Room 1
- Share
- Share on Facebook
- Share on LinkedIn