Scientific Machine Learning SeminarThe main purpose of this seminar is to build a community of SciML researchers at Northwestern University and provide a venue where students can share their work and informal ideas. We hope the short and fun talks will spark new discussions, encourage collaboration, and create opportunities for feedback. Beyond presenting, the seminar also serves as a space to learn new approaches, exchange insights, and discover ongoing projects, helping everyone broaden their perspectives in scientific machine learning.2025 Fall
In Fall 2025, the Scientific Machine Learning seminar will be held TBD. Specific information about the format of each talk will be provided in the email announcement and posted below. If you're interested in attending the seminar, please contact Yiping Lu.Topics Interested
Diffusion Model1. Basic of diffusion model
2. Discrete diffusion model
3. Posterior Sampling with Diffusion modelInterest Level:Optimizers1. Quasi-Newton Methods, Gauss-Newton Methods
2. Adam and SOAP
3. signgd and MUONInterest Level:Operator Learning1. Basic of Operator Learning
2. Mathmetics of Operator LearningInterest Level:Syallbus
Time and Location: TBD2025 Fall SCiML Seminar Speaker Title Date: TBD Probabilistic Foundations of Diffusion Models and Monte Carlo PDE Solvers: From Particle Systems to Macroscopic Dynamics We study large systems of interacting stochastic particles undergoing motion, birth, and death, and explore the probabilistic mechanisms connecting microscopic randomness to macroscopic deterministic behavior. Each particle follows a stochastic differential equation with position-dependent drift and diffusion and experiences independent birth-death events modeled by exponential clocks. The state of the system is described through its empirical measure, for which we establish a martingale decomposition of test function observables. As the number of particles tends to infinity, the empirical measure converges in probability to a deterministic limit, recovering classical advection-diffusion-reaction PDEs as a law of large numbers. This probabilistic framework forms the basic foundation for modern diffusion models and Monte Carlo PDE solvers. We further illustrate how forward and backward stochastic differential equations can be interpreted to simulate, denoise, and reconstruct distributions, and discuss Feynman–Kac representations and particle methods, emphasizing their convergence and stochastic underpinnings. Reference: Date: TBD Yiping Lu Recent Advances in Optimization: From Quasi-Newton to Geometry-Aware Algorithms Optimization lies at the heart of modern machine learning and numerical computation. In this seminar, we provide a comprehensive overview of recent developments in optimization algorithms. We begin with classical and quasi-Newton methods, including BFGS, L-BFGS, and K-FAC, highlighting their efficiency in approximating second-order information. Next, we explore adaptive algorithms inspired by Adagrad, such as Adam, Shampoo, and SOAP, which leverage per-parameter scaling to accelerate convergence. Finally, we discuss geometry-aware methods, including SignGD and MUON, which exploit alternative geometrical structures to improve optimization performance. The seminar aims to give participants both a conceptual understanding and practical insights into these cutting-edge algorithms, bridging theoretical ideas with real-world applications. Reference: Date: TBD Yiping Lu Scaling Scientific Machine Learning: Integrating Theory and Numerics in Both Training and Inference Scaling scientific machine learning (SciML) requires overcoming bottlenecks at both training and inference. On the training side, we study the statistical convergence rate and limits of deep learning for solving elliptic PDEs from random samples. While our theory predicts optimal polynomial convergence for PINNs, optimization becomes prohibitively ill-conditioned as networks widen. By adapting descent strategies to the optimization geometry, we obtain scale-invariant training dynamics that translate polynomial convergence into concrete compute and yield compute-optimal configurations. On the inference side, I will introduce Simulation-Calibrated SciML (SCaSML), a physics-informed post-processing framework that improves surrogate models without retraining or fine-tuning. By enforcing physical laws, SCaSML delivers trustworthy corrections (via Feynman-Kac simulation) with approximate confidence intervals, achieves faster and near-optimal convergence rates, and supports online updates for digital twins. Together, these results integrate theory and numerics to enable predictable, reliable scaling of SciML in both training and inference. This is based on joint work with Lexing Ying, Jose Blanchet, Haoxuan Chen, Zexi Fan, Youheng Zhu, Shihao Yang, Jasen Lai, Sifan Wang, and Chunmei Wang. Reference:
Diffusion Model
1. Basic of diffusion model
2. Discrete diffusion model
3. Posterior Sampling with Diffusion model
2. Discrete diffusion model
3. Posterior Sampling with Diffusion model
Interest Level:
Optimizers
1. Quasi-Newton Methods, Gauss-Newton Methods
2. Adam and SOAP
3. signgd and MUON
2. Adam and SOAP
3. signgd and MUON
Interest Level:
Operator Learning
1. Basic of Operator Learning
2. Mathmetics of Operator Learning
2. Mathmetics of Operator Learning
Interest Level:
Time and Location: TBD
2025 Fall SCiML Seminar | |
---|---|
Speaker | Title |
Date: TBD | Probabilistic Foundations of Diffusion Models and Monte Carlo PDE Solvers: From Particle Systems to Macroscopic Dynamics |
We study large systems of interacting stochastic particles undergoing motion, birth, and death, and explore the probabilistic mechanisms connecting microscopic randomness to macroscopic deterministic behavior. Each particle follows a stochastic differential equation with position-dependent drift and diffusion and experiences independent birth-death events modeled by exponential clocks. The state of the system is described through its empirical measure, for which we establish a martingale decomposition of test function observables. As the number of particles tends to infinity, the empirical measure converges in probability to a deterministic limit, recovering classical advection-diffusion-reaction PDEs as a law of large numbers. This probabilistic framework forms the basic foundation for modern diffusion models and Monte Carlo PDE solvers. We further illustrate how forward and backward stochastic differential equations can be interpreted to simulate, denoise, and reconstruct distributions, and discuss Feynman–Kac representations and particle methods, emphasizing their convergence and stochastic underpinnings. | |
Reference: | |
Date: TBD | Yiping Lu | Recent Advances in Optimization: From Quasi-Newton to Geometry-Aware Algorithms |
Optimization lies at the heart of modern machine learning and numerical computation. In this seminar, we provide a comprehensive overview of recent developments in optimization algorithms. We begin with classical and quasi-Newton methods, including BFGS, L-BFGS, and K-FAC, highlighting their efficiency in approximating second-order information. Next, we explore adaptive algorithms inspired by Adagrad, such as Adam, Shampoo, and SOAP, which leverage per-parameter scaling to accelerate convergence. Finally, we discuss geometry-aware methods, including SignGD and MUON, which exploit alternative geometrical structures to improve optimization performance. The seminar aims to give participants both a conceptual understanding and practical insights into these cutting-edge algorithms, bridging theoretical ideas with real-world applications. | |
Reference: | |
Date: TBD | Yiping Lu | Scaling Scientific Machine Learning: Integrating Theory and Numerics in Both Training and Inference |
Scaling scientific machine learning (SciML) requires overcoming bottlenecks at both training and inference. On the training side, we study the statistical convergence rate and limits of deep learning for solving elliptic PDEs from random samples. While our theory predicts optimal polynomial convergence for PINNs, optimization becomes prohibitively ill-conditioned as networks widen. By adapting descent strategies to the optimization geometry, we obtain scale-invariant training dynamics that translate polynomial convergence into concrete compute and yield compute-optimal configurations. On the inference side, I will introduce Simulation-Calibrated SciML (SCaSML), a physics-informed post-processing framework that improves surrogate models without retraining or fine-tuning. By enforcing physical laws, SCaSML delivers trustworthy corrections (via Feynman-Kac simulation) with approximate confidence intervals, achieves faster and near-optimal convergence rates, and supports online updates for digital twins. Together, these results integrate theory and numerics to enable predictable, reliable scaling of SciML in both training and inference. This is based on joint work with Lexing Ying, Jose Blanchet, Haoxuan Chen, Zexi Fan, Youheng Zhu, Shihao Yang, Jasen Lai, Sifan Wang, and Chunmei Wang. | |
Reference: |