Large Foundation Models: Mathematics, Algorithms, and Applications
Large foundation models have achieved remarkable success across various domains, including general applications like Natural Language Processing, image, speech, and video, as well as scientific fields such as materials science, molecular biology, and protein engineering. While the underlying techniques are firmly rooted in applied mathematics, their development has often been driven by empirical engineering practices, leading to significant practical breakthroughs. Diffusion models serve as a prime example, demonstrating both substantial engineering benefits and profound mathematical underpinnings.
This lecture series aims to bridge the gap between foundation models and their mathematical foundations, fostering interdisciplinary discussions, particularly between mathematics and machine learning.
Prerequisites:
-- For participants from a machine learning background, a solid understanding of Calculus and graduate-level Probability is required. While knowledge of stochastic processes is not strictly necessary, a basic understanding will significantly enhance comprehension.
-- For those from a mathematics background, prior knowledge of neural networks is essential. If you lack this prerequisite, please refer to the first three chapters of 'Neural Networks and Deep Learning' (http://neuralnetworksanddeeplearning.com).
Course Content:
The course will primarily cover autoregressive models, diffusion models, and discrete diffusion models, including their underlying mathematics and algorithms. We will explore their diverse applications across various domains, with a particular focus on multi-modalities.
This lecture series aims to bridge the gap between foundation models and their mathematical foundations, fostering interdisciplinary discussions, particularly between mathematics and machine learning.
Prerequisites:
-- For participants from a machine learning background, a solid understanding of Calculus and graduate-level Probability is required. While knowledge of stochastic processes is not strictly necessary, a basic understanding will significantly enhance comprehension.
-- For those from a mathematics background, prior knowledge of neural networks is essential. If you lack this prerequisite, please refer to the first three chapters of 'Neural Networks and Deep Learning' (http://neuralnetworksanddeeplearning.com).
Course Content:
The course will primarily cover autoregressive models, diffusion models, and discrete diffusion models, including their underlying mathematics and algorithms. We will explore their diverse applications across various domains, with a particular focus on multi-modalities.

讲师
日期
2025年09月15日 至 12月15日
位置
Weekday | Time | Venue | Online | ID | Password |
---|---|---|---|---|---|
周二,周四 | 10:40 - 12:15 | Shuangqing | ZOOM 01 | 928 682 9093 | BIMSA |
修课要求
Calculus, Probability I, Neural networks
听众
Undergraduate
, Advanced Undergraduate
, Graduate
, 博士后
, Researcher
视频公开
公开
笔记公开
不公开
语言
中文
, 英文
讲师介绍
胡丕丕于2014年在清华大学取得学士学位,2020年获得清华数学博士学位,2020年到2022年博士后期间在清华&BIMSA从事人工智能和数学交叉研究,2022年加入微软研究院担任高级研究员从事科学大模型工作,2025年加入BIMSA担任副教授职位。胡丕丕的研究兴趣主要是基础大模型的数学方法和应用,包括自回归扩散和离散扩散模型,科学大模型。