An optimization perspective on log-concave sampling
组织者
丘成栋
演讲者
康家熠
时间
2024年03月20日 15:00 至 15:30
地点
理科楼A-304
摘要
The primary contribution of this thesis is to advance the theory of complexity for sampling from a continuous probability density over R^d. Some highlights are included: a new analysis of the proximal sampler, taking inspiration from the proximal point algorithm in optimization; an improved and sharp analysis of the Metropolis-adjusted Langevin algorithm, yielding new state-of-the-art guarantees for high-accuracy log-concave sampling; the first lower bounds for the complexity of log-concave sampling; an analysis of mirror Langevin Monte Carlo for constrained sampling; and the development of a theory of approximate first-order stationarity in non-log-concave sampling.
演讲者介绍
Jiayi Kang received his Ph.D. in Mathematics from Tsinghua University in 2024. He joined the Beijing Institute of Mathematical Sciences and Applications (BIMSA) as an Assistant Researcher in July 2024, and became an Assistant Professor at the Hetao Institute for Mathematical and Interdisciplinary Sciences (HIMIS) in November 2025.
His research focuses on the intersection of deep learning, nonlinear filtering, and computational biology. His main research interests include: neural network-based filtering algorithms and their mathematical foundations, sampling methods in Wasserstein geometry, nonlinear filtering theory (including the Yau-Yau method) and its applications in climate science and other fields, as well as computational genomics and evolutionary system modeling. He is committed to solving complex problems in science and engineering using mathematical and machine learning methods.