机器学习与数据科学博士生系列论坛(第九十四期)—— Beyond Pre-training: Shaping LLMs with Advanced Post-Training
报告人: 王迩东(北京大学)
时间:2025-11-13 16:00-17:00
地点:腾讯会议 331-2528-5257
Abstract:
While pre-training lays the groundwork for Large Language Models (LLMs), the true key to their refinement and alignment lies in post-training. This lecture delves into the critical techniques—such as fine-tuning, reinforcement learning, and test-time scaling—that transform a base LLM into a capable, reliable, and safe AI. We will systematically explore how these methods enhance reasoning, factual accuracy, and ethical alignment. The discussion will also address pivotal challenges like catastrophic forgetting and reward hacking, providing a comprehensive overview of the current landscape and future directions in evolving LLMs beyond their initial training.
About the Speaker:
该线上论坛每两周主办一次(除了公共假期)。论坛每次邀请一位博士生就某个前沿课题做较为系统深入的介绍,主题包括但不限于机器学习、高维统计学、运筹优化和理论计算机科学。
Your participation is warmly welcomed!

欢迎扫码关注北大统计科学中心公众号,了解更多讲座信息!