Pretraining, Instruction Tuning, Alignment, Specialization: On the Source of Large Language Model Abilities

报告题目:Pretraining, Instruction Tuning, Alignment, Specialization: On the Source of Large Language Model Abilities

报告人:符尧 (爱丁堡大学)

时间:2月10日  下午2:00–4:00

地点:王选所106会议室

Abstract:  Recently, the field has been greatly impressed and inspired by Large Language Models (LLMs) like GPT-3.5. The LLMs' multi-dimensional abilities are significantly beyond many NLP researchers’ and practitioners’ expectations and thus reshaping the research paradigm of NLP. A natural question is how LLMs get there, and where these fantastic abilities come from. In this talk we try to dissect the strong LLM abilities and trace them to their sources, hoping to give a comprehensive roadmap about the evolution of LLMs. 

Bio: Yao Fu is a Ph.D. student at University of Edinburgh and a student researcher at Allen Institute for AI. Previously he finished his M.S. in Columbia University and B.S. in Peking University. Yao studies large scale probabilistic generative models for human language. His publication covers topics of large language models, emergent abilities, and complex reasoning.

CLOSE

上一篇 没有了