Large Language Models
CPSC 4770/5770 • Spring 2026 • Yale University
Welcome to CPSC 4770/5770 – Large Language Models, From Foundations to Modern Practice!
This year, we have updated the syllabus and we’ll focus heavily on Language Modeling and recent advances in the field. The curriculum spans both foundational concepts and cutting-edge developments in the field including Large Language Models (LLMs). The course begins with core neural network concepts in NLP, covering word embeddings, sequence modeling, and attention mechanisms. Students will gain a strong understanding of these building blocks while learning their practical implementations. Building on these foundations, we explore transformer architectures and their evolution, including landmark language models such as GPT. The course examines how these models enable sophisticated language understanding and generation through pre-training and transfer learning. The latter portion covers contemporary advances: LLMs, parameter-efficient fine-tuning, post-training, and efficiency techniques. We’ll analyze the capabilities and limitations of current systems while discussing emerging research directions.
Prerequisites:
Intro to ML or Intro to AI are required.
Important Note: We strongly advise against taking this course if you do not meet the prerequisites.
Resources
- Dan Jurafsky and James H. Martin. Speech and Language Processing (2024 pre-release)
- Yoav Goldberg. A Primer on Neural Network Models for Natural Language Processing
We will also using papers from major conferences in the field including ACL, EMNLP, NAACL, ICLR, NeurIPS, etc.
- Lectures: Tue/Thur 2:30PM - 3:45PM
- Lecture Location: SSS 114 - Sheffield-Sterling-Strathcona 114
- Office Hours Location: 17HH #326
- Discussion: Ed Discussions
- Instructor Arman Cohan
- Email: arman.cohan@yale.edu
- Office hours: Tues 4:15-5:15pm (by appointment)
- Appointment: Select time
- Location: 17 Hillhouse Room 326
- TA Kaili Liu
- Email: kaili.liu@yale.edu
- Office hours: Fri 9:30-11am
- Location: 17 Hillhouse, 3rd Floor
- Research Interests: LLM uncertainty/calibration, LLM post-training & evaluation, Connections between LLMs & metacognition
- TA Austin Feng
- Email: austin.feng@yale.edu
- Office hours: Wed 3:30-5pm
- Location: 17 Hillhouse 3rd Floor
- Research Interests: Multimodal foundation models, Distributed pretraining optimizers, Reasoning
- TA Aakash Patel
- Email: aakash.patel.ap2853@yale.edu
- Office hours: Mon 4-5:30pm
- Location: 17 Hillhouse 3rd Floor
- Research Interests: Biomedical foundation models -- Interpretable and multimodal models for single-cell genomics, spatial transcriptomics, fMRI, and virtual screening & drug discovery
- TA Rohan Phanse
- Email: rohan.phanse@yale.edu
- Office hours: Thu 4-5:30pm
- Location: 17 Hillhouse 3rd Floor
- Research Interests: Retrieval-Augmented Generation, Reasoning
- TA Leyao Wang
- Email: leyao.wang.lw855@yale.edu
- Office hours: Mon 2:30-4pm
- Location: Dunham 401 Desk 13
- Research Interests: Trustworthy LLMs, Human-AI Collaboration
- TA Yilun Zhao
- Email: yilun.zhao@yale.edu
- Office hours: Thu 5-6:30pm
- Location: 17 Hillhouse 3rd Floor
- Research Interests: AI4Research, Knowledge-intensive reasoning, (Multimodal) Foundation models for specialized domains