CS40008.01 NLP & LLMs

Spring 2026 @ Fudan University

View the Project on GitHub baojian/llm-26

Outline


Course Overview

Introduction

This course covers the foundations and modern frontiers of Natural Language Processing (NLP), with a heavy emphasis on Large Language Models (LLMs). You will learn the modern pipeline of building effective LLMs from basic tokenization to training, fine-tuning, and deploying modern LLM architectures.

Basic Info


Assignments and Course Project

All standard homework assignments are completed by Week 12. The final month (Weeks 13–16) is dedicated exclusively to the Course Project. Please submit your homework at https://elearning.fudan.edu.cn/

Assignment 1. Foundations of Text

Course Project


Coursework

Resources and References


GPU Resources


Weekly Schedule

Week 1 Introduction to LLMs

In our first lecture, we introduce text preprocessing, including tokenization (BPE/WordPiece), and vocab design.

Release Assignment 1

Week 2 N-Gram Language Models

In this lecture, we introduce the concept of MLE, Smoothing, Perplexity, and Language Modeling basics.

Week 3 Word Embeddings

In this lecture, we introduce text classification, Word2Vec, Distributional Hypothesis, and Intrinsic/Extrinsic evaluations.

Week 4 Neural LMs

In this lecture, we introduce neural networks and how to build NN models for sequence learning problems. We will discuss some classic models like LSTM and how the encoder-decoder style models developed and why the attention is a effective component adding to encoder-decoder model.

Week 5 Attention Mechanisms and Transformer

In this lecture, we introduce the Transformer architecture.

Week 7 LLM Pretraining

Causal LM vs MLM, Chinchilla Scaling Laws, Data Mixtures

Week 8 Fine-tuning & PEFT

SFT, LoRA/QLoRA, Adapters, Instruction Tuning

Project Proposal Due

Week 10 Evaluation

Benchmarks (MMLU/GSM8K), Contamination, LLM-as-a-judge

Week 11 Prompt Engineering

In-context learning, Chain-of-Thought, Prompt sensitivity

Week 12 Course Project Presentation

Week 13 RAG Systems

Dense Retrieval, Vector DBs, Reranking, Grounding

Week 14 Alignment & Safety

RLHF (PPO/DPO), Safety barriers, Red-teaming

Week 15 Efficiency & Systems

KV Caching, Quantization (Int8/FP4), Latency/Throughput

Week 16 Agents and Frontiers

Multimodal LLMs, Diffusion LMs, Future Directions