Skip to the content.

Prospective Students

Course Staff

Instructor (Guest) Co-instructor
Dawn Song Xinyun Chen
Professor, UC Berkeley Research Scientist, Google DeepMind

GSIs: Alex Pan & Sehoon Kim

Readers: Tara Pande & Ashwin Dara

Class Time and Location

Lecture: 3-5pm PT Monday at Latimer 120

Course Description

Large language models (LLMs) have revolutionized a wide range of domains. In particular, LLMs have been developed as agents to interact with the world and handle various tasks. With the continuous advancement of LLM techniques, LLM agents are set to be the upcoming breakthrough in AI, and they are going to transform the future of our daily life with the support of intelligent task automation and personalization. In this course, we will first discuss fundamental concepts that are essential for LLM agents, including the foundation of LLMs, essential LLM abilities required for task automation, as well as infrastructures for agent development. We will also cover representative agent applications, including code generation, robotics, web automation, medical applications, and scientific discovery. Meanwhile, we will discuss limitations and potential risks of current LLM agents, and share insights into directions for further improvement. Specifically, this course will include the following topics:

Syllabus

Date Guest Lecture Readings
(due Sunday 11:59pm before lecture on Gradescope)
Sept 9 LLM Reasoning
Denny Zhou, Google DeepMind
Intro Slides Original Recording Edited Video
- Chain-of-Thought Reasoning Without Prompting
- Large Language Models Cannot Self-Correct Reasoning Yet
- Premise Order Matters in Reasoning with Large Language Models
- Chain-of-Thought Empowers Transformers to Solve Inherently Serial Problems
All readings are optional this week.
Sept 16 LLM agents: brief history and overview
Shunyu Yao, OpenAI
Slides Original Recording Edited Video
- WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents
- ReAct: Synergizing Reasoning and Acting in Language Models
Sept 23 Agentic AI Frameworks & AutoGen
Chi Wang, AutoGen-AI
Building a Multimodal Knowledge Assistant
Jerry Liu, LlamaIndex
Chi’s Slides Jerry’s Slides Original Recording Edited Video
- AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation
- StateFlow: Enhancing LLM Task-Solving through State-Driven Workflows
Sept 30 Enterprise trends for generative AI, and key components of building successful agents/applications
Burak Gokturk, Google
Slides Original Recording Edited Video
- Google Cloud expands grounding capabilities on Vertex AI
- The Needle In a Haystack Test: Evaluating the performance of RAG systems
- The AI detective: The Needle in a Haystack test and how Gemini 1.5 Pro solves it
Oct 7 Compound AI Systems & the DSPy Framework
Omar Khattab, Databricks
Slides Original Recording Edited Video
- Optimizing Instructions and Demonstrations for Multi-Stage Language Model Programs
- Fine-Tuning and Prompt Optimization: Two Great Steps that Work Better Together
Oct 14 Agents for Software Development
Graham Neubig, Carnegie Mellon University
Slides Original Recording Edited Video
- SWE-agent: Agent-Computer Interfaces Enable Automated Software Engineering
- OpenHands: An Open Platform for AI Software Developers as Generalist Agents
Oct 21 AI Agents for Enterprise Workflows
Nicolas Chapados, ServiceNow
Slides Original Recording Edited Video
- WorkArena: How Capable Are Web Agents at Solving Common Knowledge Work Tasks?
- WorkArena++: Towards Compositional Planning and Reasoning-based Common Knowledge Work Tasks
- TapeAgents: a Holistic Framework for Agent Development and Optimization
Oct 28 Towards a unified framework of Neural and Symbolic Decision Making
Yuandong Tian, Meta AI (FAIR)
Slides Original Recording Edited Video
- Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping
- Dualformer: Controllable Fast and Slow Thinking by Learning with Randomized Reasoning Traces
- Composing Global Optimizers to Reasoning Tasks via Algebraic Objects in Neural Nets
- SurCo: Learning Linear Surrogates For Combinatorial Nonlinear Optimization Problems
Nov 4 Project GR00T: A Blueprint for Generalist Robotics
Jim Fan, NVIDIA
Slides Original Recording Edited Video
- Voyager: An Open-Ended Embodied Agent with Large Language Models
- Eureka: Human-Level Reward Design via Coding Large Language Models
- DrEureka: Language Model Guided Sim-To-Real Transfer
Nov 11 No Class - Veterans Day  
Nov 18 Open-Source and Science in the Era of Foundation Models
Percy Liang, Stanford University
Slides Original Recording Edited Video
- Cybench: A Framework for Evaluating Cybersecurity Capabilities and Risks of Language Models
Nov 25 Measuring Agent capabilities and Anthropic’s RSP
Ben Mann, Anthropic
Slides Original Recording Edited Video
- Announcing our updated Responsible Scaling Policy
- Developing a computer use model
Dec 2 Towards Building Safe & Trustworthy AI Agents and A Path for Science‑ and Evidence‑based AI Policy
Dawn Song, UC Berkeley
Slides Edited Video
- A Path for Science‑ and Evidence‑based AI Policy
- DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models
- Representation Engineering: A Top-Down Approach to AI Transparency
- Extracting Training Data from Large Language Models
- The Secret Sharer: Evaluating and Testing Unintended Memorization in Neural Networks
All readings are optional this week.

Enrollment and Grading

Please fill out the signup form if you are on the waitlist or can’t get added to the waitlist.

This is a variable-unit course. All enrolled students are expected to participate in lectures in person and complete weekly reading summaries related to the course content. Students enrolling in one unit are expected to submit an article that summarizes one of the lectures. Students enrolling in more than one unit are expected to submit a lab assignment and a project instead of the article. The project of students enrolling in 2 units should have a written report, which can be a survey in a certain area related to LLMs. The project of students enrolling in 3 units should also have an implementation (coding) component that programmatically interacts with LLMs, and the project of students enrolling in 4 units should have a very significant implementation component with the potential for either real world impacts or intellectual contributions. The grade breakdowns for students enrolled in different units are the following:

  1 unit 2 units 3/4 units
Participation 40% 16% 8%
Reading Summaries & Q/A 10% 4% 2%
Quizzes 10% 4% 2%
Article 40%    
Lab   16% 8%
Project      
  Proposal   10% 10%
  Milestone 1   10% 10%
  Milestone 2   10% 10%
  Presentation   15% 15%
  Report   15% 15%
  Implementation     20%

Lab and Project Timeline

  Released Due
Project group formation 9/9 9/16
Project proposal 9/22 9/30
Labs 10/1 10/15
Project milestone #1 10/19 10/25
Project milestone #2 10/29 11/20
Project final presentation 11/19 12/17
Project final report 11/19 12/17

Office Hours