PS200C: Causal Inference for Social Scientists — Spring 2026

Class: MW 2:00–3:15 PM, Bunche 2160 Section: Thursday 9:00–9:50 AM, on Zoom

Professor: Chad Hazlett — chazlett@ucla.edu Office Hours: Wednesday 3:30–4:45 PM and by appointment, Bunche 3264

Teaching Assistant: Barney Chen — barneychen@ucla.edu Office Hours: Mondays 5 PM at Bunche 3244 and by appointment

Course website: https://chadhazlett.github.io/ps200c-2026/ Announcements: BruinLearn — make sure you are receiving notifications.


Why causal inference?

Much of what social scientists study are causal questions. Does foreign aid promote economic growth? Do elections reduce corruption? Does media exposure shift political attitudes? Yet the gap between the causal questions we ask and what the data can tell us is vast. Bridging that gap—carefully, honestly, and with the right tools—is what this course is about.

We begin by establishing a framework for thinking precisely about causal claims: what do we mean when we say X causes Y, and what would we need to know—or assume—to support such a claim? With that foundation in place, we move quickly to the identification strategies that form the backbone of modern empirical social science: randomized experiments, selection on observables, difference-in-differences, instrumental variables, regression discontinuity, and sensitivity analysis. Along the way, we will also cover directed acyclic graphs (DAGs) as a complementary language for expressing and evaluating causal assumptions.

Approach for 2026

This year’s course reflects a deliberate shift in how we think about graduate methods training in an era where AI tools can handle much of the mechanical work—estimation, coding, debugging—that used to dominate students’ time outside of class.

Rather than assigning traditional problem sets, we are investing that time where it matters most: applying causal reasoning to real research problems. Starting early in the quarter, you will develop a final project in which you formulate a causal question, identify an appropriate strategy, and carry out an empirical analysis. The course is structured around milestones that build toward this project, with regular feedback from the instructor, TA, and peers.

In lecture, the emphasis is on building intuition for the core ideas and assumptions behind each identification strategy. Truly understanding the (heroic) assumptions required by these approaches—and what it means to do credible science given them—is the most important aspect of the course. You are expected and encouraged to use AI tools (Claude, ChatGPT, etc.) as tutors to fill in details, clarify concepts, and work through implementation. We want you to learn to think about causal inference; the tools can help you do it.

To keep everyone engaged and to create an honest signal of understanding, most class sessions will end with a short, pen-and-paper in-class activity reviewing the material covered that day. These replace exams as the primary individual assessment and also serve as the attendance mechanism.

Course Requirements and Grading

1. In-class activities (20%)

Most class sessions will include a short (5–10 minute) pen-and-paper activity at the end of lecture. These are designed to reinforce the day’s key ideas and give you immediate feedback on your understanding. They are graded primarily on completion and good-faith effort. Your lowest two scores will be dropped to accommodate absences.

2. Final project milestones (30%)

The final project is developed incrementally through a series of milestones. Each milestone is an opportunity for feedback and course-correction. The milestones are:

3. Final project (50%)

A research paper (roughly 10–15 pages) presenting your causal question, identification strategy, data, analysis, and interpretation. You should be honest about the limitations of your approach and discuss what assumptions are most vulnerable. Final projects are due during finals week. Groups will give brief presentations in the last week of class.

Diagnostic self-assessment

In the first week, we will distribute a short self-diagnostic exercise (similar to the “Problem Set 0” from prior years). This is ungraded. Its purpose is to help you identify gaps in your background—probability, statistics, regression—so you can address them early. The course website’s Resources page will include curated materials keyed to each topic in the diagnostic. Section in the first two weeks will also cover key prerequisites.

Using AI as a learning tool

You are expected to use AI tools throughout this course—not as a shortcut, but as a tutor. When something from lecture is unclear, try asking Claude or ChatGPT to explain it a different way, work through an example, or connect it to something you already know. This is a skill worth developing: learning to ask good questions of an AI is closely related to learning to think clearly about a problem.

To encourage this, each project milestone should include a brief AI learning log: one or two exchanges where you used an AI tool to clarify a concept or work through a problem related to the course, along with a sentence or two about what you learned. This is not assessed for quality—it is meant to normalize the practice and give us insight into what topics are causing confusion.

For the final project, you may use AI tools freely for coding, data wrangling, and exploring ideas. Your paper should include a short section describing how you used AI in the research process. There is no penalty for heavy use—the goal is reflective, honest engagement.

What AI tools cannot do for you is replace the hard thinking at the core of this course: choosing an identification strategy, evaluating whether its assumptions are credible in your setting, and interpreting what the results do and do not tell you. That is your job.

Section

Section will serve two roles this year. First, Barney will use section to review prerequisite and supplementary material that we defer from lecture—this is where you will get more careful treatment of technical details, estimation, and coding. Second, section will include regular project check-ins: time to discuss your project with Barney and with other groups, get feedback, and troubleshoot.

Attendance

Attendance is expected and tracked through the in-class activities. There is no separate attendance grade, but consistent absence will be reflected in your activity scores. If you need to miss class for a legitimate reason, let us know in advance.

Textbooks

The book you should buy is:

Another classic on the topic that is nice to have is:

We will also make reference to “Causal Inference: What If” by Hernán and Robins, freely available at miguelhernan.org/whatifbook.

A great review: If you want to review statistics up through regression from a perspective that is very consistent with how we will use these concepts, check out chapters 1–3 of:

A more comprehensive and annotated set of resources—including readings keyed to specific topics and the self-diagnostic—is available on the Resources page.

Topics

The following is an approximate schedule of topics. They do not correspond exactly to weeks, and the order or selection may shift. Detailed readings for each topic will be posted on the course website.

1. Introduction and the causal inference problem

2. Potential outcomes and the identification problem

3. Randomized experiments

4. Selection on observables

5. DAGs and structural causal models

6. Sensitivity analysis

7. Difference-in-differences

8. Instrumental variables

9. Regression discontinuity

10. Special topics (time permitting)

(i) Panel data without parallel trends.

(ii) Safe inference, partial identification, and new opportunities.

(iii) Mediation (or not) and the TRACE.

Course website

All materials—slides, resources, project guidelines, and section materials—will be posted at:

https://chadhazlett.github.io/ps200c-2026/

Time-sensitive announcements will be posted through the UCLA course site (BruinLearn).