Welcome to Economics 607: Econometrics III (Spring 2024) at the University of Oregon (taught by Dr. Ed Rubin).
/\\\\\\\\\\\\\\\ /\\\\\\\\\ /\\\\\ /\\\\\\\ /\\\\\\\\\\\\\\\
\/\\\/////////// /\\\//////// /\\\\//// /\\\/////\\\ \/////////////\\\
\/\\\ /\\\/ /\\\/// /\\\ \//\\\ /\\\/
\/\\\\\\\\\\\ /\\\ /\\\\\\\\\\\ \/\\\ \/\\\ /\\\/
\/\\\/////// \/\\\ /\\\\///////\\\ \/\\\ \/\\\ /\\\/
\/\\\ \//\\\ \/\\\ \//\\\ \/\\\ \/\\\ /\\\/
\/\\\ \///\\\ \//\\\ /\\\ \//\\\ /\\\ /\\\/
\/\\\\\\\\\\\\\\\ \////\\\\\\\\\ \///\\\\\\\\\/ \///\\\\\\\/ /\\\/
\/////////////// \///////// \///////// \/////// \///
Lecture Monday and Wednesday 10:00am–11:20am, Friendly 221
Lab Friday 12:00pm–12:50pm, 330 Condon
Office hours
- Ed Rubin: Mondays 4pm–5pm and by appointment, PLC 530
- Owen Jetton Tuedays and Wednesdays 12pm–1pm, 417 PLC
We will mainly use two books.
Mostly Harmless Econometrics: An Empiricist's Companion (MHE)
by Angrist and Pischke
Your new best friend. Read it.
Microeconometrics (C&T)
by Cameron and Trivedi
Also very readable and accessible.
Econometric Analysis (Greene)
by Greene
The standard—an encyclopedic resource for many of the questions MHE does not answer.
Introduction to Causal Inference (Neal)
by Brady Neal
A free, under-development, causal-inference book targeting folks who come from a prediction (think: machine learning) background.
Causal Inference in Statistics: A Primer (Pearl)
by Pearl, Glymour, and Jewell
Causal Inference: The Mixtape (Mixtape)
by Cunningham
Note: The linked slides (below) are .html
files that will only work properly if you are connected to the internet. If you're going off grid (camping + metrics?), grab the PDFs. You'll miss out on gifs and interactive plots, but the equations will actually show up.
The content of the lectures mainly follows MHE and Michael Anderson—with additional inspiration from Max Auffhammer and many other sources.
Another note on the notes: I create the slides with xaringan
in R. Thanks to Grant McDermott for encouraging me to make this switch.
Lecture 01: Research + R + You = 💖
- An introduction to empirical research via applied econometrics.
- R: Light introduction—objects, functions, and help.
Note formats: .html | .pdf | .rmd
Readings: MHE preface + MHE chapter 1
Lecture 02: The Experimental Ideal
- Neyman potential outcomes framework (Rubin causal model)
- Selection bias and experimental variation in treatment
Note formats: .html | .pdf | .rmd
Readings: MHE chapter 2
- What's the big deal about least-squares (population) regression?
- What does the CEF tell us?
- How does least-squares regression relate to the CEF?
Note formats: .html | .pdf | .rmd
Readings: MHE chapter 3.1
Lecture 04: Inference and Simulation
- How do we move from populations to samples?
- What matters for drawing basic statistical inferences about the population?
- How can we learn about inference from simulation?
- How do we run (parallelized) simulations in R?
Note formats: .html | .pdf | .rmd
Readings: MHE chapter 3
- Saturated models
- When is regression causal?
- The conditional-independence assumption
Note formats: .html | .pdf | .rmd
Readings: Still MHE chapter 3
- Omitted-variable bias
- Good and bad controls
Note formats: .html | .pdf | .rmd
Readings: Still MHE chapter 3
- Defining graphs
- Underlying theory for DAGs
- Building blocks
- Examples
Note formats: .html |
.pdf |
.rmd
Readings: Brady Neal's book, chapters 1–3 (especially 2–3)
- Matching estimators: Nearest neighbor and kernel
- Propensity-score methods: Regression control, treatment-effect heterogeneity, blocking, weighting, doubly robust
Note formats: .html |
.pdf |
.rmd
Readings: MHE chapter 3 + C&T section 25.4
Bonus: Slides outlining logistic regression.
Lecture 09: Instrumental Variables
- General research designs
- Instrumental variables (IV)
- Two-stage least squares (2SLS)
- Heterogeneous treatment effects and the LATE
Note formats: .html |
.pdf |
.rmd
Readings: MHE chapter 4 + C&T sections 4.8–4.9
Additional material: Paper on machine learning the first stage of 2SLS
Lecture 10: Regression Discontinuity
- Sharp regression discontinuities
- Fuzzy regression discontinuities
- Graphical analyses
Note formats: .html |
.pdf |
.Rmd
Readings: MHE chapter 6 + C&T sections 25.6
Lecture 11: Inference: Clustering
- General inference
- Moulton
- Cluster-robust standard errors
Note formats: .html | .pdf | .Rmd
Readings: MHE chapter 8
Lecture 12: Inference: Resampling and Randomization
- Resampling
- The bootstrap
- Permutation tests (Fisher)
- Randomization inference (Neyman-Pearson)
Note formats: .html | .pdf | .Rmd
Readings: MHE chapter 6 + C&T sections 25.6
Lecture 13: Machine learning (in one lecture)
- Prediction basics
- The bias-variance tradeoff
- In-sample vs. out-of-sample performance
- Hold-out methods (including cross validation)
- Ridge regression and lasso
- Decision trees
- Ensembles (of trees)
Note formats: .html | .pdf | .Rmd
Readings: Introduction to statistical learning
Owen Jetton will walk you through R and applications of the course content. You should attend.
Note: From previous iteration of our class.
- Object types/classes/structures
- Package management
- Math and stat. in R
- Indexing
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
Solutions:
.html | .pdf
- Data frames
- Data work with
dplyr
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
Lab 03: RStudio + Data i/o with R
- RStudio
- Getting data into and out of R
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
lm()
andlm
objectsestimatr
andlm_robust()
- Other regressions, e.g.,
glm()
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
- Default
plot()
methods ggplot2
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
- General simulation strategies
- Simulating IV in finite samples
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
Lab 07: Miscellaneous R Tips and Tricks
- The
apply
family for()
loops- Lists
- Logical vectors and
which()
Note formats: .html | .html (no pause) | .pdf | .pdf (no pause) | .Rmd
Problem sets combining econometric theory and R.
Problem set 1
Due Thursday, 18 April 2024
Problem set 2
Due Sunday, 12 May 2024
Problem set 3
Due Friday, 24 May 2024
The course has two projects:
- A research proposal that centers on a causal question.
- A presentation of a topic that extends what we cover during the course.
Building a research project/proposal.
Why? You are wrapping up your first year in the PhD. It's time to start thinking about how you could apply what you've learned.
Assignment: Pitch a project that includes a causal question of interest. Include motivation.
- This project should be something you could turn into a legitimate research project.
- Length: 150–250 words
- You should have several drafts (only submit the last one).
- Talk with your classmates (and me!).
More information here
Due: May 1, 2024; submit on Canvas
Assignment: Incorporate feedback from step 1 and write a "full" project proposal (~3 pages).
- Motivate and outline the causal question of interest.
- Explain potential sources of selection that could bias estimation.
- Describe the ideal experiment for your setting.
- Discuss a practical research design through which one could answer the question. Explain how this research design avoids selection bias.
Note: You do not need to actually estimate anything.
More information here.
Due: May 29, 2024
Assignment
- Choose a topic related to causal inference–that we do not cover in class (e.g., difference-in-differences, the Wild Clustered Bootstrap, synthetic control methods).
- Write a summary/tutorial of the topic that includes (a) the math behind the approach and (b) an empirical example.
- Present a five-minute summary of the topics to your classmates.
Why? In the course of the PhD, we want to teach you how to learn. We will not provide you with everything you need to know to be able to do research on any topic. But hopefully we provide you with a foundation and the ability to learn new things. Also: You need to learn how to communicate both in writing and in person.
Due: All material (including slides) is due June 2, 2024. Presentations will be during class on June 5, 2024.
See here for more information and example topics.
- Inference and simulation
- Matching
- Instrumental variables
- Regression discontinuity
- Inference: Clustering and resampling
The final exam has two parts:
- In class: 10:15am–12:15pm on Tuesday, June 11th (2024).
- Take-home exam: Responses due by 11:59pm Pacific on Thursday, June 13th, 2023.
We do not have a midterm exam.
Examples of past exams:
As you've hopefully figured out by now, our PhD program is not "about grades." This class is critical to building the intuition and skills that you will rely upon in your own empirical work and in communicating with others about their empirical work. Commit to (and focus on) learning the material—the theory, the intuition, and the programming.
That said, I do have to turn in grades (and there is a GPA requirement to sit for the qualifying exam). I will weight your grades as follows:
- Exams: The exam is worth 45% of your course grade.
- Project: Each of the projects is worth 12.5% of your course grade (so the projects together are worth 25% of your course grade).
- Assignments: Assignments jointly cover the remaining 30% of the grade (and may not be weighted equally).
Note: Anything you turn in with your name on it should be legitimately your own work. I encourage you to work with classmates and to get good with ChatGPT/Copilot/Google, but you still need to put things in your own words and understand what you've submitted. Submitting other people's work as your own will result in you failing this course.
Metrics books
- Hayashi's Econometrics
- Kennedy
- Mastering 'Metrics (undergrad version of Mostly Harmless)
- Stock and Waston
- Wooldridge ("Baby")
- Wooldridge (Adult?)
R resources
- RStudio: Finding Your Way To R
- My EC524 Economics & Machine Learning Course at UO
- Grant McDermott's Data Science of Economists course
- DataCamp's Introduction to R
- R for Data Science
- Advanced R
Metrics and R
- Website from 2023
- Website from 2022
- Website from 2021
- Website from 2020
- Website from 2019
- Section notes from a previous PhD-level class.
- Lecture notes from a previous undergrad class.
More
- David Card's Nobel acceptance speech, the related paper, and some slides