Data 102: Data, Inference, and Decisions
UC Berkeley
This course develops the probabilistic foundations of inference in data science, and builds a comprehensive view of the modeling and decision-making life cycle in data science including its human, social, and ethical implications. Topics include: frequentist and Bayesian decision-making, permutation testing, false discovery rate, probabilistic interpretations of models, Bayesian hierarchical models, basics of experimental design, confidence intervals, causal inference, robustness, Thompson sampling, optimal control, Q-learning, differential privacy, fairness in classification, recommendation systems and an introduction to machine learning tools including decision trees, neural networks and ensemble methods.
This class is listed as DATA C102.
Offerings
- Fall 2024 (Ramesh Sridharan, Alexander Strang)
- Spring 2024 (Ramesh Sridharan, Alexander Strang)
- Fall 2023 (Ramesh Sridharan, Aditya Guntuboyina)
- Spring 2023 (Ramesh Sridharan, Eaman Jahani)
- Fall 2022 (Ramesh Sridharan, Jacob Steinhardt)
- Spring 2022 (Ramesh Sridharan, Nika Haghtalab)
- Fall 2021 (Ramesh Sridharan, Jacob Steinhardt)
- Spring 2021 (Ramesh Sridharan, Yan Shuo Tan)
- Fall 2020 (Michael Jordan, Jacob Steinhardt)
- Spring 2020 (Jacob Steinhardt, Moritz Hardt)
- Fall 2019 (Michael Jordan, Fernando Perez)
Prerequisites
While we are working to make this class widely accessible, we currently require the following (or equivalent) prerequisites :
-
Principles and Techniques of Data Science: Data 100 covers important computational and statistical skills that will be necessary for Data 102.
-
Probability: Data 140, EECS 126, STAT 134, IEOR 172, or Math 106. Data 140 and EECS 126 are preferred. These courses cover the probabilistic tools that will form the underpinning for the concepts covered in Data 102.
-
Math: Math 54, Math 56, Math 110, both EE 16A and EE 16B, STAT 89a, or Physics 89. We will need some basic concepts like linear operators, eigenvectors, derivatives, and integrals to enable statistical inference and derive new prediction algorithms.