Kilian Weinberger is an Associate Professor in the Department of Computer Science at Cornell University. He received his Ph.D. from the University of Pennsylvania in Machine Learning under the supervision of Lawrence Saul and his undergraduate degree in Mathematics and Computer Science from the University of Oxford. During his career he has won several best paper awards at ICML (2004), CVPR (2004, 2017), AISTATS (2005) and KDD (2014, runner-up award). In 2011 he was awarded the Outstanding AAAI Senior Program Chair Award and in 2012 he received an NSF CAREER award. He was elected co-Program Chair for ICML 2016 and for AAAI 2018. In 2016 he was the recipient of the Daniel M Lazar ’29 Excellence in Teaching Award. Kilian Weinberger’s research focuses on Machine Learning and its applications. In particular, he focuses on learning under resource constraints, metric learning, machine learned web-search ranking, computer vision and deep learning. Before joining Cornell University, he was an Associate Professor at Washington University in St. Louis and before that he worked as a research scientist at Yahoo! Research in Santa Clara.
In this course, you will investigate the underlying mechanics of a machine learning algorithm’s prediction accuracy by exploring the bias variance trade-off. You will identify the causes of prediction error by recognizing high bias and variance while learning techniques to reduce the negative impacts these errors have on learning models. Working with ensemble methods, you will implement techniques that improve the results of your predictive models, creating more reliable and efficient algorithms.
These courses are required to be completed prior to starting this course:
- Problem-Solving with Machine Learning
- Estimating Probability Distributions
- Learning with Linear Classifiers
- Decision Trees and Model Selection
- Identify the cause of high prediction error by recognizing high bias or high variance
- Mitigate the negative impact of a bad bias/variance trade-off on your model
- Analyze how ensemble methods reduce bias and variance in order to improve the predictive model
- Implement bagging and boosting to improve the predictive model
How It Works
Who Should Enroll
- Data analysts
- Data scientists
- Software engineers