Introduction to Machine Learning
NPFL 054
http://ufal.mff.cuni.cz/course/npfl054
Barbora Hladká hladka@ufal.mff.cuni.cz
Martin Holub holub@ufal.mff.cuni.cz
Charles University,
Faculty of Mathematics and Physics,
Institute of Formal and Applied Linguistics
Lecture #4
Outline
• Linear regression
• Auto data set
Dataset Auto from the ISLR package
392 instances on the following 9 features
mpg Miles per gallon
cylinders Number of cylinders between 4 and 8 displacement Engine displacement (cu. inches) horsepower Engine horsepower
weight Vehicle weight (lbs.)
acceleration Time to accelerate from 0 to 60 mph (sec.) year Model year (modulo 100)
origin Origin of car (1. American, 2. European, 3. Japanese)
name Vehicle name
Dataset Auto from the ISLR package
mpg
3 4 5 6 7 8
●
●
●
●●
●●
●
●●
●●
●●
●
●
●
●
●●
●●
●●
●
●
●●
●
●●
●
●
●●
●●
●
●
●
●
●●
●
●
●
●●
●
●
●
●●
●
●●
●●
●
●●
●●
●●
●
●
●●
●
●
●
●
●●
●
●●
●
●
●
●
●●
●●
●●
●
●●
●●
●●
●
●
●
●
●
●
●
●●
●●
●
●●
●
● ● ●
●
●●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
● ●●●
●
●●
●
●
● ●
●
●
●
●
●●
●●
●
●
●
●
●●
●
●
●
●
●●
●
●●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
● ●
● ●
●
●
●
●
●
●
●
●
●●
●●
●
●●
●
●●
●●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●●
● ●
●●
●
●●
●
●
●● ●
● ●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●●
●●
●●
●●
●
●
●
●
●
●
●
●
●●
●
●
● ●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●●
●
●
●●
●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●● ●●
●
●●
● ●
●
●
●
●
● ●●●●●
●
●
●●
●
●●
●
●
●●
●●
● ●
●
●
●●●
●
●
●●
●
●
●●●
●
●●
●●
●
●●
●●●●
●
●
●● ●
●
● ●●●
●
●●
●
●
●
●
●●
●●
●●●
●● ●
●● ●●
●
●
●
●
●
●
●●●●
●
●●●
●●●
●
● ●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
● ●●
●
●●
● ●
●
● ●
●●
●
●
●
●
●
●
●●
●
● ●●●
●
●●●
●
●●
●
●
●
●
●●
●●
●
●
●
●
●●
●
●
●
●● ●●
●●
● ●
●●
●
●
●
●
●
●
●
●●●
●
●
●●
●
●
●●
● ●
●
●●
●
●
●
●
●
●●●●
●
●●
●
●●●
●
●
● ●●
●
●
● ●
● ●●
●
●
●
●
●
● ●●
●●
●
●
●●●
●
● ●●
●●●
●
●
●
●
●●
● ●
●
●
●
●
●
●
●
●
●
●●
●●●●●●
●●
●
●●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●●
●
●
●●
●
●
●●
●●●
●
●●
●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
15002500 35004500
10 20 30 40
●
●
●
●●
●●
●●●
●●●
●
●
●
●
●
●●●●●●
●
●
●●
●
●●
●
●
●●
●●
● ●
●
●
●●●
●
●
●●
●
●
●
●●
●
●●
●●
●
●●
●●●
●
●
●
●●
●
●
●
●● ●
●
●●
●
●
●
●
●●
●●●●●
●●
●●●●
●
●
●
●●
●
●
●●
●●
●
●●●
●●●
●
●●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
● ●●
●
●
● ●
●●
●
●●
●
●
●
●●
●
● ●●●
●
●●
●
●
●●
●
●
●
●
●●
●●
●
●
●
● ●●
●
●
●
●●●
●
●●
●●
●●
●
●
●
●
●
●
●
●●●
●
●
●●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
● ●●●
●
●●●
●●●●
●
●
●●
●
●
● ●
●
●●
●
●
●
●
●
●●●●
●●
●
●●●
●
●●
●
●● ●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●●
●●●●●●
● ●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●●●
●
●
●●
●
●
●●
● ●●
●
●●● ●●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
3 4 5 6 7 8
●● ●●●●●●●●●●●●●
●
● ●
●
●
●
●●●
●
●
●●
●
●●
●
●
●● ●●
●
●
●
●
●●●
●
●
●
●
●●●●●● ●●●●
●
●●
●●●● ●
● ●●●
●
●
●
●●
●●●●●●●●●
●●
●●●
●●●●●
●●
●
● ●●●
●
●●●●
●
●●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
● ●
●
●●●●
●
●
●●●●
● ●
● ● ●
●
●
●
●
●
●●
●
●
●
● ● ●
●
●
●
●
●
●●●●
●
●
●
●
●
● ●●●●●●●
●
●
●
●
●
● ●●
●
●●●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
● ●
●
●
●
●
● ●●●
●
●
●
●
●
●●●●● ●●
●
●
●
●
●
●●●
●
●●
●●●
●
●
●●●
●
●
●
●
●
●
●
●
●●
● ●●●
●
●
●
●
●
●
●
●
●
●●
●●
●●●
●●●
● ●●
●
●
●
●
●
●
●
● ●
●
●
●
●●●
● ●
●
●
●
●
●
●●●● ●
● ● ●●
●
●● ●
●
●
●
●
●
●
●
●
●
●●●●●●●●●●
● ●●●●●
●
●
●
●
●
●
●
●
●
●●●●
●
● ●●●●●●●●●●●
● ●
●
●
●●
●
●● ●● ●
cylinders
●●●●●●●●●●● ●●●
●
●
●
●
●
● ●●●●
●
●
● ●
●
●●●
●●
●
● ●
● ●
●
●●●●
●
●
●
●
●●
●●
●●
● ●●●
● ●●
● ●
●●
●●● ●●
●
●
● ● ●
●
● ●
● ●●●●●
●
●
●
● ●●●●●●● ●
●
●
●
●●
●
● ●● ●
●
●
●●
●
●
●
●
● ●
● ● ●●
●
●
●
●
●●●
● ●
● ●
● ●●
● ●●
● ●
●
● ●
● ● ●●●●●●
● ●
●
●
●
●●●
●
●●
● ●
●●
●●
●
●●
●●
●
●
●
●
● ● ●
●●●● ●●
● ●
● ●
●●
● ●
●● ●●
●
●●●
●
●●● ●
●
●
● ●
●
●
●●
● ●
●● ●
●
●●●
●
●
●
●
●
●●
●
● ●
● ●
●● ●●
●
●
●
●●●●●
● ●●
●
●
●
●
●
● ●
● ● ●
●
●
●●
●●● ●●●● ●
●
●
●
●
●
●
●
●
●
●●
●
● ●●●●
●●
●
● ●●
●
●
●
●
●
●
●●●
●
●
●
●
● ●●●●●
●
● ●
● ●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
● ●●●
●
●
●●●●●●●●●● ●●●●●●
● ●●●
●
●
●
●
●
●
●
●●●
●
●
●
●●●●●●●●
●
●
●
●
●
●●●
● ●●●
● ●
●
●
●●●●●●●●●
●
●
●
●
●
●
●●●●●
●
●
●●●
●●●
● ●●●●
● ●
●
● ●●●
●
●
●
●
●
●
●
●
●
● ●● ●●●●●
●●
●
●
● ●●●●
●
● ● ●●
●
●●
● ●● ●
●
●
●
● ●●●●●●●● ●●
●●
●●
●
●
●
●
●
●
●
●●●
●
●
●
●
● ●
● ●● ●
●
●
●
●
●
●●
● ●
● ●
●
●
●
●●● ●●
●
● ●
●●●●●● ●●
● ●●
●
●
●●●
●
●
●●
●
●
●
●●
●
●●
● ●
●
●
●
● ●●
●
●●●●●●
●
●
● ●
●●
●●
●●
●
●
●
●●
●
●●●● ●
●
●
● ●
●
●
●
●
●●
●●●
● ●●●
●
● ●●
●
●●●
●● ●●●●●●
●
●
●
●
● ●●
●
●●●
●
●
●
●
●
●●
●●●
●
●
● ●
●● ● ●●●●●
●
●
●
●
●●
●
●
●
●●
●
● ●●●●
● ●
●●
● ●
●
●
●
●
●
●
●● ●
●●
●
●
●●● ● ●●
●
●●●●●
● ●
●
● ●
●
●
● ●
●
●
●
●
● ●●●
●
●
●●
● ●●●● ●● ●●●● ●● ●
●
●
● ●
●
●●
●
●
● ●● ●●
●●
●●●●●●
●
●
●
●●
●
●
●
●●●
●● ●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
● ●
●
●●●
●
●
●
●●
●
●●●
●●●
●
●
●
●
●●
●
●●
●
●
●
●●●
●●
●●
●
●
●
●
●
●●
●
●
● ●●
●
●●
●
●
●
●
●●
●
●
●
●
●●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
● ●●
●●
●
●
●●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●●
●
●
●
●
●
●●
●
● ●
● ●
●
●
●
●
●
●
●●
●
●●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●●●●
●
●
●●
●
●●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●●
●
●●
●
●
●
●●
●
● ●●
●
●●
●●
●
●
●
●
●●
●
●●
●
●
●
●
●●●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●● ●
●
●
●
●
●
●
●
●
●
●
●●
●●●●●●●
●●●●
●
● ●●
●●
●●
●●●●●●●●●
●●
●●●
●
●●●
● ●
●
● ●
●
●
● ●
●
●
● ●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
● ●●
● ●
●
●●
●
●
●
●
●
●
●
●
●● ●●●
●
●
●
●
●●
●
●●
●
●
●
● ●
●
●●
●●
●
●
●
●
●
●●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●●
●
●
●●
●
●
●
●
● ●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●●
●
●
● ●
●
●●
●
●
●
●
● ●
●
● ●
●
●
●
●
●●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
● ●
●
●
●
●●
●
● ●●
●
●●
●●
●
●
●
●
●●
●
● ●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●●
●●
●●
●
●
●●
●
●
●● ●
●●
● ●
●●
●
●
●●
●●
●
●●
●●
●
●
●●
●
●
●
●
● ●
●
●
●●
●
●
●●
●
horsepower
50 100 150 200
●
●
●
●
●
●
●●●
●
●
●
●
●
●●●
●●
●
●●
●
●
●
●
●●
●
●●●● ●●
●
●
●
●
●●
●
●●
●
●
●
● ●
●
●●
●●
●●
●
●
●
●●
●
●
●●
●
●
●●
●
●
●
●●
●
●
●●
●
●●●
●
●
●
●●
●
●
●
●
●
●
●●
●
●●●
●●
●
●
●●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●●
●●●
●
●
●
●
●
●●
●
●●
●●
●
●●
●
●
●
●●●
●●
●
●
● ●
●
●●
●
●
●
●
●●
●
●●●●
●
●
●●
●
●●
●
●
●
●
●●
●●
●●
●
●
●
●
●
●
●●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●●●
●
●
●
●
● ●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
● ●
●
●●
●
●●
●
●●
●
●
●
●●
●
●●●
●
●●●●
●
●
●
●
●●
● ● ●
●
●
●
●
●●●
●
●●
●
●
●●●
● ●●●
● ●
●
●
●
●
●
●
● ●
● ●
● ●●
●
●
●
●●●●
●
●
●●
●●●●●●● ●●●●
●
● ●●
●●
●●
●●
●
● ●●●●●
●●
●●●
●
●●
●
●
●
●
●●
●
●
●●●
●
●●●
10 20 30 40
1500 2500 3500 4500
●
●
●
●●
●
●●
●
●
●●●
●
●
●●
●
●
●
●
●●
●
●
●
●●
●
●●
●
●
●● ●●
●
●
●●
●
●
●
●
●
●
●
●● ●●
●●
●●
●●
●●
●
●●
●
●
●
●●●●
●
●
●
●
●
●
●
●
●
● ●
●
●●
●
●
●●
●
●
●●●
●
●
●
●●
●●●
●
●●
●●
●
●●●
●●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●● ●
●
●
●
●
●
●
●
●
●
●
●●●
●●
●
●
●●
●
●
●●
●
●
●
● ●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●●
●●
●●
●●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●●
●
●●
●●
●●
●
●
●
●
●
●
●● ●
●
●
●●
●
●
● ●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
●●
●●
●●
●●
●
●
●
●
● ●●
●
●
●
●
●
●●
● ●
●●
● ●
●
● ●●
●
●
●
●
●
●
● ●
●
●
●
● ●
●
●
●
●
●
●
●
● ●
●
●●
●●
●
●●
●
●●●●
●
●
●
●●●
●
●
● ●
●
●
●
●
●
●
●
●
●
●●
●●
●●
●●●●●●
●
● ●
● ●
●
●
●
●
●
●
● ●
●●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●●
●
●●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●
●●
●●
●●
●●
●●
●●
●
●●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●●
●
●●
●
●
●●
●
●
●●
●
●
●
●
●●
●●
●
●
●●
●●
●
●●
●
● ● ●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●●
●
●
●●
●
●
●●
●
●
●
●●
● ●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●●
●●
●●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●●
●
●●
●●
●●
●
●
●
●
●
●
●●
●
●
●
● ●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
● ●
●●
●●
●●
●
●
●
●
●●
●
●
●
●
●
●
●●
●●
● ●●
●
●
●●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●●
● ●
●
●●
●
●●
●●
●
●
●
●
●●
●
●
● ●
●
●
●
●
●
●
●
●
●
●●
●●
●●
●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●
●●
●●
50 100 150 200
●
●
●
●
●
●●●●
●
●●
●
●
●
●●
●
●
●
●
●●
●
●
●
● ●
●
●●●
●
●●
● ●
●
●
●●
●
●
●
●
●
●
●
●●
●●
●●
●●
●●
● ●
●
●●
●●
●
●● ●●
●
●
●●
●
●
●
●
●●●
●
●●
●
●
●●
●
●
●●
●
●●
●
●●
●●●
●
● ●
●●
●
●●●
●
●●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●●●●
●●
●
●
● ●
●
●
●●●
●
●
●●
●
●●
●
●
●
●●
●
●
●
●
●
●
● ●
●
●
●
●
●
●●
● ●
●
●
●●
● ●
●●
●●
●●●
●
●
●●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●●●
●●
●●
●●●
●
●
●
●
●
●●●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●●
●
●
●●●
●
●
●
●
●●
●
●●
●
●
●
●
●
●
●
●
●●
●●
●●
●●
●
●
●
●
● ●●
●
●
●
●
●
●●
●●
●●●
●
●
●●●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●●●●
●
●●●
●●
●●
●
●
●
●
●●
●
●
●●
●●
●
●
●
●
●
●
●
●●
●●
●●
●
●●●●●
●
●
●
● ●
●
●
●
●
●
●
● ●
●●
weight
Linear regression
Linear regression
Linear regression is a class of regression algorithms assuming that there is at least a linear dependence between a target attribute and features.
A target hypothesis f has a form of linear function
f (x; Θ) = θ 0 + θ 1 x 1 + · · · + θ m x m (1)
– θ 0 , . . . , θ m are regression parameters
– simple linear regression if m = 1
Linear regression
Notation
y =
y 1 . . . y n
x i = h1, x i1 , . . . , x im i
Θ > =
θ 0 . . . θ m
, X =
1 x 11 . . . x 1m
1 x 21 . . . x 2m . . . . . . . . . . . . 1 x n1 . . . x nm
Now we can write y = XΘ > , f (x) = Θ > x
Parameter interpretation
Numerical feature
θ i is the average change in y for a unit change in A i holding all other features fixed
Parameter interpretation
Categorical feature with k values
Replace the feature with k − 1 dummy numerical features DA 1 , . . . , DA k−1 Example: run simple linear regression mpg ∼ origin
DA 1 DA 2
American 0 0
European 1 0
Japanase 0 1
• y = θ 0 + θ 1 DA 1 + θ 2 DA 2
• y = θ 0 + θ 1 if the car is European
• y = θ 0 + θ 2 if the car is Japanese
• y = θ 0 if the car is American
• θ 0 as the average mpg for American cars
• θ 1 as the average difference in mpg between European and American cars
• θ 2 as the average difference in mpg between Japanese and American cars
Parameter estimates Least Square Method
• residual y i − y ˆ i , where ˆ y i = ˆ f (x i ) = ˆ Θ > x i
• Loss function Residual Sum of Squares RSS( ˆ Θ) = P n
i=1 (y i − y ˆ i ) 2
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
● ●
●
●
●
● ●
●
● ●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●
● ●
●●
●
●
●●
●●●●
●
●
● ●
●
●
● ●
● ●●
●
●
●
●
●
● ●●
●
● ●
●
●
●
● ●
● ●
●
● ●
●
●
● ●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
● ●
●
●
●●
●●●
●
●●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
1500 2000 2500 3000 3500 4000 4500 5000
10 20 30 40
ISLR: Auto data set
Weight
Miles P er Gallon
Residuum ve 3D
1000 2000 3000 4000 5000 6000
01020304050
0 50
100 150
200 250
weight
horse po w er
miles per gallon
●
●
● ●●
●
●
●
● ●
●
●
●
●
●
●
● ● ●
●●●●● ●
●
●●●
●●
●
●
●
●●
●
●
●
●
●
●●●●
●
● ●
●●
● ●●
●
●
●●
●●●●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●●
●
●
●
●
●
●
●
●
●●●
●●
●
●●
●
●●
●
●
●
●
●
●
●
●
●
● ●
●
●
●●
●
●
●
●●
●
●
●
●
● ●●
●
●●●
● ●
●
●
●
●
●
●
●●
● ●
●
●
●
●
●
●
●
●
● ●
●
● ●
● ●
●
●
●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●
● ●
●
●●
●
●
●●
●
●
●
●
●●
●
●
●
●
●
●●
●
●
●
●
●●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
●
●●
●
●
●
●
●
● ●
● ●
●
●
●
●
●
●
●
●
●
●
●
●
● ●
●
●
●
●
●
●
●
●●
●
●
●●
●
●
●
●●
●
●
●●
●
● ●
●●
●
●●
●
●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
●
●●
●
●
●
●●
●
●
●
●
●
●
●
●
●
●
●
●●
●●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●
●●
●
●
●
Parameter estimates Least Square Method
Optimization problem
Θ ? = argmin Θ RSS(Θ)
The argmin operator will give Θ for which RSS(Θ) is minimal.
Parameter estimates Least Square Method
Solving the optimization problem analytically Normal Equations Calculus
Theorem
Θ ? is a least square solution to y = XΘ > ⇔ Θ ? is a solution to the Normal equation X > XΘ = X > y.
Θ ? = (X > X) −1 X > y
Computational complexity of a (m + 1) × (m + 1) matrix inversion is O(m + 1) 3 :-(
Parameter estimates Least Square Method
Solving the optimization problem numerically
Gradient Descent Algorithm
Gradient Descent Algorithm
Assume: simple regression, θ 0 = 0, θ 1 6= 0
Gradient Descent Algorithm
Assume: simple regression, θ 0 6= 0, θ 1 6= 0
theta0 30
40 50
60
theta1
−0.010
−0.005 0.000 L(theta0, theta1)
2e+05 4e+05 6e+05
Loss Function L has a minimum value at the red point
●
●
●
Contours of Loss Function
theta0
theta1
20000
40000 40000
60000 60000
80000
80000 1e+05
1e+05
120000 120000
140000 140000
160000
160000
180000 180000
2e+05
2e+05
220000 220000
240000
240000
260000
260000
280000
280000
3e+05
3e+05
320000
340000
340000
360000
360000
380000
380000
4e+05
4e+05
420000
440000
480000
5e+05
5e+05 560000 6e+05
30 40 50 60