Upload
hailos
View
30
Download
8
Embed Size (px)
DESCRIPTION
Slides from the book Learning From Data
Citation preview
Outline of the Course
1. The Learning Problem (April 3 )
2. Is Learning Feasible? (April 5 )
3. The Linear Model I (April 10 )
4. Error and Noise (April 12 )
5. Training versus Testing (April 17 )
6. Theory of Generalization (April 19 )
7. The VC Dimension (April 24 )
8. Bias-Variane Tradeo (April 26 )
9. The Linear Model II (May 1 )
10. Neural Networks (May 3 )
11. Overtting (May 8 )
12. Regularization (May 10 )
13. Validation (May 15 )
14. Support Vetor Mahines (May 17 )
15. Kernel Methods (May 22 )
16. Radial Basis Funtions (May 24 )
17. Three Learning Priniples (May 29 )
18. Epilogue (May 31 )
theory; mathematial
tehnique; pratial
analysis; oneptual
Learning From Data
Yaser S. Abu-Mostafa
California Institute of Tehnology
Leture 1: The Learning Problem
Sponsored by Calteh's Provost Oe, E&AS Division, and IST Tuesday, April 3, 2012
The learning problem - Outline
Example of mahine learning
Components of Learning
A simple model
Types of learning
Puzzle
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 2/19
Example: Prediting how a viewer will rate a movie
10% improvement = 1 million dollar prize
The essene of mahine learning:
A pattern exists.
We annot pin it down mathematially.
We have data on it.
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 3/19
Movie rating - a solution
Math movie and
viewer fators
predited
rating
o
m
e
d
y
o
n
t
e
n
t
a
t
i
o
n
o
n
t
e
n
t
b
l
o
k
b
u
s
t
e
r
?
T
o
m
C
r
u
i
s
e
i
n
i
t
?
l
i
k
e
s
T
o
m
C
r
u
i
s
e
?
p
r
e
f
e
r
s
b
l
o
k
b
u
s
t
e
r
s
?
l
i
k
e
s
a
t
i
o
n
?
l
i
k
e
s
o
m
e
d
y
?
movie
viewer
add ontributions
from eah fator
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 4/19
The learning approah
v i e w e r
m o v i e
top
bottomrating
LEARNING
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 5/19
Components of learning
Metaphor: Credit approval
Appliant information:
age 23 years
gender male
annual salary $30,000
years in residene 1 year
years in job 1 year
urrent debt $15,000
Approve redit?
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 6/19
Components of learning
Formalization:
Input: x (ustomer appliation)
Output: y (good/bad ustomer? )
Target funtion: f : X Y (ideal redit approval formula)
Data: (x1, y1), (x2, y2), , (xN , yN) (historial reords)
Hypothesis: g : X Y (formula to be used)
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 7/19
f:
(ideal credit approval function)
(historical records of credit customers)
HYPOTHESIS SET
(set of candidate formulas)
ALGORITHMLEARNING FINAL
HYPOTHESIS
UNKNOWN TARGET FUNCTION
(final credit approval formula)
TRAINING EXAMPLES
X Y
x y x yNN11( , ), ... , ( , )
H
A
g ~ f ~
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 8/19
Solution omponents
f:
(ideal credit approval function)
(historical records of credit customers)
HYPOTHESIS SET
(set of candidate formulas)
ALGORITHMLEARNING FINAL
HYPOTHESIS
UNKNOWN TARGET FUNCTION
(final credit approval formula)
TRAINING EXAMPLES
X Y
x y x yNN11( , ), ... , ( , )
H
A
g ~ f ~
The 2 solution omponents of the learning
problem:
The Hypothesis Set
H = {h} g H
The Learning Algorithm
Together, they are referred to as the learning
model .
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 9/19
A simple hypothesis set - the `pereptron'
For input x = (x1, , xd) `attributes of a ustomer'
Approve redit if
di=1
wixi > threshold,
Deny redit if
di=1
wixi < threshold.
This linear formula h H an be written as
h(x) = sign
((d
i=1
wixi
) threshold
)
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 10/19
+++
+_
_
_
_
+
+
++
+
_
_
_
_
+
`linearly separable' data
h(x) = sign
((d
i=1
wi xi
)+ w0
)
Introdue an artiial oordinate x0 = 1:
h(x) = sign
(d
i=0
wi xi
)
In vetor form, the pereptron implements
h(x) = sign(wTx)
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 11/19
A simple learning algorithm - PLA
The pereptron implements
h(x) = sign(wTx)
Given the training set:
(x1, y1), (x2, y2), , (xN , yN)
pik a mislassied point:
sign(wTxn) 6= yn
and update the weight vetor:
w w + ynxn
w+ xy
yw+ x
y= +1
xw
x
w1y=
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 12/19
Iterations of PLA
+
++
++_
_
_
_
One iteration of the PLA:
w w + yx
where (x, y) is a mislassied training point.
At iteration t = 1, 2, 3, , pik a mislassied point from
(x1, y1), (x2, y2), , (xN , yN)
and run a PLA iteration on it.
That's it!
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 13/19
The learning problem - Outline
Example of mahine learning
Components of learning
A simple model
Types of learning
Puzzle
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 14/19
Basi premise of learning
using a set of observations to unover an underlying proess
broad premise = many variations
Supervised Learning
Unsupervised Learning
Reinforement Learning
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 15/19
Supervised learning
Example from vending mahines oin reognition
25
51
Mas
s
Size
10
25
51
Mas
s
Size
10
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 16/19
Unsupervised learning
Instead of (input,orret output), we get (input, ? )
Mas
s
Size
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 17/19
Reinforement learning
Instead of (input,orret output),
we get (input,some output,grade for this output)
The world hampion was
a neural network!
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 18/19
A Learning puzzle
f = 1
f = +1
f = ?
AML Creator: Yaser Abu-Mostafa - LFD Leture 1 19/19