CS 4100: Artificial Intelligence Perceptrons and Logistic Regression Jan-Willem van de Meent, Northeastern University [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Linear Classifiers Feature Vectors Hello, Do you want free printr cartriges? Why pay more when you can get them ABSOLUTELY FREE! Just # free : 2 YOUR_NAME : 0 MISSPELLED : 2 FROM_FRIEND : 0 ... SPAM or + PIXEL-7,12 : 1 PIXEL-7,13 : 0 ... NUM_LOOPS : 1 ... “2” Some (Simplified) Biology • Very loose inspiration: human neurons Linear Classifiers • Inputs are feature values • Each feature has a weight • Sum is the activation • If the activation is: • Positive, output +1 • Negative, output -1 S f1 f2 f3 w1 w2 w3 >0? Weights • Binary case: compare features to a weight vector • Learning: figure out the weight vector from examples # free : 2 YOUR_NAME : 0 MISSPELLED : 2 FROM_FRIEND : 0 ... # free : 4 YOUR_NAME :-1 MISSPELLED : 1 FROM_FRIEND :-3 ... # free : 0 YOUR_NAME : 1 MISSPELLED : 1 FROM_FRIEND : 1 ... Dot product positive means the positive class Decision Rules Binary Decision Rule • In the space of feature vectors • Examples are points • Any weight vector is a hyperplane • One side corresponds to Y=+1 • Other corresponds to Y=-1 BIAS : -3 free : 4 money : 2 ... 0 1 0 1 2 free money +1 = SPAM -1 = HAM
4
Embed
Binary Decision RuleDecision Rules€¦ · Learning: Binary Perceptron 7 Start with weights = 0 7 For each training instance: 7Classify with current weights 7If correct: (i.e., y=y*
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
CS 4100: Artificial IntelligencePerceptrons and Logistic Regression
Jan-Willem van de Meent, Northeastern University
[These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.]
Linear Classifiers
Feature Vectors
Hello,
Do you want free printrcartriges? Why pay more when you can get them ABSOLUTELY FREE! Just
• In the space of feature vectors• Examples are points• Any weight vector is a hyperplane• One side corresponds to Y=+1• Other corresponds to Y=-1
BIAS : -3free : 4money : 2... 0 1
0
1
2
free
mon
ey
+1 = SPAM
-1 = HAM
Weight Updates Learning: Binary Perceptron• Start with weights = 0• For each training instance:
• Classify with current weights
• If correct: (i.e., y=y*), no change!
• If wrong: adjust the weight vector
Learning: Binary Perceptron• Start with weights = 0• For each training instance:
• Classify with current weights
• If correct: (i.e., y=y*), no change!• If wrong: adjust the weight vector
by adding or subtracting the feature vector. Subtract if y* is -1.
Examples: Perceptron
• Separable Case
Multiclass Decision Rule
• If we have multiple classes:• A weight vector for each class:
• Score (activation) of a class y:
• Prediction with highest score wins
Binary = multiclass where the negative class has weight zero
Learning: Multiclass Perceptron
• Start with all weights = 0• Pick training examples one by one
• Predict with current weights
• If correct: no change!• If wrong: lower score of wrong answer,
raise score of right answer
Example: Multiclass Perceptron
BIAS : 1win : 0game : 0 vote : 0 the : 0 ...
BIAS : 0win : 0game : 0vote : 0the : 0...
BIAS : 0 win : 0 game : 0 vote : 0 the : 0 ...
Question: What will the weights w be for each class after 3 updates?
y1 = “politics”, x1 = “win the vote”y2 = “politics”, x2 = “win the election”y3 = “sports”, x3 = “win the game”
Example: Multiclass Perceptron
BIAS : 1win : 0game : 0 vote : 0 the : 0 ...
BIAS : 0win : 0game : 0vote : 0the : 0...
BIAS : 0 win : 0 game : 0 vote : 0 the : 0 ...
Question: What will the weights w be for each class after 3 updates?
wpolitics f(x1) = 0wsports f(x1) = 1
wtech f(x1) = 0
+ 1 + 1+ 0+ 1+ 1
Prediction:“sports”(wrong)
- 1 - 1 - 0- 1- 1
1 1011
f(x1) =
y1 = “politics”, x1 = “win the vote”y2 = “politics”, x2 = “win the election”y3 = “sports”, x3 = “win the game”
Example: Multiclass Perceptron
BIAS : 0win : -1game : 0 vote : -1 the : -1 ...
BIAS : 1win : 1game : 0vote : 1the : 1...
BIAS : 0 win : 0 game : 0 vote : 0the : 0 ...
y1 = “politics”, x1 = “win the vote”y2 = “politics”, x2 = “win the election”y3 = “sports”, x3 = “win the game”
Question: What will the weights w be for each class after 3 updates?
wpolitics f(x1) = 3wsports f(x1) = -2
wtech f(x1) = -3
Prediction:“politics”(correct)
1 1001
f(x2) =
Example: Multiclass Perceptron
BIAS : 0win : -1game : 0 vote : -1 the : -1 ...
BIAS : 1win : 1game : 0vote : 1the : 1...
BIAS : 0win : 0game : 0 vote : 0the : 0...
y1 = “politics”, x1 = “win the vote”y2 = “politics”, x2 = “win the election”y3 = “sports”, x3 = “win the game”
Question: What will the weights w be for each class after 3 updates?
wpolitics f(x1) = 3wsports f(x1) = -2
wtech f(x1) = -3
Prediction:“politics”(wrong)
1 1101
f(x3) =
- 1 - 1- 1- 0- 1
+ 1 + 1 + 1+ 0+ 1
Example: Multiclass Perceptron
BIAS : 1win : 0game : 1 vote : -1the : 0 ...
BIAS : 0win : 0game : -1vote : 1the : 0...
BIAS : 0win : 0game : 0vote : 0the : 0...
y1 = “politics”, x1 = “win the vote”y2 = “politics”, x2 = “win the election”y3 = “sports”, x3 = “win the game”
Question: What will the weights w be for each class after 3 updates?
δ
Properties of Perceptrons
• Separability: true if there exists weights w that get the training set perfectly correct
• Convergence: if the training data are separable, a perceptron will eventually converge (binary case)
• Mistake Bound: the maximum number of mistakes (updates)(binary case) is related to the number of features kand the margin δ or degree of separability
Separable
Non-Separable
Problems with the Perceptron
• Noise: if the data isn’t separable, weights might thrash• Averaging weight vectors over time
can help (averaged perceptron)
• Mediocre generalization: finds a “barely” separating solution
• Overtraining: test / held-out accuracy usually rises, then falls• Overtraining is a kind of overfitting
Improving the Perceptron
Non-Separable Case: Deterministic DecisionEven the best linear boundary makes at least one mistake
Non-Separable Case: Probabilistic Decision
0.5 | 0.50.3 | 0.7
0.1 | 0.9
0.7 | 0.30.9 | 0.1
How to get probabilistic decisions?
• Perceptron scoring:• If very positive à want probability going to 1• If very negative à want probability going to 0