Computer Science 4442A/B Lecture Notes - Lecture 5: Linear Classifier, Joule, Wfor-Tv

33 views9 pages

Document Summary

Optimization: how to minimize a function of a single variable. Gradient direction: j(x) is the direction of the steepest increase of function j(x) Gradient direction in 1d: - j(x) points in the direction of the steepest decrease, gradient is just derivative in 1d. Gradient direction: 1d: just a derivative (example above) Let x = 3 d ( ) 4. Let a = [10, 5] ( ) 10 aj. J(x1, x2) =(x1-5)2+(x2-10)2 ( ) 10 global min aj. J(x1, x2) =(x1-5)2+(x2-10)2 x2: given j(x1, x2) = (x1 - 5)2 + (x2 - 10)2, which step size to take, which step size to take, how do we select the step size, controlled by parameter . 5: controlled by parameter a (the learning rate, controlled by parameter . From previous example: a = [10 5] a = [10 5] Let = 0. 2 a - j(a) = [10 5]+0. 2 [-10 10]=[8 7] a - j(a) = [10 5]+0. 2 [-10 10]=[8 7]

Get access

Grade+20% off
$8 USD/m$10 USD/m
Billed $96 USD annually
Grade+
Homework Help
Study Guides
Textbook Solutions
Class Notes
Textbook Notes
Booster Class
40 Verified Answers
Class+
$8 USD/m
Billed $96 USD annually
Class+
Homework Help
Study Guides
Textbook Solutions
Class Notes
Textbook Notes
Booster Class
30 Verified Answers

Related Documents