Optimization Tutorial

Post on 23-Feb-2016

36 views 0 download

Tags:

description

Optimization Tutorial. Pritam Sukumar & Daphne Tsatsoulis CS 546: Machine Learning for Natural Language Processing. What is Optimization?. Find the minimum or maximum of an objective function given a set of constraints:. Why Do We Care?. Linear ClassificationMaximum Likelihood - PowerPoint PPT Presentation

Transcript of Optimization Tutorial

1

Optimization Tutorial

Pritam Sukumar & Daphne TsatsoulisCS 546: Machine Learning for Natural Language Processing

2

What is Optimization?

Find the minimum or maximum of an objective function given a set of constraints:

3

Why Do We Care?

Linear ClassificationMaximum Likelihood

K-Means

4

Prefer Convex Problems

Local (non global) minima and maxima:

5

Convex Functions and Sets

6

Important Convex Functions

7

Convex Optimization Problem

8

Lagrangian Dual

9

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

10

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

11

Gradient Descent

12

Single Step Illustration

13

Full Gradient Descent Illustration

14

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

15

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

16

Newton’s Method

Inverse Hessian Gradient

17

Newton’s Method Picture

18

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

19

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

20

Subgradient Descent Motivation

21

Subgradient Descent – Algorithm

22

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

23

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

24

Online learning and optimization

• Goal of machine learning :– Minimize expected loss

given samples

• This is Stochastic Optimization– Assume loss function is convex

25

Batch (sub)gradient descent for ML

• Process all examples together in each step

• Entire training set examined at each step• Very slow when n is very large

26

Stochastic (sub)gradient descent

• “Optimize” one example at a time• Choose examples randomly (or reorder and

choose in order)– Learning representative of example distribution

27

Stochastic (sub)gradient descent

• Equivalent to online learning (the weight vector w changes with every example)

• Convergence guaranteed for convex functions (to local minimum)

28

Hybrid!

• Stochastic – 1 example per iteration• Batch – All the examples!• Sample Average Approximation (SAA): – Sample m examples at each step and perform SGD

on them• Allows for parallelization, but choice of m

based on heuristics

29

SGD - Issues

• Convergence very sensitive to learning rate ( ) (oscillations near solution due to probabilistic nature of sampling)– Might need to decrease with time to ensure the

algorithm converges eventually• Basically – SGD good for machine learning

with large data sets!

30

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

31

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

32

Problem Formulation

33

New Points

34

Limited Memory Quasi-Newton Methods

35

Limited Memory BFGS

36

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

37

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

38

Coordinate descent

• Minimize along each coordinate direction in turn. Repeat till minimum is found– One complete cycle of coordinate descent is the

same as gradient descent• In some cases, analytical expressions available:– Example: Dual form of SVM!

• Otherwise, numerical methods needed for each iteration

39

Dual coordinate descent

• Coordinate descent applied to the dual problem

• Commonly used to solve the dual problem for SVMs– Allows for application of the Kernel trick– Coordinate descent for optimization

• In this paper: Dual logistic regression and optimization using coordinate descent

40

Dual form of SVM

• SVM

• Dual form

41

Dual form of LR

• LR:

• Dual form (we let )

42

Coordinate descent for dual LR

• Along each coordinate direction:

43

Coordinate descent for dual LR

• No analytical expression available– Use numerical optimization (Newton’s

method/bisection method/BFGS/…)to iterate along each direction

• Beware of log!

44

Coordinate descent for dual ME

• Maximum Entropy (ME) is extension of LR to multi-class problems– In each iteartion, solve in two levels:• Outer level – Consider block of variables at a time

– Each block has all labels and one example• Inner level – Subproblem solved by dual coordinate

descent

• Can also be solved similar to online CRF (exponentiated gradient methods)

45

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

46

First Order Methods: Gradient Descent

Newton’s MethodIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Subgradient DescentIntroduction to Convex Optimization for Machine Learning, John Duchi, UC Berkeley, Tutorial, 2009

Stochastic Gradient DescentStochastic Optimization for Machine Learning, Nathan Srebro and Ambuj Tewari, presented at ICML'10

Trust RegionsTrust Region Newton method for large-scale logistic regression, C.-J. Lin, R. C. Weng, and S. S. Keerthi, Journal of Machine Learning Research, 2008

Dual Coordinate DescentDual Coordinate Descent Methods for logistic regression and maximum entropy models, H.-F. Yu, F.-L. Huang, and C.-J. Lin, . Machine Learning Journal, 2011

Linear ClassificationRecent Advances of Large-scale linear classification, G.-X. Yuan, C.-H. Ho, and C.-J. Lin. Proceedings of the IEEE, 100(2012)

47

Large scale linear classification

• NLP (usually) has large number of features, examples

• Nonlinear classifiers (including kernel methods) more accurate, but slow

48

Large scale linear classification

• Linear classifiers less accurate, but at least an order of magnitude faster– Loss in accuracy lower with increase in number of

examples• Speed usually dependent on more than

algorithm order– Memory/disk capacity– Parallelizability

49

Large scale linear classification

• Choice of optimization method depends on:– Data property• Number of examples, features• Sparsity

– Formulation of problem• Differentiability• Convergence properties

– Primal vs dual– Low order vs high order methods

50

Comparison of performance

• Performance gap goes down with increase in number of features

• Training, testing time for linear classifiers is much faster

51

Thank you!

• Questions?