CS 726: Nonlinear Optimization I - Spring 2018

Schedule

Lecture: 11:00-12:15, Computer Science 1325
Course URL: http://www.cs.wisc.edu/~swright/cs726-s18.html
Canvas Course Page: https://canvas.wisc.edu/courses/77589/

In general, classes will be held on MWF every week, and most lectures will be 60 minutes, but may take up the full 75-minutes slot on a few occasions. (I will be absent on a number of class days, and the longer lectures will make up for these absences.)

The lecture schedule is posted below - it is subject to change but each week's schedule will be finalized by the preceding week.

  • Week 1: Wed 1/24 (60 min), Fri 1/26 (60 min)
  • Week 2: Mon 1/29 (60 min), Wed 1/31 (60 min), Fri 2/2 (60 min)
  • Week 3: Mon 2/5 (60 min), Wed 2/7 (60 min), Fri 2/9 (60 min)

 


Instructor: Steve Wright

Office: 4379 CS
Email: swright at cs.wisc
Office Hours: TBD

Teaching Assistant: Ching-pei Lee

Office: WID
Email: ching-pei at cs.wisc
Office Hours: TBD

General Course Information

Prerequisite

  • Linear Algebra, some Analysis. See guidebook for specifics.
  • We'll use Matlab for some homeworks. You are responsible for teaching yourself, if you don't already know it. (We will not use highly advanced features of the language.) We'll use cvx (a Matlab add-on that allows modeling and solution of some convex optimization problems) for some homeworks.

Text

  • J. Nocedal and S. J. Wright, Numerical Optimization, Second Edition, Springer, 2006. (It's essential to get the second edition!) Here is the current list of typos.

References

  • D. P. Bertsekas, with A. Nedic and A. Ozdaglar, Convex Analysis and Optimization, Athena Scientific, Belmont, MA, 2003.
  • Nesterov, Y., Introductory Lectures on Convex Optimization, Kluwer, 2004.
  • S. Boyd and L. Vandenberghe, Convex Optimization, Cambridge University Press. Available here.
  • D. P. Bertsekas, Nonlinear Programming, Second Edition, Athena Scientific, Belmont, MA, 1999.
  • R. Fletcher, Practical Methods of Optimization, 2nd Edition, Wiley, Chichester & New York, 1987.
  • R. T. Rockafellar and R. J.-B. Wets, Variational Analysis, Springer, 1998. (This is a more advanced book and an invaluable reference.)
  • A. Ruszczynski, Nonlinear Optimization, Princeton University Press, 2006.
  • S. J. Wright, Primal-Dual Interior-Point Methods, SIAM, 1997.

Lecture Notes

These can be found on the resources page of the piazza site and are updated sporadically. They are mostly for my own benefit and hence are terse, and refer heavily to the text in places. I supply them to help you fill in any missing details in the notes that you take in class.

 


Course Outline

This will likely be adadpted as the semester proceeds, but most of the following topics will be covered.

  • Introduction
    • Optimization paradigms and applications
    • Mathematical background: convex sets and functions, linear algebra, topology, convergence rates
  • Smooth unconstrained optimization: Background
    • Taylor's theorem
    • Optimality conditions
  • First-Order Methods
    • Steepest descent. Convergence for convex and nonconvex cases.
    • Accelerated gradient. Convergence for convex case.
    • Line search methods based on descent directions
    • Conjugate gradient methods
    • Conditional gradient for optimization over closed convex sets
  • Higher-order methods
    • Newton's method
    • Line-search Newton
    • Trust-region Newton and cubic regularization
    • Conjugate gradient-Newton
    • Quasi-Newton methods
    • Limited-memory quasi-Newton
  • Stochastic optimization
    • Basic methods and their convergence properties
    • reduced-variance approaches
  • Differentiation
    • Adjoint calculations
    • Automatic differentiation
  • Derivative-free optimization
    • Model-based methods
    • Pattern-search methods
  • Least-squares and nonlinear equations
    • Linear least squares: direct and iterative methods
    • Nonlinear least squares: Gauss-Newton, Levenberg-Marquardt
    • Newton’s method for nonlinear equations
    • Merit functions for nonlinear equations, and line searches
  • Optimization with linear constraints
    • Normal cones to convex sets
    • Farkas Lemma and first-order optimality conditions (KKT)
    • Gradient projection algorithms

Assessment

(The scheme below is provisional and subject to change.)

Keep track of your grades through canvas

  • 8-10 homework assignments, 30% of grade. Some of these may be graded by class members, using a key supplied by me. Some may be graded by the TA. Some may be graded by me. Some will not be graded.
    • Homework is due at the beginning of class on the designated date.
    • No homework or project is accepted in mailbox of instructor.
    • You may discuss homework with classmates. However, you may not share any code, carry out the assignment together, or copy solutions from another person. Discussion should be verbal only. The submitted version must be worked out, written, and submitted by you alone.
    • Submitting someone else's work as your own is academic misconduct. Such cheating and plagiarism will be dealt with in accordance with University procedures (see the page on Academic Misconduct).
  • MIDTERM, approx 25% of grade. Date, Time, Location TBA
  • FINAL, approx 45% of grade. Sunday, 6 May 2018, 10:05a-12:05p. Room: TBA.

Homeworks


Previous Exams

Note that the curriculum has been changed significantly since 2010, so some questions on the earlier exams are not relevant to the current version of the course.

Handouts and Examples

These will be posted as modules on the Canvas site.

Miscellaneous