DOC PREVIEW
MIT 6 079 - Homework-6.079

This preview shows page 1 out of 2 pages.

Save
View full document
View full document
Premium Document
Do you want full access? Go Premium and unlock all 2 pages.
Access to all documents
Download any document
Ad free experience
Premium Document
Do you want full access? Go Premium and unlock all 2 pages.
Access to all documents
Download any document
Ad free experience

Unformatted text preview:

6.079/6.975, Fall 2009–10 S. Boyd & P. Parrilo Homework 10 additional problems 1. Suggestions for exercises 9.30 in Convex Optimization. We recommend the following to generate a problem instance: n = 100; m = 200; randn(’state’,1); A=randn(m,n); Of course, you should try out your code with different dimensions, and different data as well. In all cases, be sure that your line search first finds a step length for which the tentative point is in dom f; if you attempt to eva luate f o utside its domain, you’ll get complex numbers, and you’ll never recover. To find expressions for ∇f(x) and ∇2f(x), use the chain rule (see Appendix A.4); if you at tempt to compute ∂2f(x)/∂xi∂xj , you will be sorry. To compute the Newton step, you can use vnt=-H\g. 2. Suggestions for exercise 9.31 in Convex Optimization. For 9.31 a, you should try out N = 1, N = 15, and N = 30. You might as well compute and store the Cholesky factorization of the Hessian, and then back solve to get the search directions, even though you won’t really see any speedup in Matlab for such a small problem. After you evaluate the Hessian, you can find the Cholesky factorization as L=chol(H,’lower’). You can then compute a search step as -L’\(L\g), where g is the gradient at the current point. Matlab will do the right thing, i.e., it will first solve L\g using forward substitution, and then it will solve -L’\(L\g) using backward substitution. Each substitution is order n2 . To fairly compare the convergence of the three methods (i.e., N = 1, N = 15, N = 30), the horizontal axis should show the approximate total number of flops required, and not the number of iterations. You can compute the approximate number of flops using n3/3 for each factorization, and 2n2 for each solve (where each ‘solve’ involves a forward substitution step and a backwa r d substitution step). 1MIT OpenCourseWare http://ocw.mit.edu 6.079 / 6.975 Introduction to Convex Optimization Fall 2009 For information about citing these materials or our Terms of Use, visit:


View Full Document

MIT 6 079 - Homework-6.079

Download Homework-6.079
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view Homework-6.079 and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view Homework-6.079 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?