10.07.2015 Views

v2007.09.13 - Convex Optimization

v2007.09.13 - Convex Optimization

v2007.09.13 - Convex Optimization

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

568 APPENDIX D. MATRIX CALCULUSD.1.8.1.1 Example. Gradient.g(X) = w T X T Xw , X ∈ R K×L , w ∈R L . Using the tables inD.2,tr ( ∇ X g(X+ t Y ) T Y ) = tr ( 2ww T (X T + t Y T )Y ) (1612)Applying the equivalence (1610),= 2w T (X T Y + t Y T Y )w (1613)ddt g(X+ t Y ) = d dt wT (X+ t Y ) T (X+ t Y )w (1614)= w T( X T Y + Y T X + 2t Y T Y ) w (1615)= 2w T (X T Y + t Y T Y )w (1616)which is the same as (1613); hence, equivalence is demonstrated.It is easy to extract ∇g(X) from (1616) knowing only (1610):D.1.8.2tr ( ∇ X g(X+ t Y ) T Y ) = 2w T (X T Y + t Y T Y )w= 2 tr ( ww T (X T + t Y T )Y )tr ( ∇ X g(X) T Y ) = 2 tr ( ww T X T Y )(1617)⇔∇ X g(X) = 2Xww Tsecond-orderLikewise removing the evaluation at t = 0 from (1599),→Ydg 2 (X+ t Y ) = d2dt2g(X+ t Y ) (1618)we can find a similar relationship between the second-order gradient and thesecond derivative: In the general case g(X) : R K×L →R M×N from (1592) and(1595),tr(∇ X tr ( ∇ X g mn (X+ t Y ) T Y ) )TY = d2dt 2g mn(X+ t Y ) (1619)In the case of a real function g(X) : R K×L →R we have, of course,tr(∇ X tr ( ∇ X g(X+ t Y ) T Y ) )TY = d2dt2g(X+ t Y ) (1620)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!