v2009.01.01 - Convex Optimization
v2009.01.01 - Convex Optimization v2009.01.01 - Convex Optimization
610 APPENDIX D. MATRIX CALCULUS The gradient of vector-valued function v(x) : R→R N on real domain is a row-vector [ ∇v(x) = ∆ ∂v 1 (x) ∂x while the second-order gradient is ∂v 2 (x) ∂x · · · ∂v N (x) ∂x ] ∈ R N (1639) [ ∇ 2 v(x) = ∆ ∂ 2 v 1 (x) ∂x 2 ∂ 2 v 2 (x) ∂x 2 · · · ] ∂ 2 v N (x) ∈ R N (1640) ∂x 2 Gradient of vector-valued function h(x) : R K →R N on vector domain is ⎡ ∇h(x) = ∆ ⎢ ⎣ ∂h 1 (x) ∂x 1 ∂h 1 (x) ∂x 2 . ∂h 2 (x) ∂x 1 · · · ∂h 2 (x) ∂x 2 · · · . ∂h N (x) ∂x 1 ∂h N (x) ∂x 2 . ⎦ (1641) ∂h 1 (x) ∂h 2 (x) ∂h ∂x K ∂x K · · · N (x) ∂x K = [ ∇h 1 (x) ∇h 2 (x) · · · ∇h N (x) ] ∈ R K×N while the second-order gradient has a three-dimensional representation dubbed cubix ; D.1 ⎡ ∇ 2 h(x) = ∆ ⎢ ⎣ ∇ ∂h 1(x) ∂x 1 ∇ ∂h 1(x) ∂x 2 . ⎤ ⎥ ∇ ∂h 2(x) ∂x 1 · · · ∇ ∂h N(x) ∂x 1 ∇ ∂h 2(x) ∂x 2 · · · ∇ ∂h N(x) ∂x 2 . . ∇ ∂h 2(x) ∂x K · · · ∇ ∂h N(x) ⎦ (1642) ∇ ∂h 1(x) ∂x K ∂x K = [ ∇ 2 h 1 (x) ∇ 2 h 2 (x) · · · ∇ 2 h N (x) ] ∈ R K×N×K where the gradient of each real entry is with respect to vector x as in (1637). ⎤ ⎥ D.1 The word matrix comes from the Latin for womb ; related to the prefix matri- derived from mater meaning mother.
D.1. DIRECTIONAL DERIVATIVE, TAYLOR SERIES 611 The gradient of real function g(X) : R K×L →R on matrix domain is ⎡ ∇g(X) = ∆ ⎢ ⎣ = ∂g(X) ∂X 11 ∂g(X) ∂X 21 . ∂g(X) ∂X K1 ∂g(X) ∂X 12 · · · ∂g(X) [ ∇X(:,1) g(X) ∂X 22 . · · · ∂g(X) ∂X K2 · · · ∂g(X) ∂X 1L ∂g(X) ∂X 2L . ∂g(X) ∂X KL ⎤ ∈ R K×L ⎥ ⎦ ∇ X(:,2) g(X) ... ∇ X(:,L) g(X) ] ∈ R K×1×L (1643) where the gradient ∇ X(:,i) is with respect to the i th column of X . The strange appearance of (1643) in R K×1×L is meant to suggest a third dimension perpendicular to the page (not a diagonal matrix). The second-order gradient has representation ⎡ ∇ 2 g(X) = ∆ ⎢ ⎣ = ∇ ∂g(X) ∂X 11 ∇ ∂g(X) ∂X 21 . ∇ ∂g(X) ∂X K1 [ ∇∇X(:,1) g(X) ∇ ∂g(X) ∂X 12 · · · ∇ ∂g(X) ∂X 1L ∇ ∂g(X) ∂X 22 · · · ∇ ∂g(X) ∂X 2L . . ∇ ∂g(X) ∂X K2 · · · ∇ ∂g(X) ∂X KL ⎤ ∈ R K×L×K×L ⎥ ⎦ ∇∇ X(:,2) g(X) ... ∇∇ X(:,L) g(X) ] ∈ R K×1×L×K×L (1644) where the gradient ∇ is with respect to matrix X .
- Page 559 and 560: A.4. SCHUR COMPLEMENT 559 A.4.0.0.2
- Page 561 and 562: A.5. EIGEN DECOMPOSITION 561 When B
- Page 563 and 564: A.5. EIGEN DECOMPOSITION 563 dim N(
- Page 565 and 566: A.6. SINGULAR VALUE DECOMPOSITION,
- Page 567 and 568: A.6. SINGULAR VALUE DECOMPOSITION,
- Page 569 and 570: A.7. ZEROS 569 Given symmetric matr
- Page 571 and 572: A.7. ZEROS 571 (TRANSPOSE.) Likewis
- Page 573 and 574: A.7. ZEROS 573 For X,A∈ S M + [31
- Page 575 and 576: A.7. ZEROS 575 A.7.5.0.1 Propositio
- Page 577 and 578: Appendix B Simple matrices Mathemat
- Page 579 and 580: B.1. RANK-ONE MATRIX (DYAD) 579 R(v
- Page 581 and 582: B.1. RANK-ONE MATRIX (DYAD) 581 ran
- Page 583 and 584: B.2. DOUBLET 583 R([u v ]) R(Π)= R
- Page 585 and 586: B.3. ELEMENTARY MATRIX 585 If λ
- Page 587 and 588: B.4. AUXILIARY V -MATRICES 587 the
- Page 589 and 590: B.4. AUXILIARY V -MATRICES 589 18.
- Page 591 and 592: B.5. ORTHOGONAL MATRIX 591 B.5 Orth
- Page 593 and 594: B.5. ORTHOGONAL MATRIX 593 Figure 1
- Page 595 and 596: Appendix C Some analytical optimal
- Page 597 and 598: C.2. TRACE, SINGULAR AND EIGEN VALU
- Page 599 and 600: C.2. TRACE, SINGULAR AND EIGEN VALU
- Page 601 and 602: C.2. TRACE, SINGULAR AND EIGEN VALU
- Page 603 and 604: C.3. ORTHOGONAL PROCRUSTES PROBLEM
- Page 605 and 606: C.4. TWO-SIDED ORTHOGONAL PROCRUSTE
- Page 607 and 608: C.4. TWO-SIDED ORTHOGONAL PROCRUSTE
- Page 609: Appendix D Matrix calculus From too
- Page 613 and 614: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 615 and 616: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 617 and 618: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 619 and 620: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 621 and 622: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 623 and 624: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 625 and 626: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 627 and 628: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 629 and 630: D.1. DIRECTIONAL DERIVATIVE, TAYLOR
- Page 631 and 632: D.2. TABLES OF GRADIENTS AND DERIVA
- Page 633 and 634: D.2. TABLES OF GRADIENTS AND DERIVA
- Page 635 and 636: D.2. TABLES OF GRADIENTS AND DERIVA
- Page 637 and 638: D.2. TABLES OF GRADIENTS AND DERIVA
- Page 639 and 640: Appendix E Projection For any A∈
- Page 641 and 642: 641 U T = U † for orthonormal (in
- Page 643 and 644: E.1. IDEMPOTENT MATRICES 643 where
- Page 645 and 646: E.1. IDEMPOTENT MATRICES 645 order,
- Page 647 and 648: E.1. IDEMPOTENT MATRICES 647 When t
- Page 649 and 650: E.3. SYMMETRIC IDEMPOTENT MATRICES
- Page 651 and 652: E.3. SYMMETRIC IDEMPOTENT MATRICES
- Page 653 and 654: E.3. SYMMETRIC IDEMPOTENT MATRICES
- Page 655 and 656: E.5. PROJECTION EXAMPLES 655 E.4.0.
- Page 657 and 658: E.5. PROJECTION EXAMPLES 657 a ∗
- Page 659 and 660: E.5. PROJECTION EXAMPLES 659 E.5.0.
610 APPENDIX D. MATRIX CALCULUS<br />
The gradient of vector-valued function v(x) : R→R N on real domain is<br />
a row-vector<br />
[<br />
∇v(x) =<br />
∆<br />
∂v 1 (x)<br />
∂x<br />
while the second-order gradient is<br />
∂v 2 (x)<br />
∂x<br />
· · ·<br />
∂v N (x)<br />
∂x<br />
]<br />
∈ R N (1639)<br />
[<br />
∇ 2 v(x) =<br />
∆<br />
∂ 2 v 1 (x)<br />
∂x 2<br />
∂ 2 v 2 (x)<br />
∂x 2 · · ·<br />
]<br />
∂ 2 v N (x)<br />
∈ R N (1640)<br />
∂x 2<br />
Gradient of vector-valued function h(x) : R K →R N on vector domain is<br />
⎡<br />
∇h(x) =<br />
∆ ⎢<br />
⎣<br />
∂h 1 (x)<br />
∂x 1<br />
∂h 1 (x)<br />
∂x 2<br />
.<br />
∂h 2 (x)<br />
∂x 1<br />
· · ·<br />
∂h 2 (x)<br />
∂x 2<br />
· · ·<br />
.<br />
∂h N (x)<br />
∂x 1<br />
∂h N (x)<br />
∂x 2<br />
.<br />
⎦<br />
(1641)<br />
∂h 1 (x) ∂h 2 (x) ∂h<br />
∂x K ∂x K<br />
· · · N (x)<br />
∂x K<br />
= [ ∇h 1 (x) ∇h 2 (x) · · · ∇h N (x) ] ∈ R K×N<br />
while the second-order gradient has a three-dimensional representation<br />
dubbed cubix ; D.1<br />
⎡<br />
∇ 2 h(x) =<br />
∆ ⎢<br />
⎣<br />
∇ ∂h 1(x)<br />
∂x 1<br />
∇ ∂h 1(x)<br />
∂x 2<br />
.<br />
⎤<br />
⎥<br />
∇ ∂h 2(x)<br />
∂x 1<br />
· · · ∇ ∂h N(x)<br />
∂x 1<br />
∇ ∂h 2(x)<br />
∂x 2<br />
· · · ∇ ∂h N(x)<br />
∂x 2<br />
. .<br />
∇ ∂h 2(x)<br />
∂x K<br />
· · · ∇ ∂h N(x)<br />
⎦<br />
(1642)<br />
∇ ∂h 1(x)<br />
∂x K<br />
∂x K<br />
= [ ∇ 2 h 1 (x) ∇ 2 h 2 (x) · · · ∇ 2 h N (x) ] ∈ R K×N×K<br />
where the gradient of each real entry is with respect to vector x as in (1637).<br />
⎤<br />
⎥<br />
D.1 The word matrix comes from the Latin for womb ; related to the prefix matri- derived<br />
from mater meaning mother.