23.01.2014 Views

Lecture 18 Subgradients

Lecture 18 Subgradients

Lecture 18 Subgradients

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Lecture</strong> <strong>18</strong><br />

Hyperplane Theorem, there exists a nonzero vector (a, β) ∈ R n+1 such<br />

that<br />

a T (x + αd) + β(f(x) + αf ′ (x; d)) ≤ a T z + βw, (1)<br />

for all α ≥ 0, z ∈ dom f, and f(z) < w. We must have β ≥ 0 - why? We<br />

cannot have β = 0 -why?<br />

Thus, β > 0 and we can divide by β the relation in (1), and obtain with<br />

ã = a/β,<br />

ã T (x + αd) + f(x) + αf ′ (x; d) ≤ ã T z + w, (2)<br />

for all α ≥ 0, z ∈ dom f, and f(z) < w. Choosing α = 0 and letting<br />

w ↓ f(z), we see<br />

ã T x + f(x) ≤ ã T z + f(z),<br />

implying that f(x) − ã T (z − x) ≤ f(z) for all z ∈ dom f. Therefore<br />

−ã ∈ ∂f(x).<br />

Convex Optimization 13

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!