Semidefinite Programming

Two remarks on previous lecture

• Fact 9 shows that pd (or psd) can be verified by checking that every principal minor
is positive (or nonnegative). For pd we only need to examine leading principal minors.
However, this does not work for verifying psd as

shows. This matrix has leading minors equal to zero but a negative trailing minor.

• Fact 10 has introduced the Schur complement. If

This lends itself to showing that the Cholesky factorization of a pd matrix exists.

Fact 11 (Representing quadratics) Given the quadratic form

while a quadratic function of x, is a linear function of X := xxT .

Fact 12 If U and V are psd, then U • V≥0 (used in showing weak duality for SDPs). In
fact, we have the following theorem.

Theorem 1 If , then K is self-dual,

Proof: Part 1: showing want to show that if
Demonstration 1: By facts 3, 4, and 6 we can assume that one of U, V is diagonal. But then
(need only look at the diagonal entries).
Demonstration 2: By Fact 1, and the existence of matrix square roots ,

since is psd (Fact 7) and so its trace is nonnegative

Part 2: showing by contraposition, i.e., if
If , then by definition there exists a z such that (by Fact 11).
But by Corollary 1, so

By the same argument, if This shows a similarity
between nonnegative vectors and the cone of psd matrices.

Fact 13 If is compact.
Proof: The set is clearly closed, so need to show it is also bounded. Suppose
(since Then

for any So any V in the set has

Fact 14 If
Proof: First note that if UV = 0 then clearly U • V = trace (UV ) = 0. To show the
converse suppose with U • V = 0. Then

so Define to obtain the condition ATA = 0.
This implies A = 0 since the diagonal entries of ATA are the squares of the 2-norms of the
columns of A. But easily gives UV = 0.

Fact 15 If then U, V commute iff UV is symmetric, and iff they can be
simultaneously diagonalized , i.e., iff we can find Q, ,M (Q orthogonal, ,M diagonal) with

Proof: Omitted, since part of HW1.

Applications

Matrix and NLP optimization problems

We have already shown how to minimize the maximum eigenvalue of a symmetric
What if M(y) is not symmetric and not necessarily square ? For example, say we want
as small as possible. Recall:

Lemma 1

Proof: Observe that the above matrix retains linearity w.r.t. η, y. Examine two cases:
If P = 0, then both conditions reduce to
If P ≠ 0,

With η > 0, the last condition is equivalent to which holds
iffwhich is what we needed to show.
Note that we have also shown that

So, is equivalent to

which is the desired SDP in dual form.

Linear programming and nonlinear programming

Consider first an LP in dual form

and try to rewrite it as an SDP. The constraint

where C = Diag(c) and Aii = Diag(ai) and ai is the ith column of AT . This gives the equivalent
SDP in dual form.

Now consider a correspondence between primal LP and SDP. The primal SDP is given by

whereas the primal LP is given by

In general, our SDPs may have block diagonal form. Suppose and
define

If C and all Ai’s are in S then automatically for any The primal
problem involves

but we can restrict X to S without loss of generality (see HW1). Applying this to the primalform
SDP above we can assume that X is diagonal, so that it reduces to the primal -form
LP.
 

Prev Next