Notes on Symmetric Matrices

1. Symmetric Matrices

We review some basic results concerning symmetric matrices. All matrices that we discuss are over the real numbers.

Let {A} be a real, symmetric matrix of size {d \times d} and let {I} denote the {d \times d} identity matrix. Perhaps the most important and useful property of symmetric matrices is that their eigenvalues behave very nicely.

Definition 1 Let {U} be a {d \times d} matrix. The matrix {U} is called an orthogonal matrix if {U ^T U = I}. This implies that {U U ^T = I}, by uniqueness of inverses.

Fact 2 (Spectral Theorem). Let {A} be any {d \times d} symmetric matrix. There exists an orthogonal matrix {U} and a (real) diagonal matrix {D} such that

\displaystyle  A ~=~ U D U^T.

This is called a spectral decomposition of {A}. Let {u_i} be the {i}th column of {U} and let {\lambda_i} denote the {i}th diagonal entry of {D}. Then {\{u_1,\ldots,u_d\}} is an orthonormal basis consisting of eigenvectors of {A}, and {\lambda_i} is the eigenvalue corresponding to {u_i}. We can also write

\displaystyle   A ~=~ \sum_{i=1}^d \lambda_i u_i u_i^T. \ \ \ \ \ (1)

The eigenvalues {\lambda} are uniquely determined by {A}, up to reordering.

Caution. The product of two symmetric matrices is usually not symmetric.

1.1. Positive semi-definite matrices

Definition 3 Let {A} be any {d \times d} symmetric matrix. The matrix {A} is called positive semi-definite if all of its eigenvalues are non-negative. This is denoted {A \succeq 0}, where here {0} denotes the zero matrix. The matrix {A} is called positive definite if all of its eigenvalues are strictly positive. This is denoted {A \succ 0}.

There are many equivalent characterizations of positive semi-definiteness. One useful condition is

\displaystyle   A \succeq 0 \qquad\iff\qquad x^T A x \geq 0 ~~\forall x \in {\mathbb R}^d. \ \ \ \ \ (2)


\displaystyle  A \succ 0 \qquad\iff\qquad x^T A x > 0 ~~\forall x \in {\mathbb R}^d

Another condition is: {A \succeq 0} if and only if there exists a matrix {V} such that {A = V V ^T}.

The positive semi-definite condition can be used to define a partial ordering on all symmetric matrices. This is called the Löwner ordering or the positive semi-definite ordering. For any two symmetric matrices {A} and {B}, we write {A \succeq B} if {A - B \succeq 0}.

Fact 4 {A \succeq B} if and only if {v^T A v \geq v^T B v} for all vectors {v}.

Proof: Note that {v^T A v \geq v^T B v} holds if and only if {v^T (A-B) v \geq 0} holds. By (2), this is equivalent to {A-B \succeq 0} which is the definition of {A \succeq B}. \Box

The set of positive semi-definite matrices is closed under addition and non-negative scaling. (Such a set is called a convex cone.)

Fact 5 Let {A} and {B} be positive semi-definite matrices of size {d \times d}. Let {\alpha, \beta} be non-negative scalars. Then {\alpha A + \beta B \succeq 0}.

Proof: This follows easily from (2). \Box

Caution. The Löwner ordering does not have all of the nice properties that the usual ordering of real numbers has. For example, if {A \succeq B \succeq 0} then it is not necessarily true that {A^2 \succeq B^2}.

Fact 6 Let {A} and {B} be symmetric, {d \times d} matrices and let {C} be any {d \times d} matrix. If {A \succeq B} then {C A C ^T \succeq C B C ^T}. Moreover, if {C} is non-singular then the “if” is actually “if and only if”.

Proof: Since {A \succeq B}, then {A-B \succeq 0}, so there exists a matrix {V} such that {V V ^T}. Then

\displaystyle  C (A-B) C ^T ~=~ C V V ^T C ^T ~=~ C V (C V) ^T,

which shows that {C (A-B) C ^T} is positive semi-definite.

Suppose that {C} is non-singular and { C A C ^T \succeq C B C ^T }. Multiply on the left by {C^{-1}} and on the right by {(C^{-1}) ^T = (C ^T)^{-1}} to get {A \succeq B}. \Box

1.2. Spectral Norm

Just as there are many norms of vectors, there are many norms of matrices too. We will only consider spectral norm of {A}.

Definition 7 Let {A} be a symmetric {d \times d} matrix with eigenvalues {\lambda_1,\ldots,\lambda_d}. The spectral norm of {A} is denoted {\lVert A \rVert} and is defined by {\lVert A \rVert = \max_i |\lambda_i|}.

Recall that an eigenvalue of {A} is any real number {z} such that there exists a vector {u} for which {Au=zu}. Then we also have {(A+tI) u = (z+t)u}, which implies that {z+t} is an eigenvalue of {A+tI}. In fact, the entire set of eigenvalues of {A + tI} is {\{\lambda_1+t, \ldots, \lambda_d+t \}}.

Fact 8 Let {A} be a symmetric matrix. Let {\lambda_{\mathrm{min}}} and {\lambda_{\mathrm{max}}} respectively denote the smallest and largest eigenvalues of {A}. Then {\lambda_{\mathrm{min}} \cdot I \preceq A \preceq \lambda_{\mathrm{max}} \cdot I}.

Proof: Let {\lambda_1,\ldots,\lambda_d} be all the eigenvalues of {A}. We will show only the second inequality. This is equivalent to {\lambda_{\mathrm{max}} \cdot I - A \succeq 0}. As observed above, the eigenvalues of {\lambda_{\mathrm{max}} \cdot I - A} are

\displaystyle  \{ \lambda_{\mathrm{max}} - \lambda_1, \ldots, \lambda_{\mathrm{max}} - \lambda_d \}.

The minimum of these eigenvalue is

\displaystyle  \min_j \big( \lambda_{\mathrm{max}} - \lambda_j \big) ~=~ \lambda_{\mathrm{max}} - \max_j \lambda_j ~=~ 0.


In particular, for any symmetric matrix {A} we have {A \preceq \lVert A \rVert \cdot I}.

1.3. Trace

Definition 9 Let {A} be an arbitrary {d \times d} matrix (not necessarily symmetric). The trace of {A}, denoted {\mathop{\mathrm{tr}}\,(A)}, is the sum of the diagonal entries of {A}.

Fact 10 (Linearity of Trace) Let {A} and {B} be arbitrary {d \times d} matrices and let {\alpha, \beta} be scalars. Then {\mathop{\mathrm{tr}}\,(\alpha A+ \beta B) = \alpha \mathop{\mathrm{tr}}\,(A)+ \beta \mathop{\mathrm{tr}}\,(B)}.

Fact 11 (Cyclic Property of Trace) Let {A} be an arbitrary {n \times m} matrix and let {B} be an arbitrary {m \times n} matrix. Then {\mathop{\mathrm{tr}}\,(AB) = \mathop{\mathrm{tr}}\,(BA)}.

Proof: This easily follows from the definitions. The {i}th diagonal entry of {AB} is {\sum_{j=1}^m A_{i,j} B_{j,i}}, so

\displaystyle  \mathop{\mathrm{tr}}\,(AB) ~=~ \sum_{i=1}^n \sum_{j=1}^m A_{i,j} B_{j,i} ~=~ \sum_{j=1}^m \sum_{i=1}^n B_{j,i} A_{i,j},

by swapping the order of summation. This equals {\mathop{\mathrm{tr}}\,(BA)} because the {j}th diagonal entry of {BA} is {\sum_{i=1}^n B_{j,i} A_{i,j}}. \Box

Fact 12 Let {A} be a symmetric, {d \times d} matrix with eigenvalues {\{\lambda_1,\ldots,\lambda_d\}}. Then {\mathop{\mathrm{tr}}\,(A) = \sum_{i=1}^d \lambda_d}.

Claim 13 Let {A}, {B} and {C} be symmetric {d \times d} matrices satisfying {A \succeq 0} and {B \preceq C}. Then {\mathop{\mathrm{tr}}\,(A B) \leq \mathop{\mathrm{tr}}\,(A C)}.

Proof: Suppose additionally that {A} is rank-one, i.e., {A = v v^T} for some vector {v}. Then

\displaystyle  \begin{array}{rcl}  \mathop{\mathrm{tr}}\,(A B) &=& \mathop{\mathrm{tr}}\,( v v^T B ) ~=~ \mathop{\mathrm{tr}}\,( v^T B v ) ~=~ v^T B v \qquad\quad \\ &\leq& v^T C v ~=~ \mathop{\mathrm{tr}}\,(v^T C v) ~=~ \mathop{\mathrm{tr}}\,(v v^T C) ~=~ \mathop{\mathrm{tr}}\,(A C). \end{array}

Now we consider the case where {A} has arbitrary rank. Let {A = \sum_{i=1}^d \lambda_i u_i u_i^T}. Since we assume that {A \succeq 0} we can set {v_i = \sqrt{\lambda_i} u_i} and write {A = \sum_{i=1}^d v_i v_i^T}. Then

\displaystyle  \begin{array}{rcl}  \mathop{\mathrm{tr}}\,(A B) &=& \mathop{\mathrm{tr}}\,\big( \sum_i v_i v_i^T B \big) ~=~ \sum_i \mathop{\mathrm{tr}}\,( v_i v_i^T B ) \\ &\leq& \sum_i \mathop{\mathrm{tr}}\,(v_i v_i^T C) ~=~ \mathop{\mathrm{tr}}\,\big( \sum_i v_i v_i^T C \big) ~=~ \mathop{\mathrm{tr}}\,(A C). \end{array}

Here the second and third equalities are by linearity of trace, and the inequality follows from the rank-one case. \Box

Corollary 14 If {A \succeq 0} then {\mathop{\mathrm{tr}}\,(A \cdot B) \leq \lVert B \rVert \cdot \mathop{\mathrm{tr}}\,(A)}.

Proof: Apply Claim 13 with {C = \lVert B \rVert \cdot I}. We get

\displaystyle  \mathop{\mathrm{tr}}\,(A \cdot B) ~\leq~ \mathop{\mathrm{tr}}\,(A \cdot \lVert B \rVert \cdot I) ~\leq~ \lVert B \rVert \cdot \mathop{\mathrm{tr}}\,(A \cdot I),

by linearity of trace since {\lVert B \rVert} is a scalar. \Box

1.4. Functions applied to matrices

For any function {f : {\mathbb R} \rightarrow {\mathbb R}}, we can extend {f} to a function on symmetric matrices as follows. Let {A = U D U^T} be a spectral decomposition of {A} where {\lambda_i} is the {i}th diagonal entry of {D}. Then we define

\displaystyle  f(A) ~=~ U \begin{pmatrix} f(\lambda_1) & & \\ & \ddots & \\ & & f(\lambda_d) \end{pmatrix} U^T.

In other words, {f(A)} is defined simplying by applying the function {f} to the eigenvalues of {A}.

Caution. {f(A)} is not obtained applying the function {f} to the entries of {A}!

One important example of applying a function to a matrix is powering. Let {A} be any positive semi-definite matrix with spectral decomposition {U D U ^T}. For any {c \in {\mathbb R}} we can define {A^c} as follows. Let {S} be the diagonal matrix with {S_{i,i} = (D_{i,i})^c}. Then we define {A^{c} = U S U ^T}. For example, consider the case {c=1/2}. Note that

\displaystyle  A^{1/2} \cdot A^{1/2} ~=~ U S U ^T \cdot U S U ^T ~=~ U S S U ^T ~=~ U D U ^T ~=~ A.

So the square of the square root is the matrix itself, as one would expect. If {A} is non-singular, the matrix {A^{-1}} obtained by taking {c=-1} is the same as the usual matrix inverse (by uniqueness of inverses, since {A^{-1} \cdot A = I}). So we see that the inverse of a non-singular symmetric matrix is obtained by inverting its eigenvalues.

Inequalities on real-valued functions also give us inequalities on matrices.

Claim 15 Let {f : {\mathbb R} \rightarrow {\mathbb R}} and {g : {\mathbb R} \rightarrow {\mathbb R}} satisfy {f(x) \leq g(x)} for all {x \in [l,u] \subset {\mathbb R}}. Let {A} be a symmetric matrix for which all eigenvalues lie in {[l,u]} (i.e., {l I \preceq A \preceq u I}). Then {f(A) \preceq g(A)}.

Proof: Let {A = U D U^T} be a spectral decomposition of {A} where {\lambda_i} is the {i}th diagonal entry of {D}. We wish to show {f(A) \preceq g(A)}, i.e., {g(A) - f(A) \succeq 0}. By (2), this is equivalent to

\displaystyle  v^T U \begin{pmatrix} g(\lambda_1)-f(\lambda_1) & & \\ & \ddots & \\ & & g(\lambda_d)-f(\lambda_d) \end{pmatrix} U v ~\geq~ 0 \qquad \forall v \in {\mathbb R}^d.

Since {U} is non-singular this is equivalent to

\displaystyle  w^T \begin{pmatrix} g(\lambda_1)-f(\lambda_1) & & \\ & \ddots & \\ & & g(\lambda_d)-f(\lambda_d) \end{pmatrix} w ~\geq~ 0 \qquad \forall w \in {\mathbb R}^d,

where we have simply replaced {w = U v}. The left-hand side is simply {\sum_i w_i^2 \big( g(\lambda_i) - f(\lambda_i)\big)}, which is non-negative by our assumptions on {g}, {f} and the {\lambda_i}‘s. \Box

1.5. The pseudoinverse

Sometimes we would like to invert a matrix which is not quite invertible. A useful alternative is to invert the part of the matrix that is invertible (its image) and to leave alone the part of the matrix that is not invertible (its kernel).

We can do this by applying the real-valued function:

\displaystyle  f(x) ~=~ \begin{cases} 1/x &\qquad(x \neq 0) \\ 0 &\qquad(x=0) \end{cases}.

The function {f} inverts all non-zero numbers and maps {0} to {0}. So if we apply {f} to a symmetric matrix, all non-zero eigenvalues will be inverted, and the zero eigenvalues will remain unchanged. The resulting matrix is called the pseudoinverse and is denoted {A^+}. If {A} is indeed invertible then {A^+ = A^{-1}}.

Suppose {A} is not invertible. We cannot have {A \cdot A^+ = I} since {I} has full rank and {A} does not. However, we have the next best thing: {A \cdot A^+} is the identity operator on the image of {A}. More precisely {A \cdot A^+} is the orthogonal projection onto the image of {A}, which we denote {I_{\mathrm{im}~A}}. In other words, for every {v} in the image of {A}, we have {A \cdot A^+ \, v = v} and for every {v} in the kernel of {A} we have {A \cdot A^+ \, v = 0}.

Claim 16 Let {A} and {B_1,\ldots,B_k} be symmetric, positive semi-definite matices of the same size. Let {A^{+/2} = (A^+)^{1/2}} denote the square root of the pseudoinverse of {A}. Let {\ell} and {u} be arbitrary scalars. Suppose that {\mathrm{im}(B_i)} is a subspace of {\mathrm{im}(A)} for all {i}. Then

\displaystyle  \ell A \:\preceq\: \sum_{i=1}^k B_i \:\preceq\: u A \qquad\Longleftrightarrow\qquad \ell I_{\mathrm{im}~A} \:\preceq\: \frac{1}{k} \sum_{i=1}^k A^{+/2} B_i A^{+/2} \:\preceq\: u I_{\mathrm{im}~A}

Proof: {\Longrightarrow}: Since {A^{+/2} A A^{+/2} = I_{\mathrm{im}~A}}, this follows from Fact 6.

{\Longleftarrow}: This also follows from Fact 6, because

\displaystyle  A^{1/2} I_{\mathrm{im}~A} A^{1/2} ~=~ A


\displaystyle  A^{1/2} A^{+/2} B_i A^{+/2} A^{1/2} ~=~ I_{\mathrm{im}~A} B_i I_{\mathrm{im}~A} ~=~ I_{\mathrm{im}~B_i} B_i I_{\mathrm{im}~B_i} ~=~ B_i,

where the second equality follows from the assumption that {\mathrm{im}(B_i) \subseteq \mathrm{im}(A)}. \Box

1.6. Matrix exponentials

We will mainly be interested in the case {f(x) = e^x}. For any symmetric matrix {A}, note that {e^A} is positive semi-definite. Whereas in the scalar case {e^{a+b} = e^a e^b} holds, it is not necessarily true in the matrix case that {e^{A+B} = e^A \cdot e^B}. It turns out that there is a useful inequality along these lines. It is called the Golden-Thompson inequality.

Theorem 17 (Golden-Thompson Inequality) Let {A} and {B} be symmetric {d \times d} matrices. Then {\mathop{\mathrm{tr}}\,(e^{A + B}) \leq \mathop{\mathrm{tr}}\,(e^A \cdot e^B)}.

This entry was posted in Uncategorized. Bookmark the permalink.

3 Responses to Notes on Symmetric Matrices

  1. Pingback: Lecture 12: Concentration for sums of random matrices | UBC CPSC 536N

  2. Pingback: Lecture 13: Spectral sparsifiers | UBC CPSC 536N

  3. Pingback: Lecture 14: Spectral sparsifiers | UBC CPSC 536N

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s