Go Back   Science Forums Biology Forum Molecular Biology Forum Physics Chemistry Forum > General Science Forums > Physics Forum
Register Search Today's Posts Mark Forums Read

Physics Forum Physics Forum. Discuss and ask physics questions, kinematics and other physics problems.

Functions of Matrices

Functions of Matrices - Physics Forum

Functions of Matrices - Physics Forum. Discuss and ask physics questions, kinematics and other physics problems.

LinkBack Thread Tools Display Modes
Old 10-31-2006, 01:43 PM
Posts: n/a
Default Functions of Matrices

By Herman Schoenfeld
Copyright (c) 2006

In this article we shall derive an explicit formula for the evaluation
of arbitrary functions of a square matrix. The formula for a matrix
function is given in terms of the equivalent function defined for a
scalar parameter.

The mechanism which allows us to achieve this is the Maclaurin series
expansion of scalar functions, having form

[1] f(x) = SUM(n=0..INF) k_n x^n

given a set of constants k_n.

By simple substitution of x with some arbitrary square MxM matrix A, we
can define the matrix-continuation of [1] as

[2] f(A) = SUM(n=0..INF) k_n A^n

This substitution is perfectly valid since the natural-numbered powers
of square matrices always exist. By convention we define

[3] A^0 = I


A sequence of matrices (of common order)

[4] B = SUM(n) A_n

converges to B so long as the elements of B also converge. That is,

[5] B_ij = SUM(n) (A_n)_ij

Now, consider the Maclaurin series expansion of a scalar function as
seen in [1] and its matrix continuation as seen in [2]. If [1]
converges all |x| < N then [2] converges for all (square) matrices A
that have all its eigenvalues |e| < N.

We can see that this is true if we understand that a function of matrix
can be given in terms of the same function of its eigenvalues. If the
function diverges for the eigenvalue, then it diverges for the matrix
containing those eigenvalues.


If we attempt to derive a formula for a matrix function using the
Maclaurin matrix expansion as seen in [2], we will get results which
are, in most cases, incalculable with a computer and of little value

Rather, we are able to take a different approach, one which makes use
of the scalar version of the matrix function.

We do this by using an important result deriving from the
Cayley-Hamilton theorem which tells us that a function of an MxM matrix
expands as a matrix polynomial of degree (M-1)

[6] f(A) = SUM(n=0..M-1) k_n A^n

If we solve for the constants k_n in terms of the scalar function f(x),
we can get a meaningful result for the matrix function f(A) by simply
evaluating the sum in [6].

Luckily, we can solve for these constants by making use of a related
Cayley-Hamilton result. It turns out that the scalar version of [6] is
true for the same constants k_n so long as the parameter to the
function is an eigenvalue e of A . In other words,

[7] f(e) = SUM(n=0..M-1) k_n e^n

Equation [7] is useful to us because it allows us to solve the
constants k_n in terms of the scalar function f.

Now, if A is an MxM matrix, it follows from the characteristic
polynomial of A that A has M eigenvalues (not necessarily all
distinct). This means that we have M solutions for equation [7], giving
us a system of simultaneous linear equations

[8] [ f(e_1) ] [ e_1^0 ... e_1^(M-1) ] [ k0 ]
[ ... ] = [ ... ... ... ] * [ ... ]
[ f(e_m) ] [ e_m^0 ... e_M^(M-1) ] [ k_(M-1) ]

Equation [8] only represents a system of linear equations if each
equation is actually unique. Since some matrices have eigenvalues which
repeat (i.e. multiplicity > 1), then [8] is generally not a set of M
distinct linear equations.

Luckily, we can make [8] a distinct set of linear equations by
replacing all the duplicate linear equations with the derivatives of
the original linear equation. More specifically, if the subset of
eigenvalues C=(e_i, e_(i+1),...) are all the same, then for all 0 < j <
|C|, replace the (i+j)'th linear equation with the j'th derivative of
the i'th linear equation.

For example, suppose that for some parameter matrix all the eigenvalues
are the same. In that case, equation [8] becomes

[9] [ f(e_1) ] [ e_1^0 ... e_1^(M-1) ] [ k0 ]
[ ... ] = [ ... ... ... ] * [ ... ]
[ f^m(e_1) ] [ 0 ... 1 ] [ k_(M-1) ]

In some cases [8] will have only one eigenvalue with multiplicity
greater than 1, in other cases there may be multiple eigenvalues with
multiplicity greater than 1. The objective is to construct a set of
unique linear equations from the eigenvalues.

Let the square matrix containing the eigenvalues, reduced if necessary
to this unique linear equation form, be denoted B and the corresponding
column vector containing the scalar function of the eigenvalues be
denoted F. We will denote the column vector containing the constants C.

We restate the system of linear equations as

[10] F = B C

Now, by use of Cramer's rule we solve for the constants k_n with

[11] k_n = det(B_n) / det(B)

where B_n is the matrix formed by replacing the n'th column of B with

Having solved for constants k_n in terms of f(e), we proceed to solve
for matrix function by substitution of [11] into [6], giving us

[12] f(A) = SUM(n=0..M-1) A^n det(B_n)/det(B)


There is a theorem which tells that every square matrix A is similar to
a matrix J in Jordan-Canonical form. Thus, if we restate A as

[13] A = M J M^(-1)

where M is a modal matrix for A and J is in Jordan Canonical Form, then
we can use another theorem which tells us that a function of matrix
simplifies to

[14] f(A) = M f(J) M^(-1)

If J is in diagonal form, then

[15] f(J) = [ f(e_1) 0 ... 0 ]
[ 0 f(e_2) ... 0 ]
[ ... ... ... ... ]
[ 0 0 ... f(e_M)]

Otherwise, if J is a Jordan-block then

[16] f(J) = [ f(e_1)/0! f^1(e_1)/1! ... f^(M-1)(e_1)/(M-1)! ]
[ 0 f(e_2)/0! ... f^(M-2)(e_2)/(M-2)! ]
[ ... ... ... ... ]
[ 0 0 0 ... f^M(e_M)/0! ]

Although this alternative approach appears simpler, it is generally
undesirable as the decomposition [13] is computationally-expensive.
Equation [12] is thus offered as a general formula for the evaluation
of an arbitrary function of an MxM matrix.

Reply With Quote
Old 11-01-2006, 12:31 AM
Bill Hobba
Posts: n/a
Default Functions of Matrices

<[Only registered users see links. ]> wrote in message
news:1162302217.174422.11920@m7g2000cwm.googlegrou ps.com...

Why bother since it can be found in virtually any book on linear algebra?
The answer is probably obvious since you posted it before - you can't help


Reply With Quote

functions , matrices

Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Trackbacks are On
Pingbacks are On
Refbacks are On

Forum Jump

Similar Threads
Thread Thread Starter Forum Replies Last Post
Complex Elementary (Transcendental) Functions: f(z) = 1/z Dan Forum Physik 3 10-29-2008 05:52 PM
Complex Elementary Transcendental Functions: f(z) = |sin(z)| Dan Forum Physik 2 10-26-2008 09:58 AM
Electron Pilot Wave Hyper Functions Consc Physics Forum 6 01-03-2005 04:17 AM
Mechanism of sensing Wave Functions Consc Physics Forum 2 12-28-2004 08:33 PM
Discover gene functions by analyzing thousands microarrays (data included) Chang Zhu Protocols and Methods Forum 0 01-07-2004 04:38 AM

All times are GMT. The time now is 07:05 AM.

Powered by vBulletin® Version 3.8.4
Copyright ©2000 - 2015, Jelsoft Enterprises Ltd.
Copyright 2005 - 2012 Molecular Station | All Rights Reserved
Page generated in 0.13750 seconds with 16 queries