The Jacobian (named after German mathematician Karl Gustav Jacobi, 1804–1851) is generally used in conjunction with partial derivatives to provide an easy test for the existence of functional dependence (linear and nonlinear). A Jacobian determinant ∣J∣ is composed of all the first‑order partial derivatives of a system of equations, arranged in order sequence. Given
Given that the first‑order conditions zx=zy=0 are met, a sufficient condition for a multivariate function z=f(x,y) to be an optimum is
zxx,zyy>0 for a minimum; zxx,zyy<0 for a maximum.
zxxzyy>(zxy)2.
A convenient test for this second‑order condition is provided by the Hessian. A Hessian ∣H∣ is a determinant composed of all the second‑order partial derivatives, with the second‑order direct partials on the principal diagonal and the second‑order cross partials off the principal diagonal. That is,
then the second‑order conditions for a minimum are set. That is, when ∣H1∣>0 and ∣H2∣>0, the Hessian ∣H∣ is said to be positive definite. And a positive definite Hessian fulfills the second‑order conditions for a minimum.
If however the first principal minor ∣H1∣<0, while the second principal minor remains positive, then the second‑order conditions for a maximum are met. That is, when ∣H1∣<0 and ∣H2∣>0, the Hessian ∣H∣ is said to be negative definite and fulfills the second‑order conditions for a maximum.
We have ∣H1∣=4>0 and ∣H2∣=(4)(4)−(−1)2=16−1=15>0, i.e. both princpal minnores are positive and hence the Hessian is said to be positive definite and the function z is characterized by a minimum at the critical values (can you find these?).
Determinants can be used to test for positive and negative definiteness of any quadratic form. The determinant of a quadratic form is called a discriminant∣D∣. Given the quadratic form
the determinant is formed by placing the coefficients of the squared terms on the principal diagonal and dividing the coefficients of the non‑squared terms equally between the off‑diagonal positions. Hence, we have
If ∣D1∣,∣D2∣>0, ∣D∣ is positive definite and z is positive for all nonzero values of x and y. If ∣D1∣<0 and ∣D2∣>0, ∣D∣ is negative definite and z is negative for all nonzero values of x and y. If ∣D2∣=0, z is not sign definite and z may assume both positive and negative values.
Let’s take an example to test for sign definiteness of the following quadratic form
z=2x2+5xy+8y2.
We can easily form the discriminant
∣D∣=∣∣22.52.58∣∣
Then evaluating the principal minors gives
∣D1∣=2>0,∣D2∣=∣∣22.52.58∣∣=16−6.25=9.75>0.
Hence, z is positive definite, meaning that it will be greater than zero for all nonzero values of x and y.
Try these ^^
By inspecting its Discriminant, can you tell whether the function below is positive or negative for all nonzero x and y?
A quadratic form is defined as a polynomial expression in which each component term has a uniform degree.
Here are some examples:
6x2−2xy+3y2is a quadratic form in 2 variables.x2+2xy+4xz+2yz+y2+z2is a quadratic form in 3 variables.
It would be useful to determine the sign definiteness so we can make statements concerning the optimum value of the function as to whether it is a minimum or a maximum.
More generally, a quadratic form in n variables (x1,x2,…,xn) can be written as x′Ax where
x′ is a row vector [x1,x2,…,xn] and A is an n×n matrix of scalar elements.
Since both squared terms have positive coefficients, Q(x,y)>0 for all (x,y)=(0,0), the quadratic form is positive definite.
Quadratic forms are used in many areas such as:
Optimization: To determine if a critical point is a minimum, maximum, or saddle point (via the Hessian matrix).
Geometry: Describing conic sections (ellipses, hyperbolas) and quadric surfaces.
Physics: Representing energy functions (e.g., kinetic energy in mechanics).
Quadratic forms are particularly useful in determining the concavity or convexity of a differentiable function.
For a function y=f(x1,x2,…,xn), we can form the Hessian consisting of second‑order partial derivatives and construct a quadratic form as x′Hx. Then
a) The function y is strictly convex if the quadratic form is positive (implying that the Hessian is positive definite, i.e., the determinants of the principal minors are all positive). b) The function y is strictly concave if the quadratic form is negative (that is, the Hessian is negative definite, i.e. the determinants of the principal minors alternate in sign).
Let’s take an example to see this more clearly.
Say we have
z=x2+y2.
Then
zx=2x,zxx=2,zxy=0,zy=2y,zyx=0,zyy=2.
But
x′Hx=zxxx2+zyyy2+zxyzyxxy=2x2+2y2>0.
Hence the function z=x2+y2 is characterized by a minimum at its optimal value and is therefore a strictly convex function. And it is easy to see that the Hessian is positive definite (i.e. the value of the function is always positive for any non-zero values of x and y).
Since the principal minors alternate correctly in sign, the Hessian is negative definite and the function is maximized at x1≈1.04, x2≈1.22 and x3≈0.43.
Try these ^^
For each equation below find (a) critical values, and (b) the nature of the critical values using the Hessian.
To optimize a function f(x1,x2) subject to a constraint g(x1,x2) the first‑order conditions can be found by setting up what is known as the Lagrangian function F(x1,x2,λ)=f(x1,x2)−λ(g(x1,x2)−c).
The second‑order conditions can be expressed in terms of a bordered Hessian ∣Hˉ∣ as
which is the usual Hessian bordered by the first derivatives of the constraint with zero on the principal diagonal.
The order of a bordered principal minor is determined by the order of the principal minor being bordered. Hence ∣Hˉ2∣ above represents a second bordered principal minor, because the principal minor being bordered has dimensions 2×2.
For the bivariate case with a single constraint, we simply look at ∣Hˉ2∣. If this is negative, the bordered Hessian is said to be positive definite and satisfies the second‑order condition for a minimum. However, if it is positive, the bordered Hessian is said to be negative definite, and meets the sufficient conditions for a maximum.
Let’s try optimizing the following objective function
f(x1,x2)=4x12+3x22−2x1x2
subject to
x1+x2=56.
Setting up the Lagrangian function, taking the first‑order partials and solving gives x1∗=36, x2∗=20 (and λ=348). (You should verify this.)
The bordered Hessian for this optimization problem is
which is negative, hence ∣Hˉ∣ is positive definite and we have met sufficient conditions for a minimum.
For the more general case in which the objective function has say n variables, i.e., f(x1,…,xn) which is subject to some constraint g(x1,…,xn), we can set up the bordered Hessian as
where ∣Hˉ∣=∣Hˉn∣ because the n×n principal minor is bordered.
In this case, if all the principal minors are negative, i.e. ∣Hˉ2∣,∣Hˉ3∣,…,∣Hˉn∣<0, the bordered Hessian is said to be positive definite and satisfies the second‑order condition for a minimum.
On the other hand, if the principal minors alternate consistently in sign from positive to negative, i.e. ∣Hˉ2∣>0,∣Hˉ3∣<0,∣Hˉ4∣>0 etc., the bordered Hessian is negative definite, and meets the sufficient conditions for a maximum.
If aij is a technical coefficient representing the value of input i required to produce one dollar’s worth of product j, the total demand for good i can be expressed as
where bi is the final demand for product i. What is important to realize here is that the total demand for a product consists of that product being the final demand plus that product being an intermediate good required for the production of other products.
Characteristic Roots and Vectors (Eigenvalues and Eigenvectors)¶
The sign and definiteness of a Hessian and a quadratic form has been tested by using the principal minors. Sign definiteness can also be tested by using the characteristic roots of a matrix. Given a square matrix A, is possible to find a vector V=0 and a scalar λ such that
the scalar λ is called the characteristic root, latent value or eigenvalue; and the vector V is called the characteristic vector, latent vector or eigenvector. The above can be written as
where A−λI is called the characteristic matrix of A. Since we have V=0, the characteristic matrix A−λI must be singular and thus its determinant is zero.
With ∣A−λI∣=0, there will be an infinite number of solutions for V. To force a unique solution, the solution may be normalized by requiring of the elements vi of V such that ∑vi2=1.
Let’s take an example. Given a square matrix
A=[−633−6]
To find the characteristic roots (eigenvalues) of A, we simply set ∣A−λI∣=0:
∣A−λI∣=∣∣−6−λ33−6−λ∣∣=0
This means
(−6−λ)(−6−λ)−9=0λ2+12λ+27=0(λ+9)(λ+3)=0λ=−9,−3.
Since both characteristic roots λ are negative, we say A is negative definite.
Since the coefficient matrix is linearly dependent, there are infinite number of solutions. The product of the matrices gives two equations which are identical:
3v1+3v2=0⇒v2=−v1.
By normalizing we have
v12+v22=1.
Substituting v2=−v1 gives
v12+(−v1)2=1⇒2v12=1⇒v12=21.
Taking the positive square root gives v1=1/2=22 and substituting into v2=−v1 gives v2=−22. That is
D is a diagonal matrix (entries only on the main diagonal) consisting of eignevalues of A,
T is an invertible matrix whose columns are eigenvectors of A.
Note that not all matrices are diagonalizable, however.
For example, given
A=[2112]
Let us start by finding its eigenvalues.
The characteristic polynomial is given by
det(A−λI)=(2−λ)2−1=λ2−4λ+3=0.
Hence,
λ=1,λ=3.
Forλ=1:
(A−I)v=(1111)v=0.
A corresponding eigenvector is
v1=(1−1).
Forλ=3:
(A−3I)v=(−111−1)v=0.
A corresponding eigenvector is
v2=(11).
Then,
T=(1−111),D=(1003).
The inverse of T is
T−1=21(11−11).
We can verify that A=TDT−1.
TDT−1=(1−111)(1003)⋅21(11−11)=(2112)=A.
Note (1):
All symmetric matrices are diagonalizable (even with repeated eigenvalues).
Diagonalizable = invertible. For example,
[1000]
is diagonalizable but not invertible.
If A is diagonalizable, then An and eA are easy to compute.
Note (2): This is closely related to the transformation form of diagonalization, which states that if A is diagonalizable, then there exists an invertible matrix T and a diagonal matrix D such that
T−1AT=D.
Exercise. Can you show this?
Try these ^^
Given:
A=[−4−2−2−6],B=[3102]
C=[633−2],D=⎣⎡400621353⎦⎤
Find:
(a) Find the eigenvalues and eigenvectors for each of the matrices above (b) What can you say about the sign definiteness of each mattix (c) Verify A=TDT−1 (d) Find A5