Chapter 2 Optimization Math Review, ECON 510 See Syds?ter (2005, Chapters 2, 3) and Chiang (1984, Chapters 9, 11, 12 and 21). 1. Positive Definite Matrix Definite matrices are directly related to optimization. A symmetric matrix A is positive semi-definite (A ≥ 0) if x 0 Ax ≥ 0, ? x; positive definite (A>0) if x 0 Ax > 0, ? x null=0; negative semi-definite (A ≤ 0) if x 0 Ax ≤ 0, ? x; negative definite (A<0) if x 0 Ax < 0, ? x null=0. Example 2.1. A = ? ? ? 1 ?1 ?11 ? ? ? ≥ 0.  Example 2.2. Consider A = ? ? ? ab bc ? ? ? .  Example 2.3. For a symmetric matrix A, if A ≥ 0, then a ii ≥ 0 for all i; if A ≤ 0, then a ii ≤ 0 for all i.  2—1 Given a matrix A =(a ij ) n×n , for i 1 ,···,i k ∈ {1,2,···,n} with i 1 <i 2 < ···<i k , define a k ×k minor as d {i 1 ,···i k } ≡ null null null null null null null null null null null null null null null a i 1 i 1 a i 1 i 2 ··· a i 1 i k a i 2 i 1 a i 2 i 2 a i 2 i k . . . . . . . . . a i k i 1 a i k i 1 ··· a i k i k null null null null null null null null null null null null null null null . In particular, denote theprincipal minorsas d 1 ≡ d {1} ,d 2 ≡ d {1,2} ,..., d n ≡ d {1,2,...,n} . Theorem 2.1. For a symmetric matrix A, (a) A>0 ?? d k > 0, for all k. (b) A<0 ?? (?1) k d k > 0, for all k. (c) A ≥ 0 ?? d {i 1 ,···i k } ≥ 0, for all permutations {i 1 ,...,i k } and all k =1,...,n. (d) A ≤ 0 ?? (?1) k d {i 1 ,···i k } ≥ 0, for all permutations {i 1 ,...,i k } and all k =1,...,n.  Example 2.4. (a) Verify ? ? ? 1 ?1 ?11 ? ? ? ≥ 0. (b) Find conditions for A = ? ? ? ab bc ? ? ? ≥ 0.  Proposition 2.1. ? A ≥ 0 i? all its eigenvalues are positive. ? A>0 i? all its eigenvalues are strictly positive. ? If A>0, then A ?1 > 0. ? A ≥ 0 ?? A ≤ 0. 2—2 ? A>0 ?? A<0. ? If A is a tall full-rank matrix, then A 0 A>0 and AA 0 ≥ 0. ? If A>0 and |B| null=0, then B 0 AB > 0.  2. Concavity Given points x k ∈ R n , a convex combination is y ≡ λ 1 x 1 + ···+ λ m x m , λ k ≥ 0, m null k=1 λ k =1. Given any two points x, y, the intervals are [x, y] ≡ null z | z = λx +(1?λ)y, λ ∈ [0, 1] null , (x, y) ≡ null z | z = λx +(1?λ)y, λ ∈ (0, 1) null . A set S ?R n is a convex set if ? x, y ∈ S, [x, y] ? S. Proposition 2.2. (Properties of Convex Sets). 1. Any intersection of convex sets is also convex. 2. The Cartesian product of convex sets is also convex.  f : X → R is concave if X ?R n is convex and f[λx +(1?λ)y] ≥ λf(x)+(1?λ)f(y), ? λ ∈ (0, 1),x,y∈ X. If the inequality holds strictly, f is strictly concave. f is (strictly) convex if ?f is (strictly) concave. Theorem 2.2. (Properties of Concave Functions). X ? R n is convex. 1. f : X → R is concave i? {(x,t) ∈ X ×R | f(x) ≥ t} is convex. 2. Concave functions are continuous in the interior of their domains. 3. A function f : X → R is concave i? f(λ 1 x 1 + ···+ λ k x k ) ≥ λ 1 f(x 1 )+···+ λ k f(x k ), for all k ≥ 1 and all convex combinations λ 1 x 1 + ···+ λ k x k .  2—3 Theorem 2.3. Given convex X ? R n , twice di?erentiable f : X → R, 1. f is convex ?? D 2 f(x) ≥ 0, ? x ∈ X. 2. D 2 f(x) > 0, ? x ∈ X =? f is strictly convex. 3. f is concave ?? D 2 f(x) ≤ 0, ? x ∈ X. 4. D 2 f(x) < 0, ? x ∈ X =? f is strictly concave.  Corollary 2.1. Given convex X ? R n , twice di?erentiable f : X → R, and let d {i 1 ,···,i k } (x) be a k ×k principal minors of D 2 f(x), 1. f is convex ?? d {i 1 ,···,i k } (x) ≥ 0, ? x ∈ X, ? k, ?{i 1 ,···,i k }. 2. f is concave ?? (?1) k d {i 1 ,···,i k } (x) ≥ 0, ? x ∈ X, ? k, ?{i 1 ,···,i k }. 3. d k (x) > 0, ? x ∈ X, ? k. =? f is strictly convex. 4. (?1) k d k (x) > 0, ? x ∈ X, ? k. =? f is strictly concave.  Example 2.5. “f is strictly convex”>“D 2 f>0”. E.g., f(x)=x 4 .  Example 2.6. For f(x,y)=x α + y β , defined on R 2 ++ , where α, β ≥ 0, f is ? ? ? ? ? concave, if 0 ≤ α, β ≤ 1; strictly concave, if 0 < α, β < 1. Example 2.7. For Cobb-Douglas function f(x,y)=x α y β , defined on R 2 ++ , where α, β ≥ 0, f is ? ? ? ? ? concave, if α, β ≥ 0, α + β ≤ 1; strictly concave, if α, β > 0, α + β < 1.  Proposition 2.3. Let f : X → R be di?erentiable. Then, (1) f is concave ?? Df(x) · (y ?x) ≥ f(y)?f(x), for all x, y ∈ X. (2) f is strictly concave ?? Df(x) · (y ?x) >f(y) ?f(x), for all x, y ∈ X, x null= y.  2—4 3. Quasi-Concavity Given a convex set X ? R n ,f: X → R is quasi-concave if, ? x, y ∈ X, z ∈ (x, y), f(y) ≥ f(x) ? f(z) ≥ f(x). f is strictly quasi-concave if, ? x, y ∈ X, z ∈ (x, y), f(y) ≥ f(x) ? f(z) >f(x). f is (strictly) quasi-convex if ?f is (strictly) quasi-concave. Theorem 2.4. Given convex set X ?R n , 1. f : X → R is quasi-concave i? the upper level set {x ∈ X | f(x) ≥ t} is convex, ? t ∈R. 2. f : X → R is quasi-convex i? the lower level set {x ∈ X | f(x) ≤ t} is convex, ? t ∈R.  Set f ?1 (t)={x ∈ X | f(x)=t} is called a level set. Theorem 2.5. (a) Concave functions are quasi-concave; convex functions are quasi-convex. (b) Strictly concave functions are strictly quasi-concave. Strictly convex functions are strictly quasi-convex. (c) Monotonic functions defined on R are both quasi-concave and quasi-convex.  Quasi-concavity doesn’t necessarily imply concavity. strict concavity =? concavity ?? strict quasi-concavity =? quasi-concavity 2—5 Given f : X → R, define the bordered Hessian matrix B f (x): f i ≡ ?f(x) ?x i ,f ij ≡ ? 2 f(x) ?x i ?x j ,B f (x) ≡ ? ? ? ? ? ? ? ? ? ? ? 0 f 1 ··· f n f 1 f 11 ··· f 1n . . . . . . . . . f n f n1 ··· f nn ? ? ? ? ? ? ? ? ? ? ? . Theorem 2.6. Given convex X ?R n , twice di?erentiable f : X → R, and the principal minors b 1 (x),...,b n+1 (x) of B f (x), 1. For X ?R n + ,fis quasi-convex =? b k (x) ≤ 0, ? x ∈ X, ? k. 2. For X ?R n + ,fis quasi-concave =? (?1) k b k (x) ≤ 0, ? x ∈ X, ? k. 3. For X = R n + or R n ,b k (x) < 0, ? x ∈ X, ? k ≥ 2=? f is strictly quasi-convex. 4. For X = R n + or R n , (?1) k b k (x) < 0, ? x ∈ X, ? k ≥ 2=? f is strictly quasi-concave.  Example 2.8. For f(x,y)=x α + y β , defined on R 2 ++ , where α, β ≥ 0, f is ? ? ? ? ? quasi-concave, if 0 ≤ α, β ≤ 1; strictly quasi-concave, if 0 < α, β ≤ 1 and α null=1or β null=1.  Example 2.9. For Cobb-Douglas function f(x,y)=x α y β , defined on R 2 ++ , where α, β ≥ 0, f is ? ? ? ? ? quasi-concave, if α, β ≥ 0; strictly quasi-concave, if α, β > 0.  2—6 4. Unconstrained Optimization See Syds?ter (2005, Chapter 3) and Chiang (1984, Chapter 9). If there is a neighborhood N r (x ? ) of x ? with radius r such that x ? is the maximum point of f on N r (x ? ), then x ? is a local maximum point of f. For f : R n → R, if Df(?x)=0, call ?x or (?x,f(?x)) a stationary point,andf(?x) the stationary value. Given a stationary point ?x, there are three possible situations at ?x : local maximum, local minimum point,andareflection point. Example 2.10. Compare y = x 2 with y = x 3 at x =0.  Theorem 2.7. (Extreme-Value Theorem). For continuous f : R n → R andcompact A ?R n , max x∈A f(x) has at least one solution.  Theorem 2.8. Let A ? R n . (a) If x ? is an interior solution of max x∈A f(x) then (FOC) Df(x ? )=0and (SONC) D 2 f(x ? ) ≤ 0. (b) If Df(x ? )=0and (SOSC) D 2 f(x ? ) < 0, ? N r (x ? ) s.t. x ? is the maximum point of f on N r (x ? ). (c) If f is concave on A, any point x ? ∈ A satisfying Df(x ? )=0is a maximum point. (d) If f is strictly quasi-concave, a local maximum over a convex set A is the unique global maximum.  Note: the FOC and SONC are not necessary for corner solutions; they are also not su?cient for local maximization, even for interior points. Example 2.11. Find a maximum point for f(x 1 ,x 2 )=x 2 ?4x 2 1 +3x 1 x 2 ?x 2 2 .  2—7 5. Constrained Optimization See Syds?ter (2005, Chapter 3) and Chiang (1984, Chapters 12 and 21). Theorem 2.9. (Lagrange). For f : R n → R,G: R n → R m , consider problem max x∈R n f(x) s.t. G(x)=0. Let L(λ,x) ≡ f(x)+λ ·G(x) (Lagrange function). ? If x ? is a solution and if DG(x ? ) has full rank, then ? λ ∈R m (Lagrange multi- plier)s.t. FOC : D x L(λ,x ? )=0, SONC : h 0 D 2 x f(x ? )h ≤ 0, ? h satisfying DG i (x ? )h =0, ? i. ? If the FOC is satisfied, G(x ? )=0,Gis quasi-concave, and SOSC: h 0 D 2 f(x ? )h<0, for h null=0satisfying DG(x ? )h =0, then we have a unique local maximum.  Example 2.12. For a>0 and b>0, consider F(a,b) ≡ max x 1 ,x 2 ?ax 2 1 ?bx 2 2 s.t. x 1 + x 2 =1.  Theorem 2.10. Let A ∈ R n×n be symmetric, C ∈ R m×n has full rank, m<n,and b 1 ,...,b m+n be the principal minors of B ≡ ? ? ? 0 C C T A ? ? ? . Then, 1. x 0 Ax > 0 for x null=0 satisfying Cx=0 ?? (?1) m b k > 0, ? k ≥ 2m +1. 2. x 0 Ax < 0 for x null=0 satisfying Cx=0 ?? (?1) m+k b k > 0, ? k ≥ 2m +1.  2—8 Discussions: (1) What is the intuition for the FOC? (2) What is the intuition for the SOC? (3) How to verify the SOC? (4) Is the SOC related to quasi-concavity? Example 2.13. Verify the SOC for Example 2.12.  Example 2.14. Verify the SOC for Example 2.12.  Theorem 2.11. (Kuhn-Tucker). For di?erentiable f : R n → R and G : R n → R m , consider problem V ≡ max x∈R n f(x) (2.1) s.t. G(x) ≥ 0. Let L(x,λ) ≡ f(x)+λ·G(x). If x ? is a solution and Dg j (x ? ),i=1,...,m, j=1,...,k, corresponding to binding constraints are linearly independent, then there exists λ ∈ R m + such that FOC: D x L(x ? ,λ)=0, Kuhn-Tucker condition: λ ·G(x ? )=0.  Theorem 2.12. Suppose f, g i ,h j : R n → R are C 1 functions, i =1,...,m, j = 1,...,k, and k<n.Let G ≡ (g 1 ,...,g m ) T and H ≡ (h 1 ,...,h k ) T . If x ? is a solution of the following problem V ≡ max x∈R n f(x) (2.2) s.t. G(x) ≥ 0, H(x)=0, and the vectors Dg i (x ? ) and Dh j (x ? ),i=1,...,m, j =1,...,k, corresponding to binding constraints are linearly independent, then there exists a unique λ ∈ R m + and μ ∈R k such that (a) Df(x ? )+λ ·G(x ? )+μ ·H(x ? )=0; (b) λ·G(x ? )=0.  2—9 Theorem 2.13. (global maximum). Let f and g i ,i=1,...,m, be quasi-concave, where G =(g 1 ,...,g m ) T . Let x ? satisfy the Kuhn-Tucker condition and the FOC for (2.1). Then, x ? is a global maximum point if (1) Df(x ? ) null=0, and f is locally twice continuously di?erentiable, or (2) f is concave.  H : R n → R k is linear if H(x)=Ax + B, where A is matrix and B is a vector. Proposition 2.4. Let f : R n → R and G : R n → R k be concave, and H : R n → R m be linear. Also, ? x 0 s.t. G(x 0 ) >> 0. Then, x ? is a solution of max x f(x) s.t. G(x) ≥ 0 H(x)=0 i? G(x ? ) ≥ 0,H(x ? )=0, and there exist λ ∈ R m + and μ ∈ R k such that λ ·G(x ? )=0 and x ? is a solution of max x∈R n L(x,λ,μ) ≡ f(x)+λ ·G(x)+μ·H(x).  Example 2.15. For a ∈ (0, 1), consider v(p 1 ,p 2 ,m) ≡ max x 1 ,x 2 ≥0 Ax a 1 x 1?a 2 s.t. p 1 x 1 + p 2 x 2 ≤ m.  6. Envelope Theorem Theorem 2.14. (Envelope). For di?erentiable f : X × A → R,X? R n ,A? R k , and x ? (a) being an interior maximum point of F(a) ≡ max x∈X f(x,a), we have dF(a) da = ?f(x,a) ?a null null null null x=x ? (a) .  2—10 Theorem 2.15. Suppose f, g i ,h j : R n × R l → R are C 1 functions, i =1,...,m, j =1,...,k. Let G ≡ (g 1 ,...,g m ) T ,H≡ (h 1 ,...,h k ) T , and L(x,a,λ,μ) ≡ f(x,a)+λ ·G(x,a)+μ·H(x,a). If x ? (a) is a solution of the following problem F(a) ≡ max x∈R n f(x,a) s.t. G(x,a) ≥ 0, H(x,a)=0, and λ ? (a) and μ ? (a) are the corresponding Lagrange multipliers, then under some con- ditions 1 we have ?F(a) ?a = ?L(x,a,λ,μ) ?a null null null null x=x ? (a), λ=λ ? (a), μ=μ ? (a) .  (2.3) Example 2.16. Consider F(a,b) in Example 2.12.  Example 2.17. Consider the following economic problem: v(p 1 ,...,p n ,I) ≡ max x≥0 u(x 1 ,...,x n ) s.t. p 1 x 1 + ···+ p n x n ≤ I.  1 See Sydsaeter et al (2005, p.149). 2—11 Chapter 3 Dynamic Optimization Math Review, ECON 510 1. Discrete-Time Stochastic Model Good references: Sydsaeter et al (2005, Chapter 12), Stokey—Lucas (1989, p.239—259), Sargent (1987, Chapter 1). 1.1. Markov Process What is a Markov process? At t, we know x 0 ,...,x t and some knowledge of x t+1 : x t+1 ~ F t (·|x t ,x t?1 ,...,x 0 ). We call {x t } ∞ t=0 a random process. If the distribution function is time-independent, i.e., x t+1 ~ F(·|x t ,x t?1 ,...,x 0 ), for all t, we call {x t } ∞ t=0 a stationary process. If the dependence on past history has a fixed length of time, i.e., there exists an integer n such that x t+1 ~ F(·|x t ,...,x t?n+1 ), for all t, we call {x t } ∞ t=0 an nth-order Markov process. For example, a first-order Markov process {x t } ∞ t=0 is defined by x t+1 ~ F(·|x t ), for all t. 3—1 1.2. Bellman Equation Let x t ∈R n ,u t ∈R k ,f t : R n ×R k → R, and g t : R n ×R k ×R m → R n . Consider V 0 (x 0 ) ≡ max u 0 ,u 1, ... E 0 ∞ null t=0 β t f t (x t ,u t ) s.t. x t+1 = g t (x t ,u t ,ε t+1 ), x 0 is given and known, (3.1) where 0 < β < 1, null t ∈ R m is a random vector unknown until period t, E t is the expectation operator conditional on period t information Φ t . Here u t is the control and x t is the state. Example 3.1. Consider a consumer who, at each time t, consumes c t units of goods and has assets A t , with an endowment of A 0 . The consumer may save at a gross return R t . Theconsumer’sproblemis V 0 (A 0 ) ≡ max c 0 ,c 1, ... E 0 ∞ null t=0 β t u(c t ) s.t. A t+1 = R t+1 (A t ?c t ), given A 0 . Example 3.2. Consider a firm’s problem: max {l t ,s t } E 0 ∞ null t=0 null 1 1+r null t (y t ?w t l t ?s t ) s.t. y t+1 = f(k t ,l t ,ε t+1 ), k t+1 =(1?δ)k t + s t , where s t is saving, l t is labor input, k t is capital stock, w t is wage rate, and r is the interest rate.  Let V t (x t ) ≡ max u t ,u t+1, ... E t ∞ null s=t β s?t f s (x s ,u s ) s.t. x s+1 = g s (x s ,u s ,ε s+1 ), x t is given and known. (3.2) 3—2 The solution of (1.1) is the solution of the Bellman equation: V t (x t ) ≡ max u t f t (x t ,u t )+βE t V t+1 (x t+1 ) (3.3) s.t. x t+1 = g t (x t ,u t ,ε t+1 ). Conversely, the solution {u ? t ,x ? t } of (3.3) is the solution of (1.1) if the transversality condition is satisfied: lim t→∞ β t E t V t+1 (x ? t+1 )=0. (3.4) Problem (3.3) imply two equations: f t,u (x t ,u t )+βE t null V 0 t+1 (x t+1 )g t,u (x t ,u t ,ε t+1 ) null =0, (3.5) V 0 t (x t )=f t,x (x t ,u t )+βE t null V 0 t+1 (x t+1 )g t,x (x t ,u t ,ε t+1 ) null . (3.6) We look for a solution of the form: u ? t = h t (x t ). By (2.15) and (2.2), we can generally find the two functions h t and V t . Theorem 3.1. If f t = f, g t = g and {ε t } is a Markov process, then V t is time invariant.  By Theorem 3.1, a solution (h t ,V t ) from (2.15) and (2.2) will be time invariant: u ? t = h(x t ) and V t (x t )=V (x t ). Example 3.3. Solve the consumer’s problem in Example 3.1. Assuming {R t } to be a first-order Markov process. The Euler equation is: u 0 (c t )=βE t [u 0 (c t+1 )R t+1 ], (3.7) If u(c)=lnc, the optimal consumption has the feedback form: c t =(1?β)A t . Example 3.4. Solve the firm’s problem in Example 3.2. 1.3. Lagrange Method Theorem 3.2. (Kuhn-Tucker). For di?erentiable f : R n → R and G : R n → R m , let L(λ,x) ≡ f(x)+λ ·G(x). If x ? is a solution of max x f(x) s.t. G(x) ≥ 0, then there exists λ ∈R m + such that (Kuhn-Tucker condition) λ ·G(x ? )=0 and FOC: D x L(λ,x ? )=0.  3—3 The Lagrange method can solve more general problems than (1.1). In particular, the Lagrange function for (3.2) is L = E t ∞ null s=t β s?t {f s (x s ,u s )+λ s+1 · [g s (x s ,u s ,null s+1 )?x s+1 ]}. By the Kuhn-Tucker Theorem, we can again derive the two equations (3.5) and (3.6). References [1] Chiang, A.C. (1984): Fundamental Methods of Mathematical Economics,McGraw- Hill. [2] Chiang, A.C. (1992): Elements of Dynamic Optimization, McGraw-Hill. [3] Kamien, M.I. & N.L. Schwartz (1991): Dynamic Optimization, North Holland. [4] Sargent, T.J. (1987): Dynamic Macroeconomic Theory, Harvard University Press. [5] Stokey, N.L. & R.E. Lucas (1989): Recursive Methods in Economic Dynamics, Har- vard University Press. 2. Continuous-Time Deterministic Model See Kamien—Schwartz (1991) and Chiang (1992). Example 3.5. The Shortest Path. What is the shortest path to move from one (x 0 ,y 0 ) to another (x 1 ,y 1 ) in R 2 ?  Theorem 3.3. Suppose H : R×R k ×R k → R is continuous w.r.t. its first argument, continuously di?erentiable w.r.t. its second and third arguments. Let A ≡ {continuously di?erentiable functions u :[t 0 ,T] → R k } be the set of admissible controls. Then, the solution u ? of max u∈A null T t 0 H[t, u(t), ˙u(t)]dt s.t. u(t 0 )=u 0 ,u(T)=u T (3.8) 3—4 must satisfy the Euler equation: d dt H ˙u [t, u ? (t), ˙u ? (t)] = H u [t, u ? (t), ˙u ? (t)],t∈ (0,T). (3.9) If the terminal value u(T) is free, the transversality condition is H ˙u [T, u ? (T), ˙u ? (T)] = 0. (3.10) If the initial value u(t 0 ) is free, the transversality condition is H ˙u [t 0 ,u ? (t 0 ), ˙u ? (t 0 )] = 0. (3.11) If the terminal condition is u(T) ≥ 0, the transversality conditions are u ? (T)H ˙u [T, u ? (T), ˙u ? (T)] = 0,H ˙u [T, u ? (T), ˙u ? (T)] ≤ 0. (3.12) Conversely, if H(t,u, ˙u) is concave in (u, ˙u), then any u ? ∈ A satisfying (3.9) and the initial and terminal conditions is a solution of (3.8).  Here, T can be either finite or infinite. By Kamien—Schwartz (1991, p.43), the Legendre condition is H ˙u ˙u [t, u ? (t), ˙u ? (t)] ≤ 0. It is a second order necessary condition, but not su?cient even locally. Two boundary conditions are needed to pin down the two arbitrary constants in the general solution. When a boundary condition is missing, a transversality condition replaces it. Example 3.6. Solve the following agency model: max a,s(·) null v[y ?s(y)]f(y,a)dy (3.13) s.t. null u[s(y)]f(y,a)dy ≥ c(a)+ˉu.  Example 3.7. Characterize the solution of the following model in asymmetric informa- tion: max x(·) null ˉ θ θ {u[x(θ),θ]+v[x(θ),θ]}dF (θ) s.t. ˙x(θ) ≥ 0.  3—5 Theorem 3.4. Let x ∈ R n ,u∈ R k ,g: R n ×R k ×R→ R n ,f: R n ×R k ×R→ R. For problem J(x 0 ,x T ,t 0 ) ≡ max u null T t 0 f[x(t),u(t),t]dt s.t. ˙x(t)=g[x(t),u(t),t] x(t 0 )=x 0 ,x(T) ≥ 0 define the Hamiltonian as H = f(x, u,t)+λ(t) ·g(x,u,t). Under certain di?erentiability conditions, if u ? is a solution, then there exists λ : [t 0 ,T] → R n such that u ? is a solution of H u =0, (3.14) ˙ λ = ?H x , (3.15) with transversality conditions: lim t→T λx =0, λ(T) ≥ 0.  (3.16) Theorem 3.5. Let x ∈ R n ,u∈ R k ,g: R n ×R k ×R→ R n ,f: R n ×R k ×R→ R. For problem J(x 0 ,x T ,t 0 ) ≡ max u null T t 0 f[x(t),u(t)]e ?θ(t?t 0 ) dt s.t. ˙x(t)=g[x(t),u(t),t] x(t 0 )=x 0 ,x(T) ≥ 0 define the Hamiltonian as H = f(x, u)+λ(t) ·g(x,u,t). Under certain di?erentiability conditions, if u ? is a solution, then there exists λ : [t 0 ,T] → R n such that u ? is a solution of H u =0, ˙ λ = θλ?H x , with transversality condition lim t→T λxe ?θt =0, λ(T) ≥ 0.  3—6 Example 3.8. The corresponding continuous-time version of the consumer’s problem in Example 3.1 is J(A 0 ) ≡ max c null ∞ 0 u(c)e ?ρt dt s.t. ˙ A = rA?c A(0) = A 0 ,A(∞ ) ≥ 0. Characterize the optimal consumption path.  Example 3.9. Many other types of constraints can also be handled similarly. For ex- ample, consider the following problem J(x 0 ,x T ,t 0 ) ≡ max u null T t 0 f[x(t),u(t),t]dt s.t. ˙x(t)=g[x(t),u(t),t], null T t 0 h[x(t),u(t),t]dt = c, x(t 0 )=x 0 ,x(T)=x T .  3. Phase Diagram See Chiang (1984, 1992) and Sydsaeter et al (2005, p.244). 3.1. The Linear System A phase diagram is often used to analyze the solution from a dynamic system. From an optimal control problem, the Euler equation and the motion equation generally lead to two equations of the form: ? ? ? ? ? ˙x = f(x,y,t), ˙y = g(x,y,t). Phase diagrams handle the autonomous case with f = f(x,y) and g = g(x,y). 1. For simple f and g, we may be able to solve this equation system directly. 2. For complicated f and g, we may use a phase diagram to illustrate the solution path. 3—7 Example 3.10. In Example 3.8, c(t) and A(t) are determined by ˙ A = rA?c ˙c =(r ?ρ) c α(c) . The General Solution Consider a linear system ˙x = a(x? ˉx)+b(y ? ˉy), ˙y = α(x? ˉx)+β(y ? ˉy), where a, b, α, β ∈R and (ˉx, ˉy) is the steady state. The equation system can be written in a matrix form: ? ? ? ˙x ˙y ? ? ? = A ? ? ? x? ˉx y ? ˉy ? ? ? , (3.17) where A = ? ? ? d 1 a 1 a 2 d 2 ? ? ? . The eigenvalues of A are λ = 1 2 null d 1 + d 1 ? null (d 1 + d 1 ) 2 ?4|A| null , μ = 1 2 null d 1 + d 1 + null (d 1 + d 1 ) 2 ?4|A| null . We have λ ≤ μ. Let ξ and ζ be eigenvectors of λ and μ, respectively. The general solution of (3.17) is ? ? ? x(t) y(t) ? ? ? = ? ? ? ˉx ˉy ? ? ? + c 1 ξe λt + c 2 ζe μt . (3.18) where c 1 and c 2 are arbitrary constants. Stability Analysis Saddle-Point Stable Path with |A| < 0. If |A| < 0, then λ < 0 and μ > 0, implying that the system is saddle-point stable. Convergent sequences must have c 2 =0. 3—8 Thus, the initial point must be on the convergence line: ? ? ? x y ? ? ? = ? ? ? ˉx ˉy ? ? ? + τξ, for τ ∈R. Given the initial point (x 0 ,y 0 ) on the convergence line, (3.18) implies the adjustment path: ? ? ? x(t) y(t) ? ? ? = ? ? ? ˉx ˉy ? ? ? + ? ? ? x 0 ? ˉx y 0 ? ˉy ? ? ? e λt , which converges to (ˉx, ˉy). The non-convergence line is defined by ? ? ? x y ? ? ? = ? ? ? ˉx ˉy ? ? ? + τζ, for τ ∈R. Theorem 3.6. For the linear system ˙x = A(x? ˉx), where x ∈ R 2 and A ∈ R 2×2 , suppose the system is saddle-point stable with |A| < 0. Then, it has a negative eigenvalue λ < 0 and a positive eigenvalue μ > 0. Let ξ and ζ be eigenvectors of λ and μ, respectively. Then, the convergence line is x =ˉx + τξ, τ ∈R, and the non-convergence line is x =ˉx + τζ, τ ∈R. If the initial point x 0 is on the convergence line, the convergent path is x =ˉx +(x 0 ? ˉx)e λt ; and if the initial point x 0 is on the non-convergence line, the non-convergent path is x =ˉx +(x 0 ? ˉx)e μt .  Example 3.11. Draw the phase diagram for the following system: ˙x = ?(x? ˉx)+10(y ? ˉy), ˙y =20(x? ˉx)+3(y ? ˉy).  3—9 Nonstable Path with |A| > 0. If (d 1 + d 1 ) 2 < 4|A|, thetwoeigenvaluesare λ = d 1 + d 1 2 ?iθ, μ = d 1 + d 1 2 + iθ, where θ ≡ null |A|? null d 1 +d 1 2 null 2 . By (3.18), the adjustment path is ? ? ? x(t) y(t) ? ? ? = ? ? ? ˉx ˉy ? ? ? + null ξe λt , ζe μt null (ξ, ζ) ?1 ? ? ? x 0 ? ˉx y 0 ? ˉy ? ? ? . Since e λt = e d 1 +d 1 2 t [cos(θt)?isin(θt)], e μt = e d 1 +d 1 2 t [cos(θt)+isin(θt)], the convergence is totally determined by d 1 + d 1 . 1. If d 1 + d 1 =0, the adjustment path will be a cycle (a vortex with uniform fluctua- tions). 2. If d 1 + d 1 < 0, the adjustment path is convergent but with cycles (damped fluctua- tions). 3. If d 1 + d 1 > 0, the adjustment path is non-convergent and with cycles (explosive fluctuations). If (d 1 + d 1 ) 2 ≥ 4|A|, since |A| > 0, the signs of λ and μ arethesameasthesign of d 1 + d 1 . 1. If d 1 + d 1 > 0, an adjustment path starting from any point is non-stable. 2. If d 1 + d 1 < 0, an adjustment path starting from any point is convergent. Stable Path with |A| =0. If |A| =0, one of the eigenvalues must be zero and the two demarcation lines defined by A ? ? ? x? ˉx y ? ˉy ? ? ? =0 are merged into one. If d 1 + d 1 > 0, a convergent path must start from the convergence line and it stays at the initial point forever. If d 1 + d 1 < 0, any sequence is convergent, but the limit is generally not (ˉx, ˉy). 3—10 3.2. Linearization For a general nonlinear system ? ? ? ? ? ˙x = f(x,y), ˙y = g(x,y), (3.19) we can convert it into a linear system in (3.17) by linearization. Define the steady state by the state at which ˙x =0and ˙y =0, that is, the steady state (ˉx, ˉy) is the point where f(ˉx, ˉy)=0,g(ˉx, ˉy)=0. By Taylor’s expansion formula, (3.19) can be approximated by ? ? ? ? ? ˙x = f x (ˉx, ˉy)(x? ˉx)+f y (ˉx, ˉy)(y ? ˉy), ˙y = g x (ˉx, ˉy)(x? ˉx)+g y (ˉx, ˉy)(y ? ˉy). This is a linear system, which can then be analyzed using the discussions in previous section. 3.3. Adjustment Path For a system, ? ? ? ? ? ˙x = f(x,y) ˙y = g(x,y) when there is a change in f and/or g, the steady state may change. How does the economy then move towards the new steady state? For a system that is not derived by an optimization problem, two assumptions are needed in order to pin down the adjustment path: Assumption 1. One variable cannot jump, but the other can. Assumption 2. Any adjustment path can only jump at the beginning; it must move continuously afterwards. The general solution of thesystemhas two arbitraryconstants. These twoassumptions pin down these two constants. 3—11 For most dynamic optimization problems, these two assumptions are implied by op- timal behaviors. We now show in an example how to use these two assumptions to pin down the adjustment path. The Model Let A be the aggregate demand, defined by A = C + I + G, where C is aggregate demand, I is aggregate investment, and G is government spending. First, assume investment and consumption are a?ected by the long-term real interest rate r. Then, A =[a + b(Y ?T)] + (α?βr)+G, where T is tax and a, b, α and β are constants. Second, assume that the adjustment of output takes time. Namely, ˙ Y = φ(A?Y)=φ[?(1 ?b)Y ?βr + a + α?bT + G], where φ(·) is a strictly increasing function with φ(0) = 0. Third, assume that there are two types of bonds, 1. short nominal bonds that pay a nominal rate i, and 2. real consols that promise to pay one good per unit of time forever. Let Q be the real price of a consol; then the long-term return from the consol is r = 1 Q . The short-term return from the consol is 1 Q + ˙ Q Q = r? ˙r r . Assume that asset holders equalize rates of return on both assets up to a constant risk premium η, i.e., r ? ˙r r = i?π e + η, where π e is the expected inflation rate. Finally, the money demand function is in the standard form: M d P = L(i,Y ). 3—12