This fully revised 3rd edition offers an introduction to optimal control theory and its diverse applications in management science and economics. It brings to students the concept of the maximum principle in continuous, as well as discrete, time by using dynamic programming and Kuhn-Tucker theory. While some mathematical background is needed, the emphasis of the book is not on mathematical rigor, but on modeling realistic situations faced in business and economics. The book exploits optimal control theory to the functional areas of management including finance, production and marketing and to economics of growth and of natural resources. In addition, this new edition features materials on stochastic Nash and Stackelberg differential games and an adverse selection model in the principal-agent framework. The book provides exercises for each chapter and answers to selected exercises to help deepen the understanding of the material presented. Also included are appendices comprised of supplementary material on the solution of differential equations, the calculus of variations and its relationships to the maximum principle, and special topics including the Kalman filter, certainty equivalence, singular control, a global saddle point theorem, Sethi-Skiba points, and distributed parameter systems.
Optimal control methods are used to determine optimal ways to control a dynamic system. The theoretical work in this field serves as a foundation for the book, which the author has applied to business management problems developed from his research and classroom instruction. The new edition has been completely refined and brought up to date. Ultimately this should continue to be a valuable resource for graduate courses on applied optimal control theory, but also for financial and industrial engineers, economists, and operational researchers concerned with the application of dynamic optimization in their fields.