1

المرجع الالكتروني للمعلوماتية

تاريخ الرياضيات

الاعداد و نظريتها

تاريخ التحليل

تار يخ الجبر

الهندسة و التبلوجي

الرياضيات في الحضارات المختلفة

العربية

اليونانية

البابلية

الصينية

المايا

المصرية

الهندية

الرياضيات المتقطعة

المنطق

اسس الرياضيات

فلسفة الرياضيات

مواضيع عامة في المنطق

الجبر

الجبر الخطي

الجبر المجرد

الجبر البولياني

مواضيع عامة في الجبر

الضبابية

نظرية المجموعات

نظرية الزمر

نظرية الحلقات والحقول

نظرية الاعداد

نظرية الفئات

حساب المتجهات

المتتاليات-المتسلسلات

المصفوفات و نظريتها

المثلثات

الهندسة

الهندسة المستوية

الهندسة غير المستوية

مواضيع عامة في الهندسة

التفاضل و التكامل

المعادلات التفاضلية و التكاملية

معادلات تفاضلية

معادلات تكاملية

مواضيع عامة في المعادلات

التحليل

التحليل العددي

التحليل العقدي

التحليل الدالي

مواضيع عامة في التحليل

التحليل الحقيقي

التبلوجيا

نظرية الالعاب

الاحتمالات و الاحصاء

نظرية التحكم

بحوث العمليات

نظرية الكم

الشفرات

الرياضيات التطبيقية

نظريات ومبرهنات

علماء الرياضيات

500AD

500-1499

1000to1499

1500to1599

1600to1649

1650to1699

1700to1749

1750to1779

1780to1799

1800to1819

1820to1829

1830to1839

1840to1849

1850to1859

1860to1864

1865to1869

1870to1874

1875to1879

1880to1884

1885to1889

1890to1894

1895to1899

1900to1904

1905to1909

1910to1914

1915to1919

1920to1924

1925to1929

1930to1939

1940to the present

علماء الرياضيات

الرياضيات في العلوم الاخرى

بحوث و اطاريح جامعية

هل تعلم

طرائق التدريس

الرياضيات العامة

نظرية البيان

الرياضيات : نظرية التحكم :

THE PONTRYAGIN MAXIMUM PRINCIPLE-APPLICATIONS AND EXAMPLES

المؤلف:  Lawrence C. Evans

المصدر:  An Introduction to Mathematical Optimal Control Theory

الجزء والصفحة:  50-59

9-10-2016

411

HOW TO USE THE MAXIMUM PRINCIPLE. We mentioned earlier that the costate p(.) can be interpreted as a sort of Lagrange multiplier.

Calculations with Lagrange multipliers. Recall our discussion in (REVIEW OF LAGRANGE MULTIPLIERS)about finding a point x∗ that maximizes a function f, subject to the requirement that g ≤ 0. Now     x = (x1, . . . , xn)T has n unknown components we must find.

Somewhat unexpectedly, it turns out in practice to be easier to solve (1.4) for the n + 1 unknowns x1, . . . , xn and λ. We repeat this key insight: it is actually easier to solve the problem if we add a new unknown, namely the Lagrange multiplier.

Worked examples abound in multivariable calculus books.

Calculations with the costate. This same principle is valid for our much more complicated control theory problems: it is usually best not just to look for an optimal control α(.) and an optimal trajectory x(.) alone, but also to look as well for the costate p(.). In practice, we add the equations (ADJ) and (M) to (ODE)  and then try to solve for α(.), x(.) and for p(.).

The following examples show how this works in practice, in certain cases for which we can actually solve everything explicitly or, failing that, at least deduce some useful information.

EXAMPLE 1: LINEAR TIME-OPTIMAL CONTROL. For this example, let A denote the cube [−1, 1]  n in Rn. We consider again the linear dynamics:

for the payoff functional

where τ denotes the first time the trajectory hits the target point x1 = 0. We have r ≡ −1, and so

                          H(x, p, a) = f . p + r = (Mx + Na) . p − 1.

1.2 EXAMPLE 2: CONTROL OF PRODUCTION AND CONSUMPTION. We return to Example 1 in INTRODUCTION, a model for optimal consumption in a simple economy. Recall that

x(t) = output of economy at time t,

α(t) = fraction of output reinvested at time t.

We have the constraint 0 ≤ α(t) ≤ 1; that is, A = [0, 1] ⊂ R. The economy evolves according to the dynamics

where x0 > 0 and we have set the growth factor k = 1. We want to maximize the total consumption

How can we characterize an optimal control α(.)?

Introducing the maximum principle. We apply Pontryagin Maximum Principle, and to simplify notation we will not write the superscripts ∗ for the optimal control, trajectory, etc. We have n = m = 1,

                                f(x, a) = xa, g ≡ 0, r(x, a) = (1 − a)x;

and therefore

                     H(x, p, a) = f(x, a)p + r(x, a) = pxa + (1 − a)x = x + ax(p − 1).

The dynamical equation is

and the adjoint equation is

 

The terminal condition reads

(T)                                                       p(T) = gx(x(T)) = 0.

Lastly, the maximality principle asserts

Using the maximum principle. We now deduce useful information from  (ODE), (ADJ), (M) and (T).

According to (M), at each time t the control value α(t) must be selected to maximize a(p(t) − 1) for 0 ≤ a ≤ 1. This is so, since x(t) > 0. Thus

Hence if we know p(.), we can design the optimal control α(.).

So next we must solve for the costate p(.). We know from (ADJ) and (T) that

Since p(T) = 0, we deduce by continuity that p(t) ≤ 1 for t close to T, t < T. Thus α(t) = 0 for such values of t. Therefore ˙ p(t) = −1, and consequently p(t) = T – t for times t in this interval. So we have that p(t) = T − t so long as p(t) ≤ 1. And this holds for T − 1 ≤ t ≤ T But for times t ≤ T − 1, with t near T − 1, we have α(t) = 1; and so (ADJ)  becomes

Since p(T − 1) = 1, we see that p(t) = eT−1−t> 1 for all times 0 ≤ t ≤ T − 1. In particular there are no switches in the control over this time interval.

Restoring the superscript * to our notation, we consequently deduce that an optimal control is

for the optimal switching time t = T − 1.

We leave it as an exercise to compute the switching time if the growth constant k = 1.

1.3 EXAMPLE 3: A SIMPLE LINEAR-QUADRATIC REGULATOR. We take n = m = 1 for this example, and consider the simple linear dynamics

 

with the quadratic cost functional

which we want to minimize. So we want to maximize the payoff functional

For this problem, the values of the controls are not constrained; that is, A = R.

Introducing the maximum principle. To simplify notation further we again drop the superscripts ∗. We have n = m = 1,

f(x, a) = x + a, g ≡ 0, r(x, a) = −x2− a2;

and hence

H(x, p, a) = fp + r = (x + a)p − (x2+ a2)

The maximality condition becomes

We calculate the maximum on the right hand side by setting Ha = −2a + p = 0.

Thus a = p/2 , and so

The dynamical equations are therefore

Moreover x(0) = x0 , and the terminal condition is  (T) p(T) = 0.

Using the Maximum Principle. So we must look at the system of equations

the general solution of which is

Since we know x0, the task is to choose p0 so that p(T) = 0.

Feedback controls. An elegant way to do so is to try to find optimal control in linear feedback form; that is, to look for a function c(.) : [0, T] → R for which α(t) = c(t) x(t).

We henceforth suppose that an optimal feedback control of this form exists,  and attempt to calculate c(.). Now

We will next discover a differential equation that d(.) satisfies. Compute

Since p(T) = 0, the terminal condition is d(T) = 0.

So we have obtained a nonlinear first–order ODE for d(.) with a terminal boundary condition:

This is called the Riccati equation.

In summary so far, to solve our linear–quadratic regulator problem, we need to first solve the Riccati equation (R) and then set

How to solve the Riccati equation. It turns out that we can convert (R) it into a second–order, linear ODE. To accomplish this, write

for a function b(.) to be found. What equation does b(.) solve? We compute

This is a terminal-value problem for second–order linear ODE, which we can solve by standard techniques. We then set d = 2b˙ / b , to derive the solution of the Riccati equation (R).

We will generalize this example later to systems

1.4 EXAMPLE 4: MOON LANDER. This is a much more elaborate and interesting example, already introduced in INTRODUCTION. We follow the discussion of Fleming and Rishel [F-R].

Introduce the notation

h(t) = height at time t

v(t) = velocity = ˙ h(t)

m(t) = mass of spacecraft (changing as fuel is used up)

α(t) = thrust at time t.

The thrust is constrained so that 0 ≤ α(t) ≤ 1; that is, A = [0, 1]. There are also the constraints that the height and mass be nonnegative: h(t) ≥ 0,m(t) ≥ 0.

The dynamics are

with initial conditions

                                                                     

The goal is to land on the moon safely, maximizing the remaining fuel m(τ ),  where τ = τ [α()] is the first time h(τ ) = v(τ ) = 0. Since α = − m˙/ k , our intention is equivalently to minimize the total applied thrust before landing; so that

Introducing the maximum principle. In terms of the general notation, we have

Hence the Hamiltonian is

We next have to figure out the adjoint dynamics (ADJ). For our particular Hamiltonian,

Therefore

The maximization condition (M) reads

(M)

Thus the optimal control law is given by the rule:

Using the maximum principle. Now we will attempt to figure out the form of the solution, and check it accords with the Maximum Principle.

Let us start by guessing that we first leave rocket engine of (i.e., set α ≡ 0) and turn the engine on only at the end. Denote by τ the first time that h(τ ) = v(τ ) = 0,  meaning that we have landed. We guess that there exists a switching time t < τ when we turned engines on at full power (i.e., set α ≡ 1).Consequently,

Therefore, for times t ≤ t ≤ τ our ODE becomes

Suppose the total amount of fuel to start with was m1; so that m0 −m1 is the weight of the empty spacecraft. When α ≡ 1, the fuel is used up at rate k. Hence

k(τ − t) ≤ m1,

Before time t, we set α ≡ 0. Then (ODE) reads

and thus

We combine the formulas for v(t) and h(t), to discover

We deduce that the freefall trajectory (v(t), h(t)) therefore lies on a parabola

If we then move along this parabola until we hit the soft-landing curve from the previous picture, we can then turn on the rocket engine and land safely.

In the second case illustrated, we miss switching curve, and hence cannot land safely on the moon switching once.

To justify our guess about the structure of the optimal control, let us now find the costate p(.) so that α(.) and x(.) described above satisfy (ODE), (ADJ), (M).

To do this, we will have have to figure out appropriate initial conditions

p1 (0) = λ1, p2 (0) = λ2, p3 (0) = λ3.

We solve (ADJ) for α(.) as above, and find

and then adjust λ2, λ3 so that r(t) = 0.

Then r is nonincreasing, r(t) = 0, and consequently r > 0 on [0, t), r < 0 on (t, τ ]. But (M) says

Thus an optimal control changes just once from 0 to 1; and so our earlier guess of α(.) does indeed satisfy the Pontryagin Maximum Principle.

References

[B-CD] M. Bardi and I. Capuzzo-Dolcetta, Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations, Birkhauser, 1997.

[B-J] N. Barron and R. Jensen, The Pontryagin maximum principle from dynamic programming and viscosity solutions to first-order partial differential equations, Transactions AMS 298 (1986), 635–641.

[C1] F. Clarke, Optimization and Nonsmooth Analysis, Wiley-Interscience, 1983.

[C2] F. Clarke, Methods of Dynamic and Nonsmooth Optimization, CBMS-NSF Regional Conference Series in Applied Mathematics, SIAM, 1989.

[Cr] B. D. Craven, Control and Optimization, Chapman & Hall, 1995.

[E] L. C. Evans, An Introduction to Stochastic Differential Equations, lecture notes avail-able at http://math.berkeley.edu/˜ evans/SDE.course.pdf.

[F-R] W. Fleming and R. Rishel, Deterministic and Stochastic Optimal Control, Springer, 1975.

[F-S] W. Fleming and M. Soner, Controlled Markov Processes and Viscosity Solutions, Springer, 1993.

[H] L. Hocking, Optimal Control: An Introduction to the Theory with Applications, OxfordUniversity Press, 1991.

[I] R. Isaacs, Differential Games: A mathematical theory with applications to warfare and pursuit, control and optimization, Wiley, 1965 (reprinted by Dover in 1999).

[K] G. Knowles, An Introduction to Applied Optimal Control, Academic Press, 1981.

[Kr] N. V. Krylov, Controlled Diffusion Processes, Springer, 1980.

[L-M] E. B. Lee and L. Markus, Foundations of Optimal Control Theory, Wiley, 1967.

[L] J. Lewin, Differential Games: Theory and methods for solving game problems with singular surfaces, Springer, 1994.

[M-S] J. Macki and A. Strauss, Introduction to Optimal Control Theory, Springer, 1982.

[O] B. K. Oksendal, Stochastic Differential Equations: An Introduction with Applications, 4th ed., Springer, 1995.

[O-W] G. Oster and E. O. Wilson, Caste and Ecology in Social Insects, Princeton UniversityPress.

[P-B-G-M] L. S. Pontryagin, V. G. Boltyanski, R. S. Gamkrelidze and E. F. Mishchenko, The Mathematical Theory of Optimal Processes, Interscience, 1962.

[T] William J. Terrell, Some fundamental control theory I: Controllability, observability,  and duality, American Math Monthly 106 (1999), 705–719.

 

 

 

 

 

EN

تصفح الموقع بالشكل العمودي