JOURNAL OF GUIDANCE, CONTROL, AND DYNAMICS
Vol. 21, No. 2, March–April 1998
Survey of Numerical Methods for Trajectory Optimization
John T. Betts
Boeing Information and Support Services, Seattle, Washington 98124-2207
I. Introduction
I T is not surprisingthat thedevelopmentof numericalmethods fortrajectory optimization have closely paralleled the exploration
of space and the developmentof the digital computer. Space explo-
ration provided the impetus by presenting scientists and engineers
with challengingtechnicalproblems.The digital computerprovided
the tool for solving these new problems.The goal of this paper is to
review the state of the art in the eld loosely referred to as trajectory
optimization.
Presentinga surveyof a eld as diverse as trajectoryoptimization
is a daunting task. Perhaps the most dif cult issue is restricting the
scope of the survey to permit a meaningful discussion within a
limited amount of space. To achieve this goal, I made a conscious
decision to focus on the two types of methods most widely used
today,namely, direct and indirect. I begin the discussionwith a brief
review of the underlying mathematics in both direct and indirect
methods. I then discuss the complications that occur when path
and boundary constraints are imposed on the problem description.
Finally, I describe unresolved issues that are the subject of ongoing
research.
A number of recurrent themes appear throughoutthe paper. First,
the aforementioned direct vs indirect is introduced as a means of
categorizing an approach. Unfortunately, not every technique falls
neatly into one category or another. I will attempt to describe the
bene ts and de ciencies in both approaches and then suggest that
the techniquesmay ultimatelymerge. Second, I shall attempt to dis-
criminate betweenmethod vs implementation.A numericalmethod
is usually described by mathematical equations and/or algorithmic
logic. Computational results are achieved by implementing the al-
gorithm as software, e.g., Fortran code. A second level of imple-
mentation may involve translating a (pre ight) scienti c software
implementationinto an (onboard)hardwareimplementation.In gen-
eral, method and implementationare not the same, and I shall try to
emphasizethat fact. Third, I shall focus the discussionon algorithms
instead of physical models. The de nition of a trajectory problem
necessarilyentails a de nition of the dynamic environment such as
gravitational, propulsion, and aerodynamic forces. Thus it is com-
mon to use the same algorithm, with different physical models, to
solve different problems. Conversely, different algorithms may be
applied to the same physicalmodels (with entirelydifferentresults).
Finally, I shall attempt to focus on general rather than special pur-
pose methods. A great deal of research has been directed toward
solving speci c problems. Carefully specialized techniques can ei-
ther be very effectiveor very ad hoc.Unfortunately,whatworkswell
for a launch vehicle guidance problemmay be totally inappropriate
for a low-thrust orbit transfer.
John T. Betts received a B.A. degree from Grinnell College, Grinnell, Iowa, in 1965 with a major in physics
and a minor in mathematics. In 1967 he received anM.S. in astronautics from Purdue University, West Lafayette,
Indiana, with a major in orbit mechanics, and in 1970 he received a Ph.D. from Purdue, specializing in optimal
control theory. He joined The Aerospace Corporation in 1970 as a Member of the Technical Staff and from 1977–
1987 was manager of the Optimization Techniques Section of the Performance Analysis Department. He joined
The Boeing Company, serving as manager of the Operations Research Group of Boeing Computer Services from
1987–1989.In his current positionasSeniorPrincipalScientist in the AppliedResearch andTechnologyDivision,he
provides technical support to all areas of The Boeing Company.Dr. Betts is a Member of AIAA and the Society for
Industrial and Applied Mathematics with active research in nonlinear programming and optimal control theory.
E-mail: John.T.Betts@boeing.com.
Received March 29, 1997; revision received Oct. 7, 1997; accepted for publicationOct. 20, 1997.Copyright c° 1997 by theAmerican Institute of Aeronautics
and Astronautics, Inc. All rights reserved.
II. Trajectory Optimization Problem
Let us begin the discussion by de ning the problem in a fairly
general way. A trajectory optimization or optimal control problem
can be formulated as a collection of N phases. In general, the inde-
pendentvariable t for phase k is de ned in the region t .k/0 · t · t .k/f .
For many applications the independent variable t is time, and the
phases are sequential, that is, t .k C 1/0 D t .k/f ; however, neither of
those assumptions is required.Within phase k the dynamics of the
system are described by a set of dynamic variables
z D y
.k/.t/
u.k/.t/
(1)
made up of the n.k/y state variables and the n
.k/
u control variables,
respectively. In addition, the dynamics may incorporate the n.k/p
parameters p.k/, which are not dependent on t . For clarity I drop
the phase-dependentnotation from the remaining discussion in this
section. However, it is important to remember that many complex
problemdescriptionsrequire different dynamics and/or constraints,
and a phase-dependent formulation accommodates this require-
ment.
Typically the dynamics of the system are de ned by a set of
ordinary differential equations written in explicit form, which are
referred to as the state or system equations,
Py D f [ y.t/;u.t/; p; t ] (2)
where y is the n y dimension state vector. Initial conditions at time
t0 are de ned by
Ã0l · Ã[ y.t0/;u.t0/; p; t0] · Ã0u (3)
where Ã[ y.t0/;u.t0/; p; t0] ´ Ã0 , and terminal conditions at the
nal time t f are de ned by
à f l · Ã[ y.t f /; u.t f /; p; t f ] · à f u (4)
where Ã[ y.t f /; u.t f /; p; t f ] ´ Ã f . In addition, the solution must
satisfy algebraic path constraints of the form
gl · g[ y.t/;u.t/; p; t] · gu (5)
where g is a vector of size ng , as well as simple bounds on the state
variables
yl · y.t/ · yu (6)
193
D
ow
nl
oa
de
d
by
Je
t P
ro
pu
lsi
on
L
ab
or
at
or
y
on
O
ct
ob
er
2
6,
2
01
2
| ht
tp:
//a
rc.
aia
a.o
rg
| D
OI
: 1
0.2
514
/2.
423
1
194 BETTS
and control variables
ul · u.t/ · uu (7)
Note that an equality constraint can be imposed if the upper and
lower bounds are equal, e.g., .gl /k D .gu/k for some k.
Finally, it may be convenient to evaluate expressionsof the form
t f
t0
q[ y.t/;u.t/; p; t ] dt (8)
which involve the quadrature functions q. Collectively we refer to
those functions evaluated during the phase, namely,
F.t/ D
f [ y.t/;u.t/; p; t ]
g[ y.t/;u.t/; p; t ]
q[ y.t/;u.t/; p; t ]
(9)
as the vectorof continuousfunctions.Similarly, functionsevaluated
at a speci c point,such as theboundaryconditionsÃ[ y.t0/,u.t0/; t0]
and Ã[ y.t f /; u.t f /; t f ], are referred to as point functions.
The basic optimal control problem is to determine the n.k/u -
dimensionalcontrol vectorsu.k/.t/ and parametersp.k/ to minimize
the performance index
J D Á y t .1/0 ; t .1/0 ; y t .1/f ; p.1/; t .1/f ; : : : ;
y t .N /0 ; t
.N /
0 ; y t
.N /
f ; p
.N /; t .N /f (10)
Notice that the objective function may depend on quantities com-
puted in each of the N phases.
This formulation raises a number of points that deserve further
explanation. The concept of a phase, also referred to as an arc by
some authors, partitions the time domain. In this formalism the dif-
ferential equations cannot change within a phase but may change
fromone phase to another.An obviousreason to introducea phase is
to accommodatechanges in the dynamics, for example,when simu-
latingamultistagerocket.The boundaryof a phase is often calledan
event or junction point. A boundary condition that uniquely de nes
the end of a phase is sometimescalledan event criterion(and a well-
posedproblemcan have onlyone criterionat each event). Normally,
the simulation of a complicated trajectorymay link phases together
by forcing the states to be continuous, e.g., y[t .1/f ]D y[t .2/0 ]. How-
ever, for multipath or branch trajectories this may not be the case.1
The differential equations (2) have been written as an explicit sys-
tem of rst-order equations, i.e., with the rst derivative appearing
explicitlyon the left-handside,which is the standardconventionfor
aerospace applications.Although this simpli es the presentation, it
may not be eithernecessaryor desirable;e.g.,Newton’s lawFDma
is not stated as an explicit rst-order system!The objective function
(10) has been written in terms of quantities evaluated at the ends
of the phases, and this is referred to as the Mayer form.2 If the ob-
jective function only involves an integral (8), it is referred to as a
problem of Lagrange, and when both terms are present, it is called
a problem of Bolza. It is well known that the Mayer form can be
obtained from either the Lagrange or Bolza form by introducingan
additional state variable.However, again this may have undesirable
numerical consequences.
III. Nonlinear Programming
A. Newton’s Method
Essentially all numericalmethods for solving the trajectory opti-
mization problemincorporatesome typeof iterationwith a nite set
of unknowns. In fact, progress in optimal control solutionmethods
closely parallels the progressmade in the underlyingnonlinearpro-
gramming (NLP) methods. Space limitations preclude an in-depth
presentation of constrained optimization methods. However, it is
important to review some of the fundamental concepts. For more
complete information the reader is encouraged to refer to the books
by Fletcher,3 Gill et al.,4 and Dennis and Schnabel.5 The funda-
mental approach to most iterative schemes was suggested over 300
years ago by Newton. Suppose we are trying to solve the nonlinear
algebraic equations a.x/D 0 for the root x¤. Beginning with an es-
timate x we can construct a new estimate Nx according to
Nx D xC ®p (11)
where the search direction p is computed by solving the linear
system
A.x/p D ¡a.x/ (12)
The n £ n matrix A is de ned by
A D
@a1
@x1
@a1
@x2
¢ ¢ ¢ @a1
@xn
@a2
@x1
@a2
@x2
¢ ¢ ¢ @a2
@xn
:::
: : :
@an
@x1
@an
@x2
¢ ¢ ¢ @an
@xn
(13)
When the scalar step length ® is equal to 1, the iteration scheme
is equivalent to replacing the nonlinear equation by a linear ap-
proximation constructed about the point x. We expect the method
to converge provided the initial guess is close to the root x¤. Of
course, this simple scheme is not without pitfalls. First, to compute
the search direction, the matrix A must be nonsingular (invertible),
and for arbitrarynonlinear functionsa.x/ this may not be true. Sec-
ond, when the initial guess is not close to the root, the iterationmay
diverge.One common way to stabilize the iteration is to reduce the
length of the step by choosing ® such that
ka.Nx/k · ka.x/k (14)
The procedure for adjusting the step length is called a line search,
and the functionused to measureprogress(in this casekak) is called
a merit function. In spite of the need for caution, Newton’s method
enjoysbroad applicability,possiblybecause,when it works, the iter-
ates exhibit quadratic convergence.Loosely speaking, this property
means that the number of signi cant gures in x (as an estimate for
x¤) doubles from one iteration to the next.
B. Unconstrained Optimization
Let us nowconsideran unconstrainedoptimizationproblem.Sup-
pose that we must choose the n variables x to minimize the scalar
objective functionF.x/. Necessary conditionsfor x¤ to be a station-
ary point are
g.x¤/ D rx F D
@F
@x1
@F
@x2
:::
@F
@xn
D 0 (15)
Now if Newton’s method is used to nd a point where the gradi-
ent (15) is zero, we must compute the search direction using
H.x/p D ¡g.x/ (16)
where the Hessian matrix H is the symmetric matrix of second
derivativesof the objective function. Just as before, there are pitfalls
in using this method to construct an estimate of the solution. First,
we note that the conditiongD 0 is necessarybut not suf cient. Thus
a point with zero gradient can be either a maximum or a minimum.
At a minimum point the Hessian matrix is positive de nite, but this
may not be true when H is evaluated at some point far from the
solution. In fact, it is quite possible that the direction p computed
by solving Eq. (16) will point uphill rather than downhill. Second,
there is some ambiguity in the choice of a merit function if a line
D
ow
nl
oa
de
d
by
Je
t P
ro
pu
lsi
on
L
ab
or
at
or
y
on
O
ct
ob
er
2
6,
2
01
2
| ht
tp:
//a
rc.
aia
a.o
rg
| D
OI
: 1
0.2
514
/2.
423
1
BETTS 195
search is used to stabilize the method. Certainly we would hope to
reduce the objective function, that is, F.Nx/ · F .x/. However, this
may not produce a decrease in the gradient
kg.Nx/k · kg.x/k (17)
In fact, what have been described are two issues that distinguish a
direct and an indirect method for nding a minimum. For an indi-
rect method, a logical choice for the merit function is kg.x/k. In
contrast, for a direct method, one probably would insist that the
objective function is reduced at each iteration, and to achieve this
it may be necessary to modify the calculation of the search direc-
tion to ensure that it is downhill. A consequence of this is that the
region of convergence for an indirect method may be considerably
smaller than the region of convergence for a direct method. Stated
differently,an indirectmethodmay requirea better initial guess than
required by a directmethod. Second, to solve the equationsgD 0, it
is necessary to compute the expressionsg.x/! Typically this implies
that analytic expressions for the gradient are necessarywhen using
an indirect method. In contrast, nite difference approximations to
the gradient are often used in a direct method.
C. Equality Constraints
Suppose that we must choose the n variables x to minimize the
scalar objective function F.x/ and satisfy them equality constraints
c.x/ D 0 (18)
where m · n. We introduce the Lagrangian
L.x;¸/ D F.x/ ¡ ¸>c.x/ (19)
which is a scalar function of the n variables x and the m Lagrange
multipliers ¸. Necessary conditions for the point .x¤;¸¤/ to be
a constrained optimum require nding a stationary point of the
Lagrangian that is de ned by
rx L.x;¸/ D g.x/¡ G>.x/¸ D 0 (20)
and
r¸L.x;¸/ D ¡c.x/ D 0 (21)
By analogywith the development in the preceding sections,we can
useNewton’s method to nd the .nCm/ variables.x;¸/ that satisfy
the conditions (20) and (21). Proceeding formally to construct the
linear system equivalent to Eq. (12), one obtains
HL G>
G 0
p
¡ N¸ D
¡g
¡c (22)
This system requires the Hessian of the Lagrangian
HL D r2x F ¡
m
i D 1
¸ir2x ci (23)
The linear system (22) is referred to as the Kuhn–Tucker (KT) or
Karush–Kuhn–Tucker system. It is important to observe that an
equivalentway of de ning the search direction p is to minimize the
quadratic
1
2 p
>HL pC g>p (24)
subject to the linear constraints
Gp D ¡c (25)
This is referred to as a quadratic programming (QP) subproblem.
Just as in the unconstrainedand root solving applicationsdiscussed
earlier, when Newton’s method is applied to equality constrained
problems, it may be necessary to modify either the magnitude of
the step via a line search or the direction itself using a trust region
approach. However, regardless of the type of stabilization invoked
at points far from the solution, near the answer all methods try to
mimic the behavior of Newton’s method.
D. Inequality Constraints
An important generalization of the preceding problem occurs
when inequality constraints are imposed. Suppose that we must
choose the n variables x to minimize the scalar objective function
F.x/ and satisfy the m inequality constraints
c.x/ ¸ 0 (26)
In contrast to the equally constrained case, now m may be greater
than n. However, at the optimal point x¤, the constraints will fall
into one of two classes. Constraints that are strictly satis ed, i.e.,
ci .x¤/ > 0, are called inactive.The remainingactive constraints are
on their bounds, i.e., ci .x¤/D 0. If the active set of constraints is
known, then one can simply ignore the remaining constraints and
treat the problem using methods for an equality constrained prob-
lem. However, algorithms to ef ciently determine the active set of
constraints are nontrivial because they require repeated solution of
the KT system (22) as constraints are added and deleted. In spite
of the complications,methods for nonlinear programming based on
the solution of a series of quadratic programming subproblems are
widely used. A popular implementation of the so-called successive
or sequential quadratic programming (SQP) approach is found in
the software NPSOL.6;7
In summary, the NLP problem requires nding the n vector x to
minimize
F.x/ (27)
subject to the m constraints
cL · c.x/ · cU (28)
and bounds
xL · x · xU (29)
In this formulation, equality constraints can be imposed by setting
cL D cU .
E. Historical Perspective
Progress in the developmentof NLP algorithms has been closely
tied to the advent of the digital computer. Because NLP software is
a primary piece of the trajectory optimization tool kit, it has been
a pacing item in the development of sophisticated trajectory opti-
mization software. In the early 1960s most implementations were
based on a simple Newton method (11) and (12) with optimization
done parametrically, i.e., by hand. The size of a typical application
was nDm ¼ 10. In the 1970s quasi-Newton approximations3;5 be-
came prevalent.One popular approach for dealing with constraints
was to applyan unconstrainedminimizationalgorithmto a modi ed
formof theobjective,e.g.,minimize J .x; ½/D F.x/C 1
2
½c.x/>c.x/,
where½ is large.Althoughthose techniqueshavegenerallybeen su-
perseded for generaloptimization,curiouslyenough they are funda-
mental to the de nition of the merit functionsused to stabilizestate-
of-the-art algorithms. A second popular approach for constrained
problems, referred to as the reduced gradient approach, identi es a
basic set of variablesthat are used to eliminate the activeconstraints,
permitting choice of the nonbasic variables using an unconstrained
technique.Careful implementationsof this method8¡10 can be quite
effective, especially when the constraints are nearly linear and the
number of inequalities is small. Most applications in the 1970s and
early 1980s were of moderate size, i.e., nDm< 100. Current ap-
plications have incorporated advances in numerical linear algebra
that exploit matrix sparsity, thereby permitting applications with
n;m ¼ 100;000 (Refs. 11–20).
IV. Optimal Control
A. Dynamic Constraints
The optimal control problemmay be interpreted as an extension
of the nonlinear programming problem to an in nite number of
variables. For fundamental background in the associated calculus
of variations the reader should refer to Ref. 21. First let us con-
sider a simple problemwith a single phase and no path constraints.
D
ow
nl
oa
de
d
by
Je
t P
ro
pu
lsi
on
L
ab
or
at
or
y
on
O
ct
ob
er
2
6,
2
01
2
| ht
tp:
//a
rc.
aia
a.o
rg
| D
OI
: 1
0.2
514
/2.
423
1
196 BETTS
Speci cally, suppose we must choose the control functions u.t/ to
minimize
J D Á[ y.t f /; t f ] (30)
subject to the state equations
Py D f [ y.t/;u.t/] (31)
and the boundary conditions
Ã[ y.t f /; u.t f /; t f ] D 0 (32)
where the initial conditions y.t0/D y0 are given at the xed initial
time t0, and the nal time t f is free. Note that this is a very simpli ed
version of the problem (2–10), and we have purposely chosen a
problemwith only equality constraints.However, in contrast to the
previous discussion,we now have a continuous equality constraint
(31) as well as a discrete equality (32). In a manner analogousto the
de nition of the Lagrangian function (19), we form an augmented
performance index
OJ D [Á C º>Ã]t f C
t f
t
本文档为【航天器轨迹优化数值方法综述】,请使用软件OFFICE或WPS软件打开。作品中的文字与图均可以修改和编辑,
图片更改请在作品中右键图片并更换,文字修改请直接点击文字进行修改,也可以新增和删除文档中的内容。
该文档来自用户分享,如有侵权行为请发邮件ishare@vip.sina.com联系网站客服,我们会及时删除。
[版权声明] 本站所有资料为用户分享产生,若发现您的权利被侵害,请联系客服邮件isharekefu@iask.cn,我们尽快处理。
本作品所展示的图片、画像、字体、音乐的版权可能需版权方额外授权,请谨慎使用。
网站提供的党政主题相关内容(国旗、国徽、党徽..)目的在于配合国家政策宣传,仅限个人学习分享使用,禁止用于任何广告和商用目的。