Sunteți pe pagina 1din 24

Dyn Games Appl (2011) 1:534557

DOI 10.1007/s13235-011-0024-5
Numerical Solution of Orbital Combat Games Involving
Missiles and Spacecraft
Mauro Pontani
Published online: 14 July 2011
Springer Science+Business Media, LLC 2011
Abstract This research addresses the problem of the optimal interception of an optimally
evasive orbital target by a pursuing spacecraft or missile. The time for interception is to be
minimized by the pursuing space vehicle and maximized by the evading target. This problem
is modeled as a two-sided optimization problem, i.e. as a two-player zero-sum differential
game. The work incorporates a recently developed method, termed semi-direct collocation
with nonlinear programming, for the numerical solution of dynamic games. The method is
based on the formal conversion of the two-sided optimization probleminto a single-objective
one, by employing the analytical necessary conditions for optimality related to one of the
two players. An approximate, rst attempt solution for the method is provided through the
use of a genetic algorithm in a preprocessing phase. Three qualitatively different cases are
considered. In the rst example the pursuer and the evader are represented by two space-
craft orbiting the Earth in two distinct orbits. The second and the third case involve two
missiles, and a missile that pursues an orbiting spacecraft, respectively. The numerical re-
sults achieved in this work testify to the robustness and effectiveness of the method also in
solving large, complex, three-dimensional problems.
Keywords Orbital dynamic games Pursuit-evasion games Two-sided optimization
1 Introduction
The problem of the three-dimensional optimal interception of an optimally evasive orbital
target by a pursuing spacecraft (or missile) involves two competing players with contrasting
objectives. The pursuing space vehicle tries to reach the evading target as quickly as possi-
ble, whereas this latter tries to delay capture indenitely. Since the time for interception is
to be minimized by the pursuing spacecraft and maximized by the evading spacecraft, this
problem is best modeled as a two-sided optimization problem, i.e. it becomes a two-player
zero-sum differential game.
M. Pontani ()
Scuola di Ingegneria Aerospaziale, University of Rome La Sapienza, 00138 Rome, Italy
e-mail: mauro.pontani@uniroma1.it
Dyn Games Appl (2011) 1:534557 535
Zero-sum games were rst introduced by Isaacs [1] and are also referred to as pursuit-
evasion games. In the context of zero-sum games the optimal trajectories of the two
spacecraft correspond to a so-called saddle-point equilibrium solution of the game. The
necessary conditions for an open-loop saddle-point equilibrium solution are relatively
straightforward to derive and can be viewed as an extension of the necessary conditions for
optimality that hold in optimal control theory [2, 3]. Their meaningfulness in relation with
closed-loop saddle-point equilibrium solutions is closely related to the intrinsic character-
istics of the game of interest. Only a few problems with simplied dynamics are amenable
to an analytical solution [4, 5]. For problems with realistic dynamics the only choice is nu-
merical solution. Hillberg and Jrmark [6] solved an air combat maneuvering problem in
the horizontal plane with steady turn and realistic drag and thrust data. Jrmark, Merz, and
Breakwell [7] solved a qualitatively similar air combat problem employing differential dy-
namic programming, and considered only coplanar situations. A pursuit-evasion problem
between missile and aircraft has been solved using an indirect, multiple shooting method by
Breitner, Grimm and Pesch [8, 9]. Raivio and Ehtamo [10] solved a pursuit-evasion problem
for a visual identication of the target by iterating a direct method. With regard to orbital
pursuit-evasion games, past studies are often based on simplied dynamical models. Ander-
son and Grazier [11] described the construction of a closed-form solution for the barrier in a
planar pursuit-evasion game between two spacecraft, by linearizing the problem about a ref-
erence circular orbit. Kelley et al. [12] derived the impulsive maneuvers for two spacecraft
involved in an orbital combat. They argued that optimal evasion only consists of in-plane
maneuvers.
The work that follows presents a recently developed method [1316], termed semi-
direct collocation with nonlinear programming (semi-DCNLP), devoted to the numerical
solution of zero-sum dynamic games with separable dynamics of the two players. This
method is based on the formal conversion of the two-sided optimization problem into a
single-objective one, by employing the analytical necessary conditions for optimality re-
lated to one of the two players. This fact implies that the adjoint variables of one of the two
spacecraft are directly involved in the optimization process, which needs a reasonable guess
to yield an accurate saddle-point equilibrium solution. The trial-and-error selection of rst
attempt values for the (non-intuitive) adjoint variables is very challenging for the problem
at hand. In this work an approximate, rst attempt solution is provided through the use of
a genetic algorithm in a preprocessing phase. Three qualitatively different cases are con-
sidered. In the rst example the pursuer and the evader are represented by two spacecraft
orbiting Earth in two distinct orbits. The second and the third case involve two missiles, and
a missile that pursues an orbiting spacecraft, respectively.
The objective of this work is to: (i) formulate the three-dimensional orbital combat as a
dynamic game, (ii) describe and derive the analytical necessary conditions that must be sat-
ised by an open-loop equilibrium solution (while discussing their validity in relation with
closed-loop equilibrium solutions), and (iii) obtain the saddle-point equilibrium trajectories
through the joint use of a genetic algorithm and of the semi-DCNLP.
2 Problem Denition
The problem of the optimal interception of an optimally evasive orbital target consists in
the determination of the saddle-point equilibrium trajectories of the two space vehicles in-
volved in the combat scenario. Termination of the game occurs when the pursuing vehicle
(henceforth denoted with P) reaches the instantaneous position of the evading target (de-
noted with E henceforward). A plausible sufcient condition ensuring that capture ends
536 Dyn Games Appl (2011) 1:534557
Fig. 1 Local horizontal plane and related angles (a); instantaneous plane of motion and related thrust an-
gles (b)
the game is presented in the next subsection. The objective function, to be minimized by P
and maximized by E, is represented by the time for interception. In this work each player
is assumed to possess complete and instantaneous information on the state of the opponent
player.
2.1 Spacecraft Dynamics
This study employs a point-mass model to describe the three-dimensional motion of the two
space vehicles involved in the orbital game. The problem is investigated under the following
assumptions:
(a) aerodynamic forces are neglected, due to the altitudes involved in the cases that are
being considered;
(b) both spacecraft employ their maximum thrust for the entire time of ight;
(c) the two space vehicles are given modest propulsive capabilities;
(d) at the initial time t
0
, which is set to 0, the dynamical state of the two spacecraft is
specied.
Hypotheses (b) and (c) allow assuming constant thrust-to-mass ratios for both spacecraft,
denoted with (T
P
/m
P
) and (T
E
/m
E
) for P and E, respectively. This circumstance implies
also that the control is performed through the thrust direction only.
Six scalar variables describe the dynamical state of each spacecraft in an inertial Earth-
centered reference frame: radius r
i
(i = P or E), absolute longitude
i
(measured from the
vernal axis, the axis joining the equinoxes in the ecliptic plane), latitude
i
, ight path
angle
i
, velocity v
i
, heading (or coazimuth) angle
i
(dened in Fig. 1(a)). The control is
performed with the thrust direction, identied through the two angles
i
and
i
illustrated in
Fig. 1(b); by denition, /2
i
/2. If
E
denotes the Earth gravitational parameter,
the equations of motion are
r
i
= v
i
sin
i
, (1)

i
=
v
i
cos
i
cos
i
r
i
cos
i
, (2)
Dyn Games Appl (2011) 1:534557 537

i
=
v
i
cos
i
sin
i
r
i
, (3)

i
=
v
i
cos
i
r
i
+
T
i
m
i
sin
i
cos
i
v
i


E
cos
i
r
2
i
v
i
, (4)
v
i
=
T
i
m
i
cos
i
cos
i


E
sin
i
r
2
i
, (5)

i
=
T
i
m
i
sin
i
v
i
cos
i

v
i
cos
i
sin
i
cos
i
r
i
cos
i
. (6)
The state variables (r
i
,
i
,
i
,
i
, v
i
,
i
; i =P or E) form the state vector of each spacecraft:
x
P
=
_
x
1
x
2
x
3
x
4
x
5
x
6
_
T
=
_
r
P

P

P

P
v
P

P
_
T
, (7)
x
E
=
_
x
7
x
8
x
9
x
10
x
11
x
12
_
T
=
_
r
E

E

E

E
v
E

E
_
T
(8)
whereas the control vectors are
u
P
=
_
u
1
u
2
_
T
=
_

P

P
_
T
, (9)
u
E
=
_
u
3
u
4
_
T
=
_

E

E
_
T
. (10)
Equations (1)(6) can be written in compact form as
x
P
=f
P
(x
P
, u
P
, t ) and x
E
=f
E
(x
E
, u
E
, t ) . (11)
Due to assumption (d), the initial values of the state components are specied, whereas the
remaining boundary conditions are related to interception, which concludes the game at the
terminal time t
f
at which time
r
P
(t
f
) = r
E
(t
f
) i.e. x
1f
x
7f
= 0, (12)

P
(t
f
) =
E
(t
f
) i.e. x
2f
x
8f
= 0, (13)

P
(t
f
) =
E
(t
f
) i.e. x
3f
x
9f
= 0, (14)
where x
if
= x
i
(t
f
) (i = 1, . . . , 12). The boundary conditions (12)(14) are collected in a
vector of terminal constraints :

_
x
Pf
, x
Ef
, t
f
_
= 0, (15)
where x
Pf
= [x
1f
. . . x
6f
]
T
and x
Ef
= [x
7f
. . . x
12f
]
T
. The relation (15) is associated with
the target set of the problem of interest.
2.2 Formulation of the Zero-Sum Dynamic Game
The problem of interception is modeled as a two-player zero-sum dynamic game, where the
objective function (to be minimized by P and maximized by E) is represented by the time
for interception, i.e. for satisfaction of all the terminal constraints:
J =t
f
. (16)
538 Dyn Games Appl (2011) 1:534557
A condition that ensures (at least for the cases considered in this paper) that interception
concludes the game is
T
P
m
P
>
T
E
m
E
, (17)
i.e., P has superior propulsive capabilities with respect to E. As unbounded controls are
assumed for both spacecraft, it is conjectured that in general this condition implies that
interception can occur in a nite time, regardless of the initial conditions of the two players.
This circumstance implies also that no barrier that emanates from the target set can exist for
the game at hand.
In general, for zero-sum games two feedback (or, equivalently, closed-loop) strategies,

P
and
E
, can be introduced for the two players. If a closed-loop saddle-point equilibrium
exists, the strategies

P
and

E
are in saddle-point equilibrium when
J
_

P
,
E
_
J
_

P
,

E
_
J
_

P
,

E
_
,
P

P
,
E

E
, (18)
where
P
and
E
are the sets of the admissible strategies (in the neighborhoods of

P
and

E
). At a given time t , the value V of the game is dened as the outcome of the objective
function when both players employ their optimal strategies along the optimal path in the
time interval [t, t
f
]:
V = min

P
max

E
J = max

E
min

P
J (19)
provided that the operators max and min commute.
A common assumption [1, 3] is that the state space can be divided into a number of mu-
tually disjoint regions, separated by singular surfaces. These surfaces, according to the def-
inition given by Basar [3], are the loci where (i) the equilibrium strategies are not uniquely
determined by the necessary conditions, or (ii) the value function is not continuously dif-
ferentiable, or (iii) the value function is discontinuous. In the scientic literature, some spe-
cial, structural characteristics of zero-sum games are responsible of a number of singular
surfaces. For instance, state constraints can yield afferent and universal surfaces [8]. Non-
smooth data (e.g., a discontinuous thrust) can be responsible of discontinuities in the right
hand side of the state equations and transition surfaces can arise [8]. Furthermore, control
variables that appear linearly in the dynamics equations usually yield singular surfaces of
several kinds [1]. Other more complex analytical conditions [17, 18] can generate singu-
lar surfaces. For the problem at hand none of the previously mentioned circumstances is
encountered, and the non-existence of singular surfaces is conjectured. Thus, the value V
is plausibly assumed to be continuously differentiable over the entire state space (i.e. V is
assumed to be of class C
1
over the entire state space). With this assumption (which still
eludes any rigorous mathematical proof), the optimal open-loop representations (u

P
, u

E
) of
the closed-loop strategies are introduced as
u

P
(t ) =

P
(x
P
, x
E
, t ) and u

E
(t ) =

E
(x
P
, x
E
, t ) . (20)
For each player an open-loop representation of an optimal feedback strategy is the strategy
along the saddle-point equilibrium trajectory as a function of t and of the initial state only,
under the assumption that V is of class C
1
in the region of the state space under consid-
eration. In other words, if the state is contained in a region where the value function is of
class C
1
, then the open-loop strategies become open-loop representations of feedback strate-
gies. These representations are relevant because two properties relate them to the feedback
strategies:
Dyn Games Appl (2011) 1:534557 539
(a) if one of the two players deviates from his optimal open-loop strategy, his outcome
worsens,
(b) if both players employ their own optimal open-loop strategies, then the time histories of
the optimal open-loop and of the optimal feedback strategies are identical.
It is worth remarking that the determination of open-loop representations is relevant for
zero-sum games, because it represents an essential premise for the successive synthesis of
feedback strategies. Pesch et al. [19], Breitner and Pesch [20], and Lachner et al. [21] com-
puted a relevant number of open-loop solutions and employed them for the synthesis of
feedback strategies by means of special techniques (e.g., with the use of neural networks).
In the regions where the value function exists and is continuously differentiable in t and
x, V satises the following partial differential equation, referred to as Isaacs equation:
V
t
+max

E
min

P
_
V
x
P
f
P
+
V
x
E
f
E
_
= 0. (21)
Isaacs equation is written with reference to the special (separable) form (11) of the state
equations.
With regard to the dynamic game at hand, in this context the variables (
P
,
P
) and
(
E
,
E
) represent feedback strategies for P and E, respectively (
P
= [
P

P
]
T
and
E
=
[
E

E
]
T
). Isaacs equation becomes
V
t
+max

E
min

P
_
V
1
v
P
s
P
+V
2
v
P
c
P
c
P
r
P
c
P
+V
3
v
P
c
P
s
P
r
P
+V
4
_
v
P
c
P
r
P
+
T
P
m
P
s
P
c
P
v
P


E
c
P
r
2
P
v
P
_
+V
5
_
T
P
m
P
c
P
c
P


E
s
P
r
2
P
_
+V
6
_
T
P
m
P
s
P
v
P
c
P

v
P
c
P
s
P
c
P
r
P
c
P
__
+max

E
min

P
_
V
7
v
E
s
E
+V
8
v
E
c
E
c
E
r
E
c
E
+V
9
v
E
c
E
s
E
r
E
+V
10
_
v
E
c
E
r
E
+
T
E
m
E
s
E
c
E
v
E


E
c
E
r
2
E
v
E
_
+V
11
_
T
E
m
E
c
E
c
E


E
s
E
r
2
E
_
+V
12
_
T
E
m
E
s
E
v
E
c
E

v
E
c
E
s
E
c
E
r
E
c
E
__
= 0 (22)
and holds (x, t ), since V is assumed to be of class C
1
over the entire state space. The
symbol V
j
denotes the derivative of the value function V with respect to the state component
x
j
(j = 1, . . . , 12), whereas s[] = sin[] and c[] = cos[]. It is worth noticing that for the
problem of interest the two operators max and min are interchangeable due to separability
of the dynamical system. For the same reason (22) reduces to:
V
t
+
T
P
m
P
min

P
_
V
4
sin
P
cos
P
v
P
+V
5
cos
P
cos
P
+V
6
sin
P
v
P
cos
P
_
+
T
E
m
E
max

E
_
V
10
sin
E
cos
E
v
E
+V
11
cos
E
cos
E
+V
12
sin
E
v
E
cos
E
_
+r.t. = 0, (23)
where r.t. represents the remaining terms, all of which are independent of the (feedback)
control variables
P
and
E
. Introducing the unit vector
P
by

P
=
_
V
2
5
+
_
V
4
v
P
_
2
+
_
V
6
v
P
cos
P
_
2
_

1
2
_
V
5
V
4
v
P
V
6
v
P
cos
P
_
T
(24)
540 Dyn Games Appl (2011) 1:534557
it is then relatively straightforward to nd the control
P
that minimizes the sec-
ond term of (23). In fact, if the thrust direction of P is denoted with

T
P
, then

T
P
=
[cos
P
cos
P
sin
P
cos
P
sin
P
]
T
and the second term in (23) can be rewritten as
T
P
m
P
_
_
V
4
v
P
_
2
+V
2
5
+
_
V
6
v
P
cos
P
_
2
min
u
P
_

T
P

T
P
_
. (25)
The dot product
T
P

T
P
is minimized to 1 if
cos
P
cos
P
= V
5
__
V
4
v
P
_
2
+V
2
5
+
_
V
6
v
P
cos
P
_
2
_

1
2
, (26)
sin
P
cos
P
=
V
4
v
P
__
V
4
v
P
_
2
+V
2
5
+
_
V
6
v
P
cos
P
_
2
_

1
2
, (27)
sin
P
=
V
6
v
P
cos
P
__
V
4
v
P
_
2
+V
2
5
+
_
V
6
v
P
cos
P
_
2
_

1
2
. (28)
The three relations (26)(28) lead to deriving
P
(which is constrained to [/2, /2]) and

P
as functions of the state variable x
P
and {V
j
}
j=4,5,6
. The same steps can be repeated for

E
(taking into account that max replaces min in (23)) and lead to the following relation-
ships:
cos
E
cos
E
=V
11
__
V
10
v
E
_
2
+V
2
11
+
_
V
12
v
E
cos
E
_
2
_

1
2
, (29)
sin
E
cos
E
=
V
10
v
E
__
V
10
v
E
_
2
+V
2
11
+
_
V
12
v
E
cos
E
_
2
_

1
2
, (30)
sin
E
=
V
12
v
E
cos
E
__
V
10
v
E
_
2
+V
2
11
+
_
V
12
v
E
cos
E
_
2
_

1
2
(31)
which allow obtaining
E
(constrained to [/2, /2]) and
E
as functions of the state
variable x
E
and {V
j
}
j=10,11,12
. Due to (26)(31), Isaacs equation becomes
V
t
+V
1
v
P
s
P
+V
2
v
P
c
P
c
P
r
P
c
P
+V
3
v
P
c
P
s
P
r
P
+V
4
_
v
P
c
P
r
P


E
c
P
r
2
P
v
P
_
V
5

E
s
P
r
2
P
V
6
v
P
c
P
s
P
c
P
r
P
c
P

T
P
m
P
_
_
V
4
v
P
_
2
+V
2
5
+
_
V
6
v
P
c
P
_
2
+V
7
v
E
s
E
+V
8
v
E
c
E
c
E
r
E
c
E
+V
9
v
E
c
E
s
E
r
E
+V
10
_
v
E
c
E
r
E


E
c
E
r
2
E
v
E
_
V
11

E
s
E
r
2
E
V
12
v
E
c
E
s
E
c
E
r
E
c
E
+
T
E
m
E
_
_
V
10
v
E
_
2
+V
2
11
+
_
V
12
v
E
c
E
_
2
= 0. (32)
As r
i
, v
i
, cos
i
, and cos
i
(i = P or E) never vanish and due to continuity of the partial
derivatives {V
j
}
j=1,...,12
, for the game at hand Isaacs equation holds in the entire state space.
Dyn Games Appl (2011) 1:534557 541
The partial differential equation (32) cannot be directly solved in closed form and this
circumstance prevents directly deriving the feedback control laws in the form (20). In dif-
ferential game contexts, it is a common practice [3] to employ the necessary conditions for
open-loop saddle-point strategies. Then, if the value function is of class C
1
over the region of
the state space under consideration, then the open-loop strategies become open-loop repre-
sentations of feedback strategies. This research is aimed at determining open-loop strategies,
which are conjecturally considered open-loop representations of feedback strategies, under
the reasonable assumption that for the game of interest the value function has class C
1
over
the entire state space.
The necessary conditions for open-loop saddle-point solutions involve ordinary differ-
ential equations, and can be regarded as extensions of the necessary conditions for a local
minimum that hold in optimal control theory. First, a Hamiltonian H and a function of ter-
minal conditions are introduced as
H =
T
P
f
P
+
T
E
f
E
, =t
f
+
T
, (33)
where
P
,
E
, and are the adjoint variables conjugate to the state equations (11), and
to the boundary conditions (15), respectively. For the Lagrange multipliers
P
and
E
the
following adjoint equations hold [3]:

P
=
_
H
x
P
_
T
=
_
f
P
x
P
_
T

P
, (34)

E
=
_
H
x
E
_
T
=
_
f
E
x
E
_
T

E
(35)
with the respective boundary conditions:

P
(t
f
) =
_

x
Pf
_
T
, (36)

E
(t
f
) =
_

x
Ef
_
T
. (37)
Open-loop control variables can be determined through the following pair of relations,
u

P
= argmin
u
P
H, (38)
u

E
= argmax
u
E
H (39)
that can be regarded as the extension of the Pontryagin minimum principle to dynamic
games. As the terminal time t
f
is unspecied, the following transversality condition holds:
H(t
f
) +

t
f
= 0. (40)
Equations (11), (15), and (34)(40) dene the two-point boundary value problem associated
with the zero-sum dynamic game. The unknowns are the state vectors x
P
(t ) and x
E
(t ),
the control vectors u
P
(t ) and u
E
(t ), the Lagrange multipliers
P
(t ),
E
(t ), and , and the
terminal time t
f
.
542 Dyn Games Appl (2011) 1:534557
With regard to the orbital game at hand, (34)(35) yield 12 scalar adjoint equations,
which are not reported for the sake of brevity (cf. [16]). If the subscript f denotes the
value of the corresponding variable at t
f
, and using the terminal constraints (12)(14), the
boundary conditions for the adjoint variables
P
(t ) and
E
(t ) are

1f
=
1

2f
=
2

3f
=
3

4f
=
5f
=
6f
= 0

7f
=
1

8f
=
2

9f
=
3

10f
=
11f
=
12f
= 0
(41)
or equivalently

4f
=
5f
=
6f
= 0, (42)

10f
=
11f
=
12f
= 0, (43)

1f
+
7f
= 0, (44)

2f
+
8f
= 0, (45)

3f
+
9f
= 0. (46)
Then, for the control variables the necessary conditions (38)(39) yield
_

P
_
T
= argmin
u
P
_

4
sin
P
cos
P
v
P
+
5
cos
P
cos
P
+
6
sin
P
v
P
cos
P
_
, (47)
_

E
_
T
= argmax
u
E
_

10
sin
E
cos
E
v
E
+
11
cos
E
cos
E
+
12
sin
E
v
E
cos
E
_
. (48)
These relations are formally identical to those used to determine the feedback strategies as
functions of the partial derivatives of V , with the only difference that the adjoint variables
{
j
}
j=4,5,6,10,11,12
replace {V
j
}
j=4,5,6,10,11,12
. Therefore, the optimal open-loop control laws
are given by

P
= arcsin
_

6
v
P
cos
P
__

4
v
P
_
2
+
2
5
+
_

6
v
P
cos
P
_
2
_

1
2
_
, (49)
sin

P
=

4
v
P
cos

P
__

4
v
P
_
2
+
2
5
+
_

6
v
P
cos
P
_
2
_

1
2
, (50)
cos

P
=

5
cos

P
__

4
v
P
_
2
+
2
5
+
_

6
v
P
cos
P
_
2
_

1
2
, (51)

E
= arcsin
_

12
v
E
cos
E
__

10
v
E
_
2
+
2
11
+
_

12
v
E
cos
E
_
2
_

1
2
_
, (52)
sin

E
=

10
v
E
cos

E
__

10
v
E
_
2
+
2
11
+
_

12
v
E
cos
E
_
2
_

1
2
, (53)
cos

E
=

11
cos

E
__

10
v
E
_
2
+
2
11
+
_

12
v
E
cos
E
_
2
_

1
2
. (54)
Dyn Games Appl (2011) 1:534557 543
Lastly, the transversality condition (not reported for the sake of brevity) holds, because the
terminal time t
f
is unspecied.
3 Method of Solution
The semi-direct collocation with nonlinear programming (semi-DCNLP) algorithm converts
the two-sided optimization problem into a single-objective one, by employing the analytical
necessary conditions for optimality related to one of the two players [13]. Then the semi-
DCNLP algorithm transforms the continuous optimization problem into a discrete problem,
in which the system governing equations are translated into nonlinear algebraic (constraint)
equations involving the discrete parameters. The problem thus becomes a nonlinear pro-
gramming (NLP) problem. The numerical NLP solver must be initialized with a guess or
approximate solution (of reasonably good quality) if it is to converge to an accurate open-
loop saddle-point equilibrium solution. The guess solution affects the semi-DCNLP conver-
gence. As the costate variables usually have a non-intuitive meaning, the selection of rst
attempt values for them is very challenging, especially for large problems. In this research,
as well as in other papers published in the literature (cf. [1316]), a genetic (or evolutionary)
algorithm is employed as a preprocessing technique to overcome this difculty. The use of
a genetic algorithm (GA) is intended to provide a rst attempt approximate solution to
the problem. Then this guess is employed by the semi-DCNLP algorithm to generate an
actual, accurate (open-loop) saddle-point equilibrium solution. This section describes both
the evolutionary preprocessing and the semi-DCNLP algorithm.
3.1 Genetic Algorithm Preprocessing
Genetic algorithms represent a systematic approach to providing a starting guess for the
semi-DCNLP algorithm because they do not require any a priori information about the
solution. The unknown parameters involved in the problem form an individual. A popu-
lation is composed of a large number of individuals. Each individual corresponds to a set
of values of the unknown parameters and is evaluated with respect to a given objective
(or tness) function. The starting population is randomly generated and suitable reproduc-
tion mechanismssuch as crossover, elitism, and mutation (cf. [22, 23])are employed to
improve the population generation after generation. After a specied (large) number of gen-
erations, the GA is expected to produce the best individual, which contains the parameters
associated with the optimal approximate solution to the problem. Genetic algorithms are
characterized by a poor numerical accuracy, due to the representation of parameters through
a nite number of digits. This can be ameliorated in part by using real genetic algorithms.
Yet, this property is not a limitation when they are employed as preprocessing techniques,
i.e. just to provide a reasonable guess for the subsequent use of the semi-DCNLP.
In this study, the GA preprocessing considers all the equations that form the TPBVP
associated with the zero-sum game. In particular:
each individual is composed of all the unknown values of the costate variables at the
initial time t
0
(= 0), and includes also the (unknown) time of ight:
_

i
(0)
_
i=1,...,12
; t
f
(55)
the control variables are expressed as functions of the state and costate variables through
(49)(54);
544 Dyn Games Appl (2011) 1:534557
the state equations (1)(6) and the adjoint equations for
P
(t ) and
E
(t ) are integrated
numerically for each individual;
the 13 boundary conditions (12)(14), (40), and (42)(46) are assimilated to scalar con-
straints of the form c
l
(x
Pf
, x
Pf
,
Pf
,
Pf
, t
f
) = 0 (l = 1, . . . , 13);
the following functional, related to constraint violation, represents the objective function

J for the GA, and is evaluated for each individual:

J =
9

l=1
k
l
c
2
l
(k
l
>0). (56)
It is worth remarking that the number of unknown parameters is exactly equal to the number
of constraints (i.e. 13). In this research the C package NSGA-II, developed by Deb [23], has
been employed, with the following settings: a population composed of 500 individuals, and
100 generations to select the best individual.
3.2 Semi-DCNLP Algorithm
The semi-direct collocation with nonlinear programming (semi-DCNLP) algorithm converts
the dual-sided optimization problem, formulated as zero-sum game, into a single-objective
optimization problem. It is based on the following points:
the control of the evader is found from the necessary conditions (52)(54), and can be
expressed as u
E
=u
E
(x
E
,
E
);
the control of the pursuer is found numerically;
an extended state x ( n-dimensional vector) is dened with the inclusion of the adjoint
variables of the evader
x =
_
x
T
P
x
T
E

T
E
_
T
; (57)
a new control variable, including u
P
only, is introduced: u =u
P
( m-dimensional vector).
Hence, the extended state equations for x can be formally written by taking into account the
state equations (11) and the adjoint equations (35) for
E
(t ):

x =
_
f
T
P
f
T
E

T
E
_
f
E
x
E
__
T
=

f, (58)
where f
T
E
=f
T
E
(x
E
, u
E
(x
E
,
E
, t ), t ) =f
T
E
(x
E
,
E
, t ).
The extended boundary conditions include the original boundary conditions of the prob-
lem (15) and the boundary conditions related to the adjoint variables of the evader, collected
in
EXT
:

=
_

T
EXT
_
T
= 0. (59)
The additional term
EXT
consists of the boundary conditions related to
E
only, after elim-
inating the components of from (36)(37). For the problem of interest
EXT
includes the
left hand side of (43) and (40) (after introducing (42) and (44)(46)), i.e.
EXT
has 4 com-
ponents. As an immediate consequence, the q-dimensional vector

has seven components
( q = 7).
With these steps the zero-sumgame has been converted into the following optimal control
problem:
min
u
J subject to (58) and (59). (60)
Dyn Games Appl (2011) 1:534557 545
The corresponding extended Hamiltonian is

H =

f =
T
P(e)
f
P
+
T
E(e)
f
E

T
(e)
_
f
E
x
E
_
T

E
, (61)
where

= [
T
P(e)

T
E(e)

T
(e)
]
T
. The extended terminal function

now includes also
EXT
:

=t
f
+
T

=t
f
+
T
+
T
EXT

EXT
, (62)
where = [
T

T
EXT
]
T
. The solution of the problem (60) also satises the necessary condi-
tions for an open-loop representation of a saddle-point equilibrium solution if the following
condition holds (cf. Appendix A):

EXT
= 0. (63)
The continuous problem (60) is then discretized in time through collocation [24, 25] and
solved numerically. More specically:
the time interval [t
0
, t
f
] is partitioned into N subintervals (N = 10 in this study);
in each subinterval, the state and the control variables are discretized in time (i.e. only
their values at discrete times are employed by the algorithm);
equations (58) are translated into nonlinear algebraic equations by means of high-order
quadrature rules (in this research the highly accurate GaussLobatto fth-order quadra-
ture rules, cf. [25]).
The resulting nonlinear programming problem is solved by a numerical solver (in this work
the Fortran package NPSOL [26]).
With the fth-order GaussLobatto quadrature rules, each state component is represented
by the values at the initial, at the central, and at the terminal point of each subinterval. There-
fore, the extended state x is represented by (2 nN + n) parameters (378 in this study). Each
control component is represented by the respective values at the initial, at the central, and
at terminal point of each subinterval, and also by two additional values corresponding to
two collocation points (cf. [25] for further details). Therefore, the control vector u is rep-
resented through (4 mN + m) parameters (82 in this work). The fth-order GaussLobatto
rules allow translating the continuous problem into 2 nN nonlinear constraints (360 in this
research). The NLP solver is expected to yield the optimal values of the parameters, and then
the state components are interpolated through fth-degree polynomials, which represent the
continuous accurate approximations of their optimal time histories.
4 Numerical Results
In this study canonical units have been employed; the Earth radius R
E
is the distance
unit (DU), whereas the time unit is such that the Earth gravitational parameter
E
equals
1 DU
3
/TU
2
. Hence, 1 DU = 6378.165 km and 1 TU = 806.8 sec. In canonical units,
1 DU/TU
2
1 g = 9.798 10
3
km/sec
2
.
Three problems have been considered and solved through the method described in
Sect. 3:
(a) the optimal interception of an optimally evading spacecraft by a pursuing spacecraft;
(b) the optimal interception of an optimally evading missile by a pursuing missile;
(c) the optimal interception of an optimally evading spacecraft by a pursuing missile.
546 Dyn Games Appl (2011) 1:534557
Fig. 2 Preprocessed (from the GA optimizer) and optimal (from the sDCNLP optimizer) control laws of the
pursuer (a) and of the evader (b)
In all of the cases, the saddle-point trajectories do not include any atmospheric arc, be-
cause the altitude of both players is always greater than 100 km. This means that for combat
scenarios involving missiles only the exoatmospheric phase of ight is considered, and the
dynamic game is assumed to regard only the last stage of each missile.
The numerical solver yields the values of the adjoint variables conjugate to the state
equations (at discrete times), as well as those conjugate to the boundary conditions. For all
of the problems considered in this paper the condition (63) is satised with a numerical
accuracy of 10
4
.
4.1 Spacecraft vs. Spacecraft
At the initial time t
0
, the orbital elements of the two spacecraft involved in the dynamic
game and the respective (constant) thrust-to-mass ratios are the following:
a
P
(t
0
) = 6978.165 km; e
P
(t
0
) = 0; i
P
(t
0
) = 40 deg;

P
(t
0
) = 0 deg;
P
(t
0
) =
P
(t
0
) +M
P
(t
0
) = 310 deg;
(T
P
/m
P
) = 0.02 g;
a
E
(t
0
) = 6978.165 km; e
E
(t
0
) = 0; i
E
(t
0
) = 60 deg;

E
(t
0
) = 30 deg;
E
(t
0
) =
E
(t
0
) +M
E
(t
0
) = 290 deg;
(T
E
/m
E
) = 0.01 g,
where a
i
is the semi-major axis (SMA), e
i
is the eccentricity, i
i
is the inclination,
i
is
the right ascension of the ascending node (RAAN),
i
is the argument of perigee, and M
i
is the mean anomaly (i = P or E). Figure 2 illustrates the preprocessed and the optimal
control laws of the two spacecraft. Note that for the case of the evader the GA solution for
the optimal control corresponds quite well to the nal solution but that for the pursuer the
correspondence is actually quite poor. Nonetheless the initial guess is sufcient to allow
the NLP solver to converge to a satisfactory solution. Figure 3 portrays the corresponding
saddle-point trajectories, whereas Figs. 4 and 5 show the time histories of the osculating
orbital elements. Interception occurs in 28.1 minutes at an altitude of 540.8 km.
Dyn Games Appl (2011) 1:534557 547
Fig. 3 Spacecraft vs. spacecraft:
saddle-point trajectories leading
to interception
4.2 Missile vs. Missile
At the initial time t
0
, the orbital elements of the two spacecraft involved in the dynamic
game and the respective (constant) thrust-to-mass ratios are
a
P
(t
0
) = 5000 km; e
P
(t
0
) = 0.469; i
P
(t
0
) = 150 deg;

P
(t
0
) = 150 deg;
P
(t
0
) = 88.8 deg;
M
P
(t
0
) = 117.0 deg; (T
P
/m
P
) = 0.1 g;
a
E
(t
0
) = 5000 km; e
E
(t
0
) = 0.476; i
E
(t
0
) = 30 deg;

E
(t
0
) = 0 deg;
E
(t
0
) = 90 deg;
M
E
(t
0
) = 120.3 deg; (T
E
/m
E
) = 0.05 g
these vehicles have the characteristics of missiles in that their thrust acceleration is much
larger than the vehicles of the previous example and the orbits in this case will impact the
surface of the Earth. Figure 6 illustrates the preprocessed and the optimal control laws of the
two players. Figure 7 portrays the corresponding saddle-point trajectories, whereas Figs. 8
and 9 show the time histories of the osculating orbital elements. The initial altitudes (at t
0
)
of the pursuing and of the evading missile are 300 km and 400 km, respectively. Interception
occurs in 14.1 minutes at an altitude of 877.9 km.
4.3 Missile vs. Spacecraft
At the initial time t
0
, the orbital elements of the two spacecraft involved in the dynamic
game and the respective (constant) thrust-to-mass ratios are
a
P
(t
0
) = 3939.1 km; e
P
(t
0
) = 0.798; i
P
(t
0
) = 70 deg;
548 Dyn Games Appl (2011) 1:534557
Fig. 4 Osculating orbital elements of the pursuing spacecraft
Fig. 5 Osculating orbital elements of the evading spacecraft

P
(t
0
) = 60 deg;
P
(t
0
) = 127.5 deg;
M
P
(t
0
) = 116.9 deg; (T
P
/m
P
) = 0.1 g;
a
E
(t
0
) = 6678.165 km; e
E
(t
0
) = 0; i
E
(t
0
) = 56.5 deg;

E
(t
0
) = 0 deg;
E
(t
0
) =
E
(t
0
) +M
E
(t
0
) = 39.8 deg;
(T
E
/m
E
) = 0.05 g.
Dyn Games Appl (2011) 1:534557 549
Fig. 6 Preprocessed (from the GA optimizer) and optimal (from the sDCNLP optimizer) control laws of the
pursuer (a) and of the evader (b)
Fig. 7 Missile vs. missile:
saddle-point trajectories leading
to interception
Figure 10 illustrates the preprocessed and the optimal control laws of the two players. Fig-
ure 11 portrays the corresponding saddle-point trajectories, whereas Figs. 12 and 13 show
the time histories of the osculating orbital elements. The initial altitude (at t
0
) of the pursuing
missile is 100 km. Interception occurs in 12.7 minutes at an altitude of 317.2 km.
5 Concluding Remarks
Combat scenarios involving two competing space vehicles are best modeled as zero-sum
dynamic games. Algorithms devoted to the numerical solution of optimal control problems
cannot be employed to solve directly zero-sum games. This research describes an effective
numerical method tailored to solving zero-sum games with separable dynamics: the semi-
direct collocation with nonlinear programming algorithm (semi-DCNLP). More specically,
under the assumption that they exist, this work addresses the determination of open-loop
representations of feedback saddle-point equilibrium solutions. These representations are
550 Dyn Games Appl (2011) 1:534557
Fig. 8 Osculating orbital elements of the pursuing missile
Fig. 9 Osculating orbital elements of the evading missile
sought by employing the necessary conditions that hold for open-loop strategies, under the
plausible conjecture that for the game at hand the value function is continuously differ-
entiable over the entire state space. The semi-DCNLP converts the zero-sum game into an
optimal control problem, and then solves this converted problem employing collocation. The
method under consideration has already been successfully applied to a variety of aerospace
problems [1316] and here is used to solve the problem of optimal interception of an opti-
mally evasive target by a pursuing spacecraft or missile. The two space vehicles are assumed
to start maneuvering simultaneously and each of them is supposed to possess complete and
Dyn Games Appl (2011) 1:534557 551
Fig. 10 Preprocessed (from the GA optimizer) and optimal (from the sDCNLP optimizer) control laws of
the pursuer (a) and of the evader (b)
Fig. 11 Missile vs. spacecraft:
saddle-point trajectories leading
to interception
instantaneous information on the state of the opponent player. In real life the evader is un-
likely to possess this information, which it needs to execute the optimal evasion. However,
the solution from game theory for the optimal strategy of the evading target can provide
from a practical point of viewthe worst-case-scenario faced by the pursuer, which is
very useful to know.
The semi-DCNLP requires a reasonable guess for the non-intuitive adjoint variables of
one of the two players. This guess is provided through a genetic algorithm and in the three
examples that have been solved it is occasionally found to have only a poor correspondence
to the nal, converged solution. This circumstance testies to the effectiveness and robust-
ness of the semi-DCNLP, which apparently needs only a feasible (approximate) solution as
a guess, i.e. a solution that fullls the conditions for termination with a fair accuracy.
Only a small number of cases of optimal interception are solved here. Of course the
number of possible initial conditions and thrust capabilities for the vehicles is innite, so
that even a large number of solved cases would not be much more useful. The solved cases
552 Dyn Games Appl (2011) 1:534557
Fig. 12 Osculating orbital elements of the pursuing missile
Fig. 13 Osculating orbital elements of the evading spacecraft
represent three possible combat scenarios involving missiles and spacecraft and prove the
validity and usefulness of the analysis, as well as the effectiveness of the method of solution.
Dyn Games Appl (2011) 1:534557 553
Appendix A: Formal Conversion of a Zero-Sum Game into an Optimal Control
Problem
This section has the purpose of proving that the solution of the optimal control problem (60)
yields an open-loop representation of a saddle-point equilibrium solution of the original
zero-sum game if the condition (63) holds.
The following necessary conditions for optimality are associated with the optimal control
problem under consideration:

P(e)
=
_


H
x
P
_
T
=
_
f
P
x
P
_
T

P(e)
, (64)

E(e)
=
_


H
x
E
_
T
=
_
f
E
x
E
_
T

E(e)
+
_

x
E
__
f
E
x
E
_
T

E
__
T

(e)
, (65)

(e)
=
_

E
_
T
=
_
f
E
x
E
_
T

(e)
(66)
with boundary conditions given by (59) and including also

(e)f
=
_

E(e)f
_
T
=
_

EXT

E(e)f
_
T

EXT
. (67)
If
EXT
= 0, then, due to homogeneity of (66),
(e)
= 0 t , and the Hamiltonian

H reduces
to

H =

f =
T
P(e)
f
P
+
T
E(e)
f
E
(68)
whereas the function of terminal conditions simplies to

=t
f
+
T
. (69)
These two expressions are formally identical to the corresponding expressions that hold in
the denition of the necessary conditions for an open-loop saddle-point equilibrium solu-
tion (33). As the same differential equations and boundary conditions hold for
P
and
P(e)
and for
E
and
E(e)
, these pairs of variables are identical, i.e.
P

P(e)
and
E

E(e)
.
This circumstance implies that solving the optimal control problem (60) is equivalent to
identifying an open-loop representation of a saddle-point equilibrium solution for the origi-
nal zero-sum game, provided that the condition (63) holds.
It is worth remarking that the same analytical developments can be derived if
P
is con-
sidered for inclusion in x. This means that the roles of P and E are interchangeable in this
context.
Appendix B: Relations Between Orbital Elements and State Components
State components (r
i
,
i
,
i
,
i
, v
i
,
i
) and orbital elements (a
i
, e
i
, i
i
,
i
,
i
, M
i
) represent
a set of six variables, which describe the dynamic state of each spacecraft. Once the orbital
elements are known, the state components are unequivocally determined and vice versa.
This appendix deals with the formal derivation of all the relationships needed to calculate
the state components from the orbital elements and vice versa.
554 Dyn Games Appl (2011) 1:534557
Fig. 14 Set of rotation angles
associated with R
A
First of all, it is worth remarking that the ranges where angular variables are dened are
the following:

i
<,

2

i


2
,
i
<,

2

i


2
, (70)

i
<, 0 i
i
,
i
<,
i
<. (71)
With reference to Fig. 14, the Earth-centered inertial frame is identied by ( c
1
, c
2
, c
3
): c
1
is the vernal axis and ( c
1
, c
2
) belong to the Earth equatorial plane. This frame and the or-
bital frame ( r
i
,

i
,

h
i
) (where

h
i
denotes the unit vector aligned with the specic angular
momentum h
i

=r
i
v
i
, cf. Fig. 14) are related through the rotation matrix R
A
, dened by

r
i

h
i

c
i
c
i
c
i
c
i
s
i
c
i
s
i
s
i
s
i
c
i
c
i
c
i
s
i
s
i
s
i
s
i
c
i
s
i
s
i
c
i
s
i
c s
i
c
i
c
i
s
i
s
i
c
i
c
i

. ,, .
R
A

c
1
c
2
c
3

(72)
where s[] = sin[] and c[] = cos[]. The rotation R
A
is written in terms of the angles
i
,
i
,
and
i
, and results as the composition of three elementary rotations: the rst (counterclock-
wise by the angle
i
) about axis 3, the second (clockwise by the angle
i
) about axis 2, the
third (counterclockwise by the angle
i
) about axis 1.
Similarly, with reference to Fig. 15, the orbital frame ( r
i
,

i
,

h
i
) can be obtained from the
inertial frame ( c
1
, c
2
, c
3
) through an alternative rotation R
B
, written in terms of the angles

i
, i
i
, and
i
, where
i
(=
i
+f
i
) is the argument of latitude (f
i
denotes the true anomaly):

r
i

h
i

c
i
c
i
s
i
ci
i
s
i
c
i
s
i
+s
i
ci
i
c
i
s
i
si
i
s
i
c
i
c
i
ci
i
s
i
c
i
ci
i
c
i
s
i
s
i
c
i
si
i
si
i
s
i
si
i
c
i
ci
i

. ,, .
R
B

c
1
c
2
c
3

. (73)
This rotation results as the composition of three elementary rotations: the rst (counter-
clockwise by the angle
i
) about axis 3, the second (counterclockwise by the angle i
i
)
Dyn Games Appl (2011) 1:534557 555
Fig. 15 Set of rotation angles
associated with R
B
about axis 1, the third (counterclockwise by the angle
i
) about axis 3. The two matrices
R
A
and R
B
must coincide and this fact implies that the corresponding elements must be
identical. As a result, one obtains
cos
i
cos
i
= cos
i
cos
i
sin
i
cos i
i
sin
i
, (74)
cos
i
sin
i
= cos
i
sin
i
+sin
i
cos i
i
cos
i
, (75)
sin
i
= sin
i
sini
i
, (76)
sin
i
cos
i
= cos
i
sini
i
, (77)
cos
i
cos
i
= cos i
i
, (78)
cos
i
sin
i
cos
i
+sin
i
sin
i
= sini
i
sin
i
, (79)
cos
i
sin
i
sin
i
sin
i
cos
i
= sini
i
cos
i
. (80)
B.1
If the orbital elements (a
i
, e
i
, i
i
,
i
,
i
, M
i
) are specied, the state components (r
i
, v
i
,
i
,

i
,
i
,
i
) can be deduced in the following fashion. First of all, the numerical solution of
Keplers equation produces the eccentric anomaly E
i
:
M
i
=E
i
e
i
sinE
i
E
i
(81)
which is directly related to the true anomaly f
i
through the well-known formulas:
sinf
i
=
sinE
i
_
1 e
2
i
1 e
i
cos E
i
and cos f
i
=
cos E
i
e
i
1 e
i
cos E
i
. (82)
The polar equation of elliptic orbits yields the radius r
i
:
r
i
=
a
i
(1 e
2
i
)
1 +e
i
cos f
i
. (83)
Then, from the vis viva equation [27] one obtains the velocity
v
i
=
_
2
E
r
i


E
a
i
. (84)
556 Dyn Games Appl (2011) 1:534557
The ight path angle
i
can be deduced from the radial component of velocity, v
ri
:
v
ri
=
_

E
a
i
(1 e
2
i
)
e
i
sinf
i
=v
i
sin
i

i
. (85)
Finally, the three angles (
i
,
i
,
i
) can be calculated from (
i
, i
i
,
i
) through the relation-
ships (74)(78).
B.2
If the state components (r
i
, v
i
,
i
,
i
,
i
,
i
) are specied, the instantaneous (osculating) or-
bital elements (a
i
, e
i
, i
i
,
i
,
i
, M
i
) can be deduced in the following fashion. First of all,
from the vis viva equation one obtains the SMA a
i
:
a
i
=

E
r
i
2
E
r
i
v
2
i
. (86)
In terms of the state components, the specic angular momentum magnitude

h
i
is written as

h
i
=r
i
v
i
cos
i
. However this quantity is also:

h
i
=
_

E
a
i
(1 e
2
i
). Hence, the eccentricity
is given by:
e
i
=
_
1
(r
i
v
i
cos
i
)
2

E
a
i
. (87)
The true anomaly f
i
can be found by considering the polar equation of the ellipse:
r
i
=
a
i
(1 e
2
i
)
1 +e
i
cos f
i
cos f
i
=
a
i
(1 e
2
i
) r
i
r
i
e
i
(88)
in conjunction with the radial component of velocity:
v
ri
=
_

E
a
i
(1 e
2
i
)
e
i
sinf
i
=v
i
sin
i
sinf
i
=
v
i
sin
i
e
i
_
a
i
(1 e
2
i
)

E
. (89)
The counterparts of the relationships (82) yield the eccentric anomaly E
i
:
sinE
i
=
sinf
i
_
1 e
2
i
1 +e
i
cos f
i
and cos E
i
=
cos f
i
+e
i
1 +e
i
cos f
i
. (90)
Then, one can obtain the mean anomaly M
i
= E
i
e
i
sinE
i
. The three angles (
i
, i
i
,
i
)
can be calculated from (
i
,
i
,
i
) through the relationships (76)(80). Finally, once
i
and
f
i
are known, the argument of perigee
i
is simply
i
=
i
f
i
.
References
1. Isaacs R (1965) Differential games. Wiley, New York
2. Bryson AE, Ho YC (1975) Applied optimal control. Hemisphere, New York
3. Basar T, Olsder GJ (1999) Dynamic noncooperative game theory. SIAM, Philadelphia
4. Breakwell JV, Merz AW (1977) Minimum required capture radius in a coplanar model of the aerial
combat problem. AIAA J 15(8):10891094
Dyn Games Appl (2011) 1:534557 557
5. Guelman M, Shinar J, Green A (1990) Qualitative study of a planar pursuit evasion game in the atmo-
sphere. J Guid Control Dyn 13(6):11361142
6. Hillberg C, Jrmark B (1983) Pursuit-evasion between two realistic aircraft. AIAA Atmospheric Flight
Mechanics Conference, Gatlinburg, Paper AIAA-83-2119
7. Jrmark B, Merz AW, Breakwell JV (1981) The variable speed tail-chase aerial combat problem. J Guid
Control Dyn 4(3):323328
8. Breitner MH, Pesch HJ, Grimm W (1993) Complex differential games of pursuit-evasion type with state
constraints, Part 1: Necessary conditions for open-loop strategies. J Optim Theory Appl 78(3):419441
9. Breitner MH, Pesch HJ, Grimm W (1993) Complex differential games of pursuit-evasion type with state
constraints, Part 2: Numerical computation of open-loop strategies. J Optim Theory Appl 78(3):443463
10. Raivio T, Ehtamo H (2000) Visual aircraft identication as a pursuit-evasion game. J Guid Control Dyn
23(4):701708
11. Anderson GM, Grazier VW (1975) A closed-form solution for the barrier in pursuit-evasion problems
between two low thrust orbital spacecraft and its application. In: Aerospace sciences meeting, Pasadena,
CA, January 1975
12. Kelley HJ, Cliff EM, Lutze FH (1981) Pursuit-evasion in orbit. J Astronaut Sci 29:277288
13. Horie K, Conway BA (2006) Optimal ghter pursuit-evasion maneuvers found via two-sided optimiza-
tion. J Guid Control Dyn 29(1):105112
14. Horie K, Conway BA (2004) Genetic algorithm pre-processing for numerical solution of differential
games problems. J Guid Control Dyn 27(6):10751078
15. Pontani M, Conway BA (2008) Optimal interception of evasive missile warheads: numerical solution of
the differential game. J Guid Control Dyn 31(4):11111122
16. Pontani M, Conway BA (2009) Numerical solution of the three-dimensional orbital pursuit-evasion
game. J Guid Control Dyn 32(2):474487
17. Cardialaguet P, Quincampoix M, Saint-Pierre P (1995) Numerical methods for optimal control and dif-
ferential games. Ceremade CNRS URQA 749, University of ParisDauphine, Paris, France
18. Cardialaguet P, Quincampoix M, Saint-Pierre P (1999) Set-valued numerical analysis for optimal control
and differential games. In: Bardi M, Raghavan TES, Parthasarathy T (eds) Stochastic and differential
games: theory and numerical methods. Annals of the international society of dynamic games. Birkhuser,
Boston, pp 177247
19. Pesch HJ, Gabler I, Miesbach S (1995) Synthesis of optimal strategies for differential games by neural
networks. In: Olsder GJ (ed) New trends in dynamic games and applications. Annals of the international
society of dynamic games. Birkhuser, Boston, pp 111141
20. Breitner MH, Pesch HJ (1994) Reentry trajectory optimization under atmospheric uncertainty as a dif-
ferential game. In: Basar T, Haurie A (eds) Advances in dynamic games and applications. Annals of the
international society of dynamic games. Birkhuser, Boston, pp 7086
21. Lachner R, Breitner MH, Pesch HJ (2000) Differential game, numerical solution, and synthesis of strate-
gies. In: Filar JR, Gaitsgory V, Mizukami K (eds) Advances in dynamic games and applications. Annals
of the international society of dynamic games. Birkhuser, Boston, pp 115135
22. Goldberg DE (1989) Genetic algorithms in search, optimization, and machine learning. Addison-Wesley,
Boston
23. Deb K (2001) Multi-objective optimization using evolutionary algorithms. Wiley, Chichester
24. Hargraves CR, Paris SW (1987) Direct trajectory optimization using nonlinear programming and collo-
cation. J Guid Control Dyn 10(4):338342
25. Herman AL, Conway BA (1996) Direct optimization using collocation based on high-order Gauss
Lobatto quadrature rules. J Guid Control Dyn 19(3):592599
26. Gill PE, Murray W, Saunders MA, Wright MH (1986) Users guide for NPSOL (Version 4.0): A Fortran
package for nonlinear programming, SOL 86-2, Stanford University
27. Prussing JE, Conway BA (1993) Orbital mechanics. Oxford University Press, New York

S-ar putea să vă placă și