Module I·Article I·~4 min read
What is a Differential Game: History and Formulation
Introduction to Differential Games
Turn this article into a podcast
Pick voices, format, length — AI generates the audio
The Birth of the Theory: From Missiles to Economics
In the 1950s, the Cold War confronted military analysts with a new challenge: how to intercept a highly maneuverable missile? How to evade an interceptor? This is not an optimal control problem in the usual sense—the "target" has its own will and actively resists. Rufus Isaacs, working at RAND Corporation, realized: this is an entirely new class of mathematical problems. He invented the theory of differential games and wrote the classical book "Differential Games" (1965). Today, differential games are not just about missiles. It's a mathematical language for competitive markets, negotiations, autonomous vehicles, and predator-prey biology.
Key Difference from Optimal Control
In the optimal control problem: one player controls the system, minimizing the cost. Nature is "not against"—there is no opponent. In a differential game: two (or more) players control a shared system, and their goals conflict.
This makes the problem fundamentally more complex: the optimal strategy of one player depends on the strategy of the other, and vice versa. It's a "loop": you need to find strategies that are simultaneously optimal given the opponent's strategies.
Formal Formulation
Two-player zero-sum (the most studied case):
Dynamics: $\dot{x} = f(x, u, v)$, $x \in \mathbb{R}^n$ (state), $u \in U$ (control of player P, minimizer), $v \in V$ (control of player E, maximizer).
Functional: $J = g(x(T)) + \int_0^T F(x(t), u(t), v(t)) ,dt$.
Players' Goals:
- Player P (pursuer/minimizer): $\min_u \max_v J$
- Player E (evader/maximizer): $\max_v \min_u J$
Game Value $V(x_0)$ exists if Isaacs' condition is satisfied (see the next article):
$ V = \min_u \max_v J = \max_v \min_u J $
What Does “Strategy” Mean?
Unlike static games, players have three types of strategies:
Open-loop strategy: $u = u(t)$—control as a function of time only. Planned in advance, does not react to the system state. Mathematically simpler, but unrealistic in practice.
Feedback-form strategy: $u = \alpha(x, t)$—control as a function of the current state. Each player "sees" $x$ and reacts. This is a realistic model for real systems.
Output feedback strategy: $u = \alpha(y, t)$, where $y = h(x)$—incomplete observation. The most complex case.
Important Fact: for zero-sum games under Isaacs' condition, the value of the game is the same for open-loop and feedback strategies!
Classification of Games
By sum: zero-sum ($J_1 + J_2 = 0$—interests are completely opposed), nonzero-sum (each has their own $J$), cooperative (players can make agreements).
By information: full information (both see $x$), incomplete (part of $x$ is hidden).
By horizon: finite ($T < \infty$), infinite ($T = \infty$, pursuit until capture problem).
By dynamics: linear ($f = Ax + Bu + Cv$), nonlinear, stochastic (with noise).
Connection with Optimal Control Theory
When $v = 0$ (no second player), a differential game degenerates into an optimal control problem. All optimal control tools—the Pontryagin principle, Bellman dynamic programming equation—are carried over into game theory with modifications.
Pontryagin principle $\to$ Isaacs minimax principle. Bellman dynamic programming equation $\to$ Hamilton-Jacobi-Isaacs (HJI) equation.
Real-Life Examples
Aviation: interceptor (P) and target (E). The interceptor wants to minimize the distance to the target. The target wants to maximize it. Optimal strategy for P: fly toward the "anticipation point," not directly.
Economics: two companies set prices for competing products. The price of the first affects the demand for the second and vice versa. Dynamic model $\to$ differential game.
Autonomous vehicles: two cars at an intersection. Each wants to pass without crashing. This is a Stackelberg or Nash equilibrium problem in differential games.
Biology: predator and prey in three-dimensional space. Strategically, optimal "chase" is not always "in a straight line."
Historical Development
After Isaacs, the theory developed in several directions. In the 1970s-80s, Bensoussan, Friedman, Elliott-Kalton developed mathematically rigorous foundations. In the 1980s-90s, viscosity solutions of Crandall-Lions for the HJI equation resolved the existence and uniqueness of the game value. In the 2000s, the rise of mean field games (Lasry-Lions and Caines-Huang) provided tools for games with millions of players (crowd models, markets, biology). In the 2010s, connections to reinforcement learning appeared: Q-learning algorithms for Markov games, multi-agent RL.
Modern Numerical Methods
- Finite difference schemes for HJI: Lax-Friedrichs, ENO/WENO upwind, level-set (Osher-Sethian)—standard for low-dimensional problems ($n \leq 4$)
- Semi-Lagrangian method (Falcone, Ferretti): effective for problems with discontinuities
- Adaptive grids: AMR (Adaptive Mesh Refinement) for localized refinement
- Neural network approximations: Deep Galerkin, PINNs for HJI in high dimensions—a breakthrough 2018–2023
- Reach-avoid analysis: Hamilton-Jacobi reachability in libraries hj_reachability (Python), helperOC (MATLAB)
Applications
Differential games are used in autonomous transport control (Tesla, Waymo use game-theoretic planning for interaction with other drivers), in robust financial modeling (Avellaneda, Zariphopoulou), in cybersecurity (attacker vs defender), in power grids (DR agents compete for capacity).
§ Act · what next