Markov Decision Processes (MDPs) Overview
- Markov Decision Processes (MDPs) are discrete-time stochastic models for sequential decision-making, defined by states, actions, rewards, and discount factors.
- They underpin reinforcement learning and operations research, providing frameworks for robust, risk-sensitive, and high-dimensional control problems.
- Recent developments include measurized MDPs, probabilistic constraints, and representation learning techniques that enhance policy synthesis and optimization.
A Markov Decision Process (MDP) is a discrete-time stochastic control process that models sequential decision-making under uncertainty. An MDP is defined by a state space, an action space, a transition mechanism, a reward function, and a discount factor. The aim is to synthesize policies that optimize an expected cumulative objective. MDPs form the core mathematical object in stochastic control, reinforcement learning, and operations research. The contemporary literature extends the MDP framework to incorporate general state/action spaces, robust and risk-sensitive objectives, learning from incomplete information, and domain-specific constraints.
1. Formal Structure and Canonical Properties
Let be a Borel state space and a Borel action space. The canonical, discrete-time MDP is a tuple
where
- is a Markov transition kernel: for each , , is a probability measure on ;
- is a one-stage reward;
- is a discount factor.
A policy is a measurable mapping assigning actions (possibly randomized) to each state. The value function under policy is
The optimal value function satisfies the Bellman optimality equations:
For finite and countable, these reduce to the classical Bellman recurrences. Similar expressions underly the average-reward and constrained MDPs.
2. Existence, Measurability, and Semicontinuous–Semicompact Framework
When and are general Borel spaces, issues of measurability and continuity become paramount. The semicontinuous–semicompact framework, as developed by Hernández-Lerma and Lasserre, casts the MDP in a setting where the reward is upper semicontinuous and bounded above, and the transition kernel depends continuously (in the weak topology) on . Under these and mild integrability assumptions, the Bellman operator admits fixed points in the space of bounded Borel-measurable functions; optimal measurable selectors exist, yielding stationary optimal policies. This facilitates extension to constrained and infinite-horizon models without the technicalities of universally measurable selectors (Adelman et al., 6 May 2024).
3. Lifting and the Measurized MDP Formalism
A crucial generalization is the measurized MDP, whereby the state space is lifted from points to probability measures on , formulated within the weak topology. The measurized MDP is specified by the tuple
where is the set of probability measures on , is the set of Markov decision rules (stochastic kernels), encodes deterministic transitions on via
and is the lifted one-stage reward. The Bellman equations in this setting admit the same structure, with value functions (bounded Borel-measurable functions on ), and Borel-measurable selectors guaranteeing optimal stationary product policies. For Dirac- measures, this framework recovers the original MDP, without loss of fidelity:
The framework naturally incorporates external random shocks, risk or chance constraints, and supports nuanced approximation architectures (Adelman et al., 6 May 2024).
4. Constraints and Value Function Approximations
The lifting approach enables constraints and approximations not expressible in the classical space:
- Risk constraints (e.g., CVaR): By restricting to policies satisfying
with the cost and defined in terms of the conditional tail expectation. The Bellman equation then optimizes over this risk-constrained policy set.
- Probabilistic state constraints: For instance, bounding the variance further restricts feasible , affecting the supremum in the Bellman update.
- Value function approximations: By expanding in basis functions (moments, Laplace transforms, divergences) and solving for weights in
the MDP solution reduces to a convex optimization over the function class (Adelman et al., 6 May 2024).
5. Representation Learning and State Compression
A central technical challenge in high-dimensional domains is constructing low-dimensional, Markovian feature representations. The formalism of feature Markov Decision Processes (MDPs) encodes any history compression (from history space to state space) and defines penalized code-length cost functions to score candidate . The optimal minimizes this cost, yielding an induced process that is approximately Markov (0812.4580). Alternating deep neural networks (ADNN) further automate this discovery, learning encoders such that the reduced process is itself Markov and sufficient for optimal control; conditional independence criteria (residual tests, Brownian distance covariance) ensure fidelity with the original process, and group-lasso regularization yields sparsity for interpretability (Wang et al., 2017).
6. Robust, Risk-Sensitive, and Distributionally Robust Extensions
MDPs are generalized to address model uncertainty (robust MDPs, RMDPs) and risk via recursive risk measures:
- Robust MDPs: Transition probabilities are uncertain within nonempty convex sets , yielding robust Bellman equations of the form
which is equivalent to a zero-sum stochastic game with nature (Suilen et al., 18 Nov 2024).
- Risk-sensitive MDPs: The objective is the recursive application of coherent, law-invariant risk measures ; the value function solves
Under mild contractivity, unique fixed points and optimal stationary policies exist; the dual representation of provides a direct link to distributionally robust control (Bäuerle et al., 2020).
- Distributionally robust chance-constrained MDPs: Ambiguity in the reward distribution is handled via moment, -divergence or Wasserstein sets. The robust optimization reduces to tractable SOCP/MISOCP or copositive programs, yielding policies with formal chance constraint guarantees (Nguyen et al., 2022).
7. Learning Under Partial Knowledge, Unawareness, and Nonstandard Extensions
MDPs have been extended to address partial knowledge, unawareness, non-cumulative objectives, and non-stationarity:
- Learning with Unawareness: In MDPs with unawareness (MDPUs), the agent is initially unaware of part of the action set. Discovery is modelled via an “explore” action and a stochastic discovery process . Near-optimal play is possible iff , with polynomial-time learning characterized by the growth rate of this sum (Halpern et al., 2014).
- Non-cumulative Objectives: For decision processes with non-cumulative return functions (e.g., maximizing the maximum reward over time), construction of a “lifted” MDP with an augmented state capturing the statistic of the past reward sequence enables reduction to standard RL and dynamic programming algorithms—retaining optimality (Nägele et al., 22 May 2024).
- Non-stationary or Externally Modulated MDPs: When transitions depend on external temporal processes or histories, the formalism augments the state with finite-memory histories; under suitable decay of exogenous effects, optimality is achieved via truncated history, with error bounded in terms of total variation (Ayyagari et al., 2023).
In summary, the theory of MDPs encompasses discrete- and continuous-parameter models, broad classes of uncertainties and constraints, and advanced frameworks for representation, approximation, and robust/learning-based control. Lifting to measure spaces, compositional risk constraints, and function-approximation architectures exemplify ongoing innovations (Adelman et al., 6 May 2024). The contemporary MDP is not only a canonical model for sequential stochastic decision making but also a substrate on which generalizations for robustness, risk sensitivity, high-dimensionality, unawareness, and statistical learning are rigorously built.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days free