Natural Critic-Actor Algorithm in CMDPs
- Natural Critic-Actor Algorithm is an online reinforcement learning method that reverses conventional update timescales and leverages natural policy gradients for CMDPs.
- It employs a three-timescale structure with linear value approximation, average cost estimation, and Lagrange multiplier updates to ensure both stability and constraint satisfaction.
- The algorithm achieves improved non-asymptotic sample complexity and robust performance on safety-critical tasks compared to traditional actor-critic frameworks.
The Natural Critic-Actor Algorithm is a class of online reinforcement learning algorithms for long-run average cost constrained Markov decision processes (CMDPs) that innovates on the standard actor-critic architecture by reversing the conventional update timescale ordering and incorporating natural policy gradient methods. Employing linear function approximation, online average cost estimation, and Lagrange multiplier updates for inequality constraints, this algorithm exhibits a three-timescale "critic-actor" structure and achieves improved non-asymptotic sample complexity while ensuring stability and constraint satisfaction. The design supports practical deployment in large-scale, safety-critical or constraint-dominated RL domains.
1. Algorithm Structure and Key Features
The Constrained Natural Critic-Actor (C–NCA) algorithm proceeds at each step by updating four sets of variables:
- Average cost estimate and estimates for each constraint cost,
- Critic parameter (for the linear function approximator of the differential value function),
- Actor parameter (parameterizing the stochastic policy),
- Lagrange multipliers (for inequality constraints).
The distinguishing element is the critic–actor timescale reversal: the actor (policy) parameters receive the fastest updates (same rates as average cost estimation), followed by the critic on an intermediate scale, and finally the Lagrange multipliers on the slowest scale. This ordering is formalized by diminishing step-size sequences (actor and average cost, e.g., , ), (critic, with slower than ), and (multiplier, with slowest).
Update Rules (primary recursions)
- Critic update:
where projects to a compact set, are fixed features, and is the temporally consistent TD error:
- Actor update (natural gradient):
with and the Fisher information matrix estimate:
- Average cost and constraint cost update:
Analogous updates for each constraint cost estimate .
- Lagrange multiplier:
with a projection operator.
The combination of these recursions forms a multi-timescale stochastic approximation system in which the actor's natural gradient increment leverages the curvature of the policy manifold for greater robustness, and the critic's TD error is computed with respect to a linear value-function approximator using features.
2. Function Approximation and Compatibility
The algorithm adopts linear function approximation both for the differential value function (critic) and for computing the compatible features needed in the natural gradient update:
- Differential value function: where .
- Compatible features: Actor updates use , aligning the function class of the advantage estimate with the policy gradient.
This compatibility is critical: it ensures the natural gradient step direction aligns with the true policy gradient under certain regularity conditions, yields unbiased updates, and underpins the theoretical analysis of both convergence and constraint satisfaction.
3. Convergence and Sample Complexity Analysis
A rigorous non-asymptotic analysis is provided for the convergence behavior of each component:
- Critic: For step-size choices respecting the prescribed timescale separation (e.g., , , with ), it is established that the average squared error in the critic (parameter ) admits the bound:
- Actor and Average Cost: For similar scaling, the error in the policy and the average cost estimate is shown to decay at optimized rates (e.g., ).
- Sample complexity improvement: By modifying the step-sizes with a factor (for example, ), the sample complexity of driving the mean-squared critic error below is , matching that of the best known single-timescale unconstrained approaches and improving on prior multi-timescale analyses.
- Constraint satisfaction: The projected multiplier recursions are proven to ensure, with high probability, that long-run average constraints are respected, i.e., the learned policy satisfies after sufficient iterations.
4. Comparison with Related Architectures
The C–NCA algorithm generalizes both the two-timescale critic-actor (CA) architecture (where the policy is updated at a faster rate than the value function, emulating value iteration) and the standard actor-critic (AC) approach (where the critic tracks a slow policy). The multi-timescale ordering used here inherits the critic-actor stability properties established for unconstrained discounted and average reward problems (Bhatnagar et al., 2022).
Distinguishing features compared to actor-critic include:
- Stability of natural gradient updates for fast policy improvement,
- Lower practical complexity for policy parameter tuning,
- Superior or matched sample complexity with flexible performance through step-size modification,
- Capacity for effective handling of safety or inequality constraints in CMDPs.
5. Empirical Performance and Safety-Critical Applications
Empirical evaluation on Safety–Gym benchmarks (including SafetyAntCircle1–v0, SafetyCarGoal1–v0, and SafetyPointPush1–v0) demonstrates that the C–NCA (especially the "Modified" version employing the improved step-size schedule) is competitive with, and in several cases outperforms, state-of-the-art constrained RL algorithms (C–AC, C–NAC, C–CA, C–DQN, etc.) in terms of maximizing average reward while maintaining rigorous satisfaction of constraint thresholds. These experiments validate the theoretical findings and reinforce the utility of natural critic–actor approaches for constraint-dominated, safety-oriented RL settings.
6. Mathematical Formulation and Algorithmic Details
A representative subset of the algorithm's update rules is: where denotes the projection onto for bounded multipliers, and are intermediate-step sizes, and , are fast and slow step sizes respectively.
7. Significance and Future Directions
The Natural Critic-Actor algorithm with constraint-handling and function approximation advances reinforcement learning theory by:
- Establishing the first non-asymptotic convergence and sample complexity guarantees for the critic-actor order with natural policy gradient in constrained average cost MDPs,
- Demonstrating that step-size refinement can achieve nearly optimal sample complexity, a key milestone for RL algorithms with complex constraints,
- Confirming, via benchmark tasks, that the algorithm delivers stable and competitive performance alongside strict constraint satisfaction.
Potential extensions include exploring nonlinear function approximation (e.g., neural network critics), application to large-scale multi-agent systems, and further variance reduction techniques to relax technical conditions on Markovian noise or step-size scheduling. The approach may serve as a foundation for scalable, deployable RL in high-stakes domains where constraint satisfaction and safety are paramount.