Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Control Synthesis from Linear Temporal Logic Specifications using Model-Free Reinforcement Learning (1909.07299v2)

Published 16 Sep 2019 in cs.RO, cs.AI, and cs.LG

Abstract: We present a reinforcement learning (RL) framework to synthesize a control policy from a given linear temporal logic (LTL) specification in an unknown stochastic environment that can be modeled as a Markov Decision Process (MDP). Specifically, we learn a policy that maximizes the probability of satisfying the LTL formula without learning the transition probabilities. We introduce a novel rewarding and path-dependent discounting mechanism based on the LTL formula such that (i) an optimal policy maximizing the total discounted reward effectively maximizes the probabilities of satisfying LTL objectives, and (ii) a model-free RL algorithm using these rewards and discount factors is guaranteed to converge to such policy. Finally, we illustrate the applicability of our RL-based synthesis approach on two motion planning case studies.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Alper Kamil Bozkurt (9 papers)
  2. Yu Wang (939 papers)
  3. Michael M. Zavlanos (65 papers)
  4. Miroslav Pajic (59 papers)
Citations (115)