RMLGym: a Formal Reward Machine Framework for Reinforcement Learning

  • Hisham Unniyankal
  • , Francesco Belardinelli
  • , Angelo Ferrando
  • , Vadim Malvone

Research output: Contribution to journalConference articlepeer-review

Abstract

Reinforcement learning (RL) is a powerful technique for learning optimal policies from trial and error. However, designing a reward function that captures the desired behavior of an agent is often a challenging and tedious task, especially when the agent has to deal with complex and multi-objective problems. To address this issue, researchers have proposed to use higher-level languages, such as Signal Temporal Logic (STL), to specify reward functions in a declarative and expressive way, and then automatically compile them into lower-level functions that can be used by standard RL algorithms. In this paper, we present RMLGym, a tool that integrates RML, a runtime verification tool, with OpenAI Gym, a popular framework for developing and comparing RL algorithms. RMLGym allows users to define reward functions using RML specifications and then generates reward monitors that evaluate the agent’s performance and provide feedback at each step. We demonstrate the usefulness and flexibility of RMLGym by applying it to a famous benchmark problem from OpenAI Gym, and we analyze the strengths and limitations of our approach.

Original languageEnglish
Pages (from-to)1-16
Number of pages16
JournalCEUR Workshop Proceedings
Volume3579
Publication statusPublished - 1 Jan 2023
Event24th Workshop "From Objects to Agents", WOA 2023 - Roma, Italy
Duration: 6 Nov 20238 Nov 2023

Keywords

  • OpenAI Gym
  • Reinforcement Learning
  • Runtime Verification

Fingerprint

Dive into the research topics of 'RMLGym: a Formal Reward Machine Framework for Reinforcement Learning'. Together they form a unique fingerprint.

Cite this