Attention ensemble mixture: a novel offline reinforcement learning algorithm for autonomous vehicles

Research output: Contribution to journalArticlepeer-review

Abstract

Offline Reinforcement Learning (RL), which optimizes policies from previously collected datasets, is a promising approach for tackling tasks where direct interaction with the environment is infeasible due to high risk or cost of errors, such as autonomous vehicle (AV) applications. However, offline RL faces a critical challenge: extrapolation errors arising from out-of-distribution (OOD) data. In this paper, we propose Attention Ensemble Mixture (AEM), a novel offline RL algorithm that leverages ensemble learning and an attention mechanism. Ensemble learning enhances the confidence of Q-function predictions, while the attention mechanism evaluates the uncertainty of selected actions. By assigning appropriate attention weights to each Q-head, AEM effectively down-weights OOD actions and up-weights in-distribution actions. We further introduce three key improvements to enhance the robustness and generality of AEM: attention-weighted Bellman backups, KL divergence regularization, and delayed attention updates. Extensive comparative experiments demonstrate that AEM outperforms several state-of-the-art ensemble offline RL algorithms, while ablation studies underscore the significance of the proposed enhancements. In AV tasks, AEM exhibits superior performance compared to other methods, excelling in both offline and online evaluations.

Original languageEnglish
Article number508
JournalApplied Intelligence
Volume55
Issue number6
DOIs
Publication statusPublished - 1 Apr 2025

Keywords

  • Attention
  • Autonomous vehicle
  • Deep q-learning
  • Ensemble learning
  • Offline RL

Fingerprint

Dive into the research topics of 'Attention ensemble mixture: a novel offline reinforcement learning algorithm for autonomous vehicles'. Together they form a unique fingerprint.

Cite this