Passer à la navigation principale Passer à la recherche Passer au contenu principal

Attention ensemble mixture: a novel offline reinforcement learning algorithm for autonomous vehicles

Résultats de recherche: Contribution à un journalArticleRevue par des pairs

Résumé

Offline Reinforcement Learning (RL), which optimizes policies from previously collected datasets, is a promising approach for tackling tasks where direct interaction with the environment is infeasible due to high risk or cost of errors, such as autonomous vehicle (AV) applications. However, offline RL faces a critical challenge: extrapolation errors arising from out-of-distribution (OOD) data. In this paper, we propose Attention Ensemble Mixture (AEM), a novel offline RL algorithm that leverages ensemble learning and an attention mechanism. Ensemble learning enhances the confidence of Q-function predictions, while the attention mechanism evaluates the uncertainty of selected actions. By assigning appropriate attention weights to each Q-head, AEM effectively down-weights OOD actions and up-weights in-distribution actions. We further introduce three key improvements to enhance the robustness and generality of AEM: attention-weighted Bellman backups, KL divergence regularization, and delayed attention updates. Extensive comparative experiments demonstrate that AEM outperforms several state-of-the-art ensemble offline RL algorithms, while ablation studies underscore the significance of the proposed enhancements. In AV tasks, AEM exhibits superior performance compared to other methods, excelling in both offline and online evaluations.

langue originaleAnglais
Numéro d'article508
journalApplied Intelligence
Volume55
Numéro de publication6
Les DOIs
étatPublié - 1 avr. 2025

Empreinte digitale

Examiner les sujets de recherche de « Attention ensemble mixture: a novel offline reinforcement learning algorithm for autonomous vehicles ». Ensemble, ils forment une empreinte digitale unique.

Contient cette citation