LESSON: Learning to Integrate Exploration Strategies for Reinforcement Learning via an Option Framework
Abstract
A unified exploration framework in RL using option-critic models integrates diverse strategies for adaptive exploration-exploitation trade-offs, validated in MiniGrid and Atari environments.
In this paper, a unified framework for exploration in reinforcement learning (RL) is proposed based on an option-critic model. The proposed framework learns to integrate a set of diverse exploration strategies so that the agent can adaptively select the most effective exploration strategy over time to realize a relevant exploration-exploitation trade-off for each given task. The effectiveness of the proposed exploration framework is demonstrated by various experiments in the MiniGrid and Atari environments.
Get this paper in your agent:
hf papers read 2310.03342 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper