Poster Session 3 · Thursday, December 4, 2025 11:00 AM → 2:00 PM
#415
Reinforcement Learning with Action Chunking
Abstract
We present Q-chunking, a simple yet effective recipe for improving reinforcement learning (RL) algorithms for long-horizon, sparse-reward tasks. Our recipe is designed for the offline-to-online RL setting, where the goal is to leverage an offline prior dataset to maximize the sample-efficiency of online learning.
Effective exploration and sample-efficient learning remain central challenges in this setting, as it is not obvious how the offline data should be utilized to acquire a good exploratory policy. Our key insight is that action chunking, a technique popularized in imitation learning where sequences of future actions are predicted rather than a single action at each timestep, can be applied to temporal difference (TD)-based RL methods to mitigate the exploration challenge.
Q-chunking adopts action chunking by directly running RL in a chunked action space, enabling the agent to:
- leverage temporally consistent behaviors from offline data for more effective online exploration
- use unbiased -step backups for more stable and efficient TD learning.
Our experimental results demonstrate that Q-chunking exhibits strong offline performance and online sample efficiency, outperforming prior best offline-to-online methods on a range of long-horizon, sparse-reward manipulation tasks.