How to store entire observation sequences instead of individual transitions for a Deep Recurrent Q-Learning agent in Reinforcement Learning Toolbox?
I have a Deep Recurrent Q-Learning application using the Reinforcement Learning Toolbox in MATLAB R2022b. I would like to store entire sequences instead of individual transitions in the replay buffer for the agent. For the agent parameter "ExperienceBuffer", how should the data be constructed to then sample minibatch of sequences instead of minibatch of transitions?I have a Deep Recurrent Q-Learning application using the Reinforcement Learning Toolbox in MATLAB R2022b. I would like to store entire sequences instead of individual transitions in the replay buffer for the agent. For the agent parameter "ExperienceBuffer", how should the data be constructed to then sample minibatch of sequences instead of minibatch of transitions? I have a Deep Recurrent Q-Learning application using the Reinforcement Learning Toolbox in MATLAB R2022b. I would like to store entire sequences instead of individual transitions in the replay buffer for the agent. For the agent parameter "ExperienceBuffer", how should the data be constructed to then sample minibatch of sequences instead of minibatch of transitions? drqn, sequences, replay, buffer, minibatch MATLAB Answers — New Questions