Pretrained Visual Representations in Reinforcement Learning
Visual reinforcement learning (RL) has made significant progress in recent years, but the choice of visual feature extractor remains a crucial design decision. This paper compares the performance of RL algorithms that train a convolutional neural network (CNN) from scratch with those that utilize pre-trained visual representations (PVRs). We evaluate the Dormant Ratio Minimization (DRM) algorithm, a state-of-the-art visual RL method, against three PVRs: ResNet18, DINOv2, and Visual Cortex (VC). We use the Metaworld Push-v2 and Drawer-Open-v2 tasks for our comparison. Our results show that the choice of training from scratch compared to using PVRs for maximising performance is task-dependent, but PVRs offer advantages in terms of reduced replay buffer size and faster training times. We also identify a strong correlation between the dormant ratio and model performance, highlighting the importance of exploration in visual RL. Our study provides insights into the trade-offs between training from scratch and using PVRs, informing the design of future visual RL algorithms.
Funding
EPSRC Centre for Doctoral Training in Agri-Food Robotics: AgriFoRwArdS
Engineering and Physical Sciences Research Council
Find out more...History
School affiliated with
- School of Engineering and Physical Sciences
Publication Title
Towards Autonomous Robotic Systems 25th Annual Conference, TAROS 2024, London, UK, August 21–23, 2024, Proceedings (ed. by Md Nazmul Huda, Mingfeng Wang, Tatiana Kalganova)Publisher
Springer ChamExternal DOI
ISSN
0302-9743eISSN
1611-3349ISBN
978-3-031-72058-1eISBN
978-3-031-72059-8Date Accepted
2024-06-21Date of Final Publication
2024-11-03Event Name
Towards Autonomous Robotic Systems: 25th Annual Conference, TAROS 2024Event Dates
21-23 August 2024Event Organiser
Brunel University LondonOpen Access Status
- Not Open Access