Projects per year
Abstract
Inspired by the recent success of sequence modeling in RL and the use of
masked language model for pre-training, we propose a masked model for
pre-training in RL, RePreM (Representation Pre-training with Masked
Model), which trains the encoder combined with transformer blocks to
predict the masked states or actions in a trajectory. RePreM is simple
but effective compared to existing representation pre-training methods
in RL. It avoids algorithmic sophistication (such as data augmentation
or estimating multiple models) with sequence modeling and generates a
representation that captures long-term dynamics well. Empirically, we
demonstrate the effectiveness of RePreM in various tasks, including
dynamic prediction, transfer learning, and sample-efficient RL with both
value-based and actor-critic methods. Moreover, we show that RePreM
scales well with dataset size, dataset quality, and the scale of the
encoder, which indicates its potential towards big RL models.
Original language | English |
---|---|
Pages (from-to) | 6879-6887 |
Number of pages | 9 |
Journal | Proceedings of the AAAI Conference on Artificial Intelligence |
Volume | 37 |
Issue number | 6 |
DOIs | |
Publication status | Published - 27 Jun 2023 |
Event | 37th AAAI Conference on Artificial Intelligence, AAAI 2023 - Washington, United States Duration: 7 Feb 2023 → 14 Feb 2023 |
Fingerprint
Dive into the research topics of 'RePreM: representation pre-training with masked model for reinforcement learning'. Together they form a unique fingerprint.Projects
- 1 Finished
-
DE21 : Scalable and Deep Anomaly Detection from Big Data with Similarity Hashing
1/01/21 → 31/12/23
Project: Research