[TOC]

  1. Title: Masked World Models for Visual Control 2022
  2. Author: Younggyo Seo et. al.
  3. Publish Year: 2022
  4. Review Date: Fri, Jul 1, 2022

https://arxiv.org/abs/2206.14244?context=cs.AI

https://sites.google.com/view/mwm-rl

Summary of paper

Motivation

TL:DR: Masked autoencoders (MAE) has emerged as a scalable and effective self-supervised learning technique. Can MAE be also effective for visual model-based RL? Yes! with the recipe of convolutional feature masking and reward prediction to capture fine-grained and task-relevant information.

Some key terms

Decouple visual representation learning and dynamics learning

compared to DreamerV2, the new model try to decouple visual learning and dynamics learning in model based RL.

image-20220701130107888

Early convolutional layers and masking out convolutional features instead of pixel patches

image-20220701125427531

Potential future work

We can use this model on our project