menu
techminis

A naukri.com initiative

google-web-stories
Home

>

ML News

>

Learning T...
source image

Arxiv

2d

read

296

img
dot

Image Credit: Arxiv

Learning The Minimum Action Distance

  • This paper introduces a framework for learning a state representation for Markov decision processes solely from state trajectories.
  • The framework does not require reward signals or actions executed by the agent.
  • The proposed framework focuses on learning the minimum action distance (MAD), which is the minimum number of actions needed to move between states.
  • MAD serves as a fundamental metric capturing the environment's structure and assists in goal-conditioned reinforcement learning and reward shaping.
  • The self-supervised learning approach constructs an embedding space where the distances between states correspond to their MAD.
  • The approach is evaluated on various environments with known MAD values, including deterministic and stochastic dynamics, discrete and continuous state spaces, and noisy observations.
  • Empirical results show that the proposed method efficiently learns accurate MAD representations and outperforms existing state representation methods in terms of quality.

Read Full Article

like

17 Likes

For uninterrupted reading, download the app