Skip to content

orybkin/lexa-benchmark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

LEXA Benchmark

Codebase for the self-supervised goal reaching benchmark introduced in the LEXA paper (Discovering and Achieving Goals via World Models, NeurIPS 2021).

Setup

Create the conda environment by running : conda env create -f environment.yml

Alternatively, you can update an existing conda environment by running : conda env update -f environment.yml

Modify the python path
export PYTHONPATH=<path to lexa-benchmark>

Export the following variables for rendering
export MUJOCO_RENDERER=egl; export MUJOCO_GL=egl

Please follow these instructions to install mujoco

Usage

WARNING! The success criteria defined in this benchmark should not be used as a reward for the agent. The criteria were tuned to roughly match human intuition, but produce some false positives, which the agent can exploit if it observes the success variable.

Bibtex

If you find this code useful, please cite:

@misc{lexa2021,
    title={Discovering and Achieving Goals via World Models},
    author={Mendonca, Russell and Rybkin, Oleh and
    Daniilidis, Kostas and Hafner, Danijar and Pathak, Deepak},
    year={2021},
    Booktitle={NeurIPS}
}

Acknowledgements

This benchmark is built on top of the following environments: Adept, MetaWorld, and DeepMind Control Suite.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages