This repository contains code for the following return transformation methods for Upside Down RL and Decision Transformer:
- Install
stochastic_offline_envsfrom here. - Install dependencies with
pip install -r requirements.txt. - Make sure pytorch >= 1.10 is installed.
- Install the package with
pip install -e .. - Install the included
decision_transformerpackage. This is only necessary if you want to use the transformed returns with the included modified decision transformer implementation.
- Run
download_esper_datasets.pyto save thestochastic_offline_envsdatasets in a format that Decision Transformer understands. - Use the
--rtg path/to/returnsflag to use the generated returns or leave it out to use the original returns.
return_transforms operates on offline RL datasets. It saves a file with the transformed returns in the specified directory.
To use return_transforms on a dataset, run the following command:
python return_transforms/generate.py --env_name tfe --config configs/esper/tfe.yaml --device cuda --n_cpu 10 --ret_file data/tfe.ret
Then, you can use the included fork of Decision Transformer (in the decision_transformer directory) to train on the transformed returns.
python experiment.py --env tfe --dataset default -w True --max_iters 2 --num_steps_per_iter 25000 --rtg ../data/tfe.ret
Configurations are included for all included stochastic_offline_envs in the configs/esper directory.