Skip to content

GIST-DSLab/LDCQ

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

57 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Latent Diffusion Constrained Q-Learning (LDCQ)

Training and visualizing of diffusion models from Reasoning with Latent Diffusion in Offline Reinforcement Learning (NeurIPS 2023).

Installation

Install requirement.txt

pip install --upgrade pip
pip install -r requirement.txt

Requirements (pip)

tqdm
matplotlib
wandb
ipdb
arcle == 0.2.5

Dataset

추가하기!

Training

Training Code for ARCLE Environment

cd training
  1. Training skill with:
./gpu0_train_1_skill_model.sh
  1. Collect data to train diffusion model with:
./gpu0_train_2_collect_diffusion_data.sh
  1. Training diffusion model with:
./gpu0_train_3_diffusion.sh
  1. Collect data to train offline Q-learning with:
./gpu0_train_4_collect_q_learning.sh
  1. Training Q-network with:
./gpu0_train_5_q_learning.sh

Test

cd ../eval/
./gpu0_test_ARCLE.sh

Reference

@inproceedings{ldcq,
  title = {Reasoning with Latent Diffusion in Offline Reinforcement Learning},
  author = {Siddarth Venkatraman, Shivesh Khaitan, Ravi Tej Akella, John Dolan, Jeff Schneider, Glen Berseth},
  booktitle = {Conference on Neural Information Processing Systems},
  year = {2023},
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published