This repo generates the data for NeoRL.
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Icarus f7217196fa add description for tianshou 5 months ago
offlinedata update 5 months ago
.gitignore add data parse script 1 year ago add description for tianshou 5 months ago fix a bug 1 year ago


OfflineData is the repository to train policies and generate datasets for NeoRL benchmarks.

Install OfflineData

1. Install offlinedata

git clone
cd OfflineData
pip install -e .

2. Install neorl

Please install neorl for getting environments:

git clone
cd NeoRL
pip install -e .

3. Install tianshou

We use tianshou, a popular RL framework, to train the behavioral policies. Please install the tianshou through GitHub:

pip install git+

If you use mujoco environment, please make sure you install mujoco and mujoco_py.


You can use neorl to get all standardized environments, like:

import neorl

env = neorl.make("halfcheetah-meidum-v3")

env = neorl.make("citylearn")

You can use the following environments now:

Env Name observation shape action shape have done max timesteps
HalfCheetah-v3 18 6 False 1000
Hopper-v3 12 3 True 1000
Walker2d-v3 18 6 True 1000
ib 182 3 False 1000
finance 181 30 False 2516
citylearn 74 14 False 1000


1.Train policy

python --task env_name

The policy models labelled by trajectories total reward will be saved in models. Some models are pre-saved in the folder.

2.Sample data

You can also skip the first step and sample data use our pre-saved policy models.

(1) Deterministic policy sampling:

python --task env_name

(2) stochastic policy sampling:

python --task env_name --add_noise

The datasets will be saved in datasets.