Skip to content

Commit fb24992

Browse files
committed
v0.5.1
1 parent abdf68a commit fb24992

File tree

4 files changed

+24
-4
lines changed

4 files changed

+24
-4
lines changed

CHANGELOG

+20
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,23 @@
1+
2024.02.04(v0.5.1)
2+
- env: add MADDPG pettingzoo example (#774)
3+
- env: polish NGU Atari configs (#767)
4+
- env: fix bug in cliffwalking env (#759)
5+
- env: add PettingZoo replay video demo
6+
- env: change default max retry in env manager from 5 to 1
7+
- algo: add QGPO diffusion-model related algorithm (#757)
8+
- algo: add HAPPO multi-agent algorithm (#717)
9+
- algo: add DreamerV3 + MiniGrid adaption (#725)
10+
- algo: fix hppo entropy_weight to avoid nan error in log_prob (#761)
11+
- algo: fix structured action bug (#760)
12+
- algo: polish Decision Transformer entry (#754)
13+
- algo: fix EDAC policy/model bug
14+
- fix: env typos
15+
- fix: pynng requirements bug
16+
- fix: communication module unittest bug
17+
- style: polish policy API doc (#762) (#764) (#768)
18+
- style: add agent API doc (#758)
19+
- style: polish torch_utils/utils API doc (#745) (#747) (#752) (#755) (#763)
20+
121
2023.11.06(v0.5.0)
222
- env: add tabmwp env (#667)
323
- env: polish anytrading env issues (#731)

README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@
3535
[![Hugging Face](https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Models-yellow)](https://huggingface.co/OpenDILabCommunity)
3636
[![Open in OpenXLab](https://cdn-static.openxlab.org.cn/header/openxlab_models.svg)](https://openxlab.org.cn/models?search=opendilab)
3737

38-
Updated on 2023.12.05 DI-engine-v0.5.0
38+
Updated on 2024.02.04 DI-engine-v0.5.1
3939

4040

4141
## Introduction to DI-engine
@@ -51,7 +51,7 @@ It provides **python-first** and **asynchronous-native** task and middleware abs
5151
- Offline RL algorithms: BCQ, CQL, TD3BC, Decision Transformer, EDAC, Diffuser, Decision Diffuser, SO2
5252
- Model-based RL algorithms: SVG, STEVE, MBPO, DDPPO, DreamerV3, MuZero
5353
- Exploration algorithms: HER, RND, ICM, NGU
54-
- LLM + RL Algorithms: PPO-max, DPO, MPDPO
54+
- LLM + RL Algorithms: PPO-max, DPO, MODPO,PromptPG
5555
- Other algorithms: such as PER, PLR, PCGrad
5656

5757
**DI-engine** aims to **standardize different Decision Intelligence environments and applications**, supporting both academic research and prototype applications. Various training pipelines and customized decision AI applications are also supported:

conda/meta.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
{% set data = load_setup_py_data() %}
22
package:
33
name: di-engine
4-
version: v0.5.0
4+
version: v0.5.1
55

66
source:
77
path: ..

ding/__init__.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
import os
22

33
__TITLE__ = 'DI-engine'
4-
__VERSION__ = 'v0.5.0'
4+
__VERSION__ = 'v0.5.1'
55
__DESCRIPTION__ = 'Decision AI Engine'
66
__AUTHOR__ = "OpenDILab Contributors"
77
__AUTHOR_EMAIL__ = "[email protected]"

0 commit comments

Comments
 (0)