Skip to content

feat(tj): update unizero ppo implementation#473

Open
tAnGjIa520 wants to merge 11 commits intoopendilab:mainfrom
tAnGjIa520:unizero-ppo-updates
Open

feat(tj): update unizero ppo implementation#473
tAnGjIa520 wants to merge 11 commits intoopendilab:mainfrom
tAnGjIa520:unizero-ppo-updates

Conversation

@tAnGjIa520
Copy link
Contributor

Main Changes

  • Replace custom GAE implementation with ding library's gae function
  • Fix data duplication issue when concatenating Segment data
  • Fix length inconsistency between advantage_segment and reward_segment
  • Optimize Buffer sample logic to prevent out-of-range errors
  • Add online training config and implementation plan documentation

主要改动

  • 使用 ding 库的 gae 函数替代自定义 GAE 实现
  • 修复 Segment 数据拼接时的重复问题
  • 修复 advantage_segmentreward_segment 长度不一致问题
  • 优化 Buffer sample 逻辑,避免超出范围
  • 新增在线训练配置和实现方案文档

xiongjyu and others added 11 commits August 27, 2025 13:11
- Replace manual GAE computation with ding.rl_utils.gae_data and gae
- Keep original implementation as _batch_compute_gae_for_pool_bak for backup
- Add test script to verify GAE computation correctness
- Fix lunarlander_env.py to handle both int and numpy array actions
- Add lunarlander_disc_unizero_ppo_config.py for PPO training
Co-authored-by: Cursor <cursoragent@cursor.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants