Skip to content

MonkeScripts/PPO

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Proximal Policy Optimization (PPO) for CartPole-v1

Overview

This repository contains an implementation of PPO for solving the CartPole-v1 environment using PyTorch. The code is adapted from a tutorial with significant enhancements.

Key Features

  • Modern implementation using Gymnasium (successor to OpenAI Gym)
  • Performance monitoring with TensorBoard
  • Experiment tracking via Weights & Biases
  • Optimized hyperparameters for CartPole-v1

Demo

rl-video-step-400.mp4

Some findings

As a fellow neanderthal, I needed to make notes for myself:

Here is a notion link with some of my findings: Notion Link

Setup

Environment dependencies are managed through Conda. To get started:

conda env create -f environment.yml

About

An implementation of PPO for classic Cartpole-v1

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages