Skip to content

gth828r/ppo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 

Repository files navigation

Introduction

This repository contains an implementation of PPO. Initially the plan was to explore how CQL affects PPO, but with a deeper understanding of both PPO and CQL, it appears to make more sense to explore CQL applied to the soft actor critic algorithm. The primary reason is that PPO's design was not specifically for offline RL, while SAC's is designed for offline RL. CQL is specifically a tool to use for helping generalize with offline RL.

See https://github.com/gth828r/sac-cql for exploration of CQL applied to SAC.

About

An implementation of PPO in Tensorflow

Topics

Resources

Stars

Watchers

Forks

Packages

No packages published