Skip to content

Commit

Permalink
fix ReadME.md
Browse files Browse the repository at this point in the history
  • Loading branch information
wisnunugroho21 committed Sep 13, 2020
1 parent 827593c commit eaab23a
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -60,7 +60,7 @@ There are two primary variants of PPO: PPO-Penalty and PPO-Clip.

* PPO-Clip doesn’t have a KL-divergence term in the objective and doesn’t have a constraint at all. Instead relies on specialized clipping in the objective function to remove incentives for the new policy to get far from the old policy.

This repository use PPO-Clip
OpenAI use PPO-Clip
You can read full detail of PPO in [here](https://spinningup.openai.com/en/latest/algorithms/ppo.html)

## Random Network Distillation
Expand Down

0 comments on commit eaab23a

Please sign in to comment.