Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Policy Optimization and PPO #2424

Open
BrianPulfer opened this issue Jan 7, 2023 · 3 comments
Open

Policy Optimization and PPO #2424

BrianPulfer opened this issue Jan 7, 2023 · 3 comments

Comments

@BrianPulfer
Copy link
Contributor

Dear all,

While the book currently has a small section on Reinforcement Learning covering MDPs, value iteration, and the Q-Learning algorithm, the book still does not cover an important family of algorithms: Policy optimization algorithms.

It'd be great to include an overview of the taxonomy of algorithms as the one provided by OpenAI's spinning UP

For that, I propose that we cover Proximal Policy Optimization (PPO) since:

  • It is very popular in the ML community
  • It is a state-of-the-art algorithm
  • It is relatively easy to implement and grasp.

I have already written a medium post about it. My idea would be to use the environment used for the Q-learning algorithm to train the PPO model.

@astonzhang
Copy link
Member

@rasoolfa FYI

@rasoolfa
Copy link
Contributor

rasoolfa commented Jan 8, 2023

Hi @BrianPulfer,

Thank you so much for the note and suggestion.
I'd like to note that our goal for the first run of the RL section is to cover fundamental concepts which are essential for more advanced materials and then start discussing advanced topics.
That said, we'll release a couple of more RL notebooks in coming weeks covering deep RL including both on-policy and off-policy methods, and advanced topics.

Rasool

@BrianPulfer
Copy link
Contributor Author

Dear @rasoolfa,

Thank you for the answer. Please let me know if I can help with anything related to this, I'd love to!

Regards,
Brian

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants