Replies: 1 comment 5 replies
-
|
Hi @hiraz01 Currently, it is necessary to modify the source code to support different learning rate for the involved optimizers. |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Is it possible to use two separate optimizers and learning rates for the policy and value optimizers without directly changing the current PPO implementation? I looked through the source and it appears that I could change it directly, but I am not sure how best way to go about doing this. I am trying to replicate results from DeepMimic environment where they use separate LRs in tensorflow 1. Goal is to have additional flexibility and use IsaacLab with skrl.
Beta Was this translation helpful? Give feedback.
All reactions