-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Train new policy from scratch #21
Comments
Hi - you could use the current policy as here's an example of an RL training script. the snippet in the documentation you linked would be a way to instantiate the environment, but you could otherwise train the policy using the more standard RL pipeline shown in the stable baselines link. |
Thank you for your answer @mfe7, |
The CADRL training scripts haven't been maintained since ~2017 and the CADRL code is just meant to be available for inference. We do have a separate repo for training policies using GA3C-CADRL https://github.com/mit-acl/rl_collision_avoidance in case that is helpful |
Hello @mfe7 ,
In these days I am trying to train a new RL policy, following the documentation. I am not able to execute a train because I don't know what I need to use to execute the following line:
rl_action = model.sample(obs)
How should I replace
model
? There is an example file that I can follow to achieve this goal? Please, I need help because I am working on a course project at my university.The text was updated successfully, but these errors were encountered: