You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'd like to suggest a couple features for V3, in case it hasn't been suggested already:
All RL algorithms are able to automatically normalize input features, similar to how it currently works with DDPG. I believe it is better this way, rather than forcing the user to use wrappers, such as VecNormalize.
All RL algorithms are able to wrap the given environment (or list of environments) into DummyVecEnv (or list of DummyVecEnv). Again, I believe it is better for this task to fall under the developer rather than the user.
The text was updated successfully, but these errors were encountered:
All RL algorithms are able to automatically normalize input features, similar to how it currently works with DDPG. I believe it is better this way, rather than forcing the user to use wrappers, such as VecNormalize.
I would argue against that one. In the current master version (and in v3), all algorithms now support normalization through VecNormalize wrapper. The main issue for me would be unecessary complexity and clarity of what is happening. Unless you have an elegant way to implement that feature in pytorch (I would be happy to see it).
Also,VecNormalize allows to normalize reward and is nice to have it separated from the RL algorithm (because it acts/and uses info only from the environment).
All RL algorithms are able to wrap the given environment (or list of environments) into DummyVecEnv (or list of DummyVecEnv). Again, I believe it is better for this task to fall under the developer rather than the user.
What do you mean exactly? VecEnv will be the default for v3 (cf #576#733 ). However, if you mean that all algorithms should support multiprocessing this is something else.
It would be nice to have unless it adds to much complexity. Also, it would remove the per-episode training capability for SAC or TD3 for instance, which is essential in a robotic setting.
I'd like to suggest a couple features for V3, in case it hasn't been suggested already:
All RL algorithms are able to automatically normalize input features, similar to how it currently works with DDPG. I believe it is better this way, rather than forcing the user to use wrappers, such as VecNormalize.
All RL algorithms are able to wrap the given environment (or list of environments) into DummyVecEnv (or list of DummyVecEnv). Again, I believe it is better for this task to fall under the developer rather than the user.
The text was updated successfully, but these errors were encountered: