The rrllib wrapper should provide a single agent environment with action_space
and observation_space
attributes
#69
Labels
enhancement
New feature or request
🐛 Bug
Many algorithms based on OpenAI gym domains expect the domain to define the
action_space
andobservation_space
attributes, and especially ray's rllib's single agent algorithms like AlphaZero.See for instance ray's rllib's AlphaZero implementation which makes use of those attributes.
The current implementation of scikit-decide's rllib wrapper provides only a multi-agent environment wrapper via
AsRLlibMultiAgentEnv
which does not define theaction_space
andobservation_space
attributes (which is fine for rllib's multi-agent environments). Therefore scikit-decide's rllib wrapper should additionally provide a single-agent environment wrapper for algorithms like rllib's AlphaZero, which defines theaction_space
andobservation_space
attributes.To Reproduce
Define a scikit-decide RL domain and pass it to ray's rllib's AlphaZero algorithm.
The following exception will be thrown when solving the domain:
AttributeError: 'AsRLlibMultiAgentEnv' object has no attribute 'action_space'
Expected behavior
No exception is thrown because an environment wrapper like
AsRLlibSingleAgentEnv
(to be defined) should define theaction_space
andobservation_space
attributes.The text was updated successfully, but these errors were encountered: