-
Notifications
You must be signed in to change notification settings - Fork 74
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to use RL algorithms with continuous action space #49
Comments
Hi @AizazSharif , From the error logs, it looks like the DDPG's critic network's concat operation is failing to concat tensors of different rank: |
Thanks for the reply @praveen-palanisamy. I will look into it and let you know. |
I also wanted to ask whether it is possible to have one agent with discrete and another with continuous actions in a same driving scenario? @praveen-palanisamy |
Hi @AizazSharif , |
Hi @praveen-palanisamy |
Hi, |
Hi @SExpert12 I used 4 discrete (PPO DQN A3C IMPALA) and 2 continuous (TD3 DDPG) algorithms. I have usually tried two to three agents per scenario in my experiments. |
Hi, Thanks for reply |
Hi @SExpert12 Here is the training code for two agents learning independently in a three-way scenario. The testing for the trained policies above can be looked at in the following script. |
Thanks. |
Sorry @SExpert12, the links I shared were examples of multi-agent settings. You can find a continuous agent example in the following link. The agent here is trained in the presence of other NPC agents using continuous action space. Let me know if you have more questions. |
Okay. |
Hi, https://github.com/T3AS/MAD-ARL/blob/main/examples/step_1_training_victims.py 024-07-08 09:21:32.363670: W tensorflow/strea024-07-08 09:21:32.363670: W tensorflow/stream_executor/platform/default/dso_loader.cc:55] Could not load dynamic library 'libnvinfer.so.6'; dlerror: libnvinfer.so.6: cannot open shared object file: No such file or directory; LD_LIBRARY_PATH: /home/ryzen/miniconda3/envs/MAD-ARL/lib/python3.7/site-packages/cv2/../../lib64: 2024-07-08 09:21:35,507 ERROR trial_runner.py:521 -- Trial PPO_HomoNcomIndePOIntrxMASS3CTWN3-v0_00000: Error processing event. Traceback (most recent call last): 2024-07-08 09:21:35,507 ERROR trial_runner.py:521 -- Trial PPO_HomoNcomIndePOIntrxMASS3CTWN3-v0_00000: Error processing event. Traceback (most recent call last): How to solve this now? |
Hi @praveen-palanisamy
I have been working on macad-gym successfully over the past few months using PPO and many other algorithms. Now I am trying to use DDPG using RLlib which requires continuous action space.
I have changed the boolean "discrete_actions": False within environment config, but its still a issue since the policy function is passing Discrete(9) and I do not know the alternative for continuous action space.
I also followed the guide mentioned here but now its giving me the following error.
error.txt
Any help in this regard would be appreciated.
Thanks.
The text was updated successfully, but these errors were encountered: