You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It's a really nice work on benchmarking multi-agent RL algorithms. When I was running the code, I had some confusion about the parameters of the onpolicy algorithm. Especially when using MAPPO algorithm, the winning rate is always 0 during training 2s_vs_1sc and 5m_vs_6m. I tried to use mappo.yaml in the source code and the relevant parameters given in the paper, but there was no effect. Could you please provide the parameters for training the MAPPO algorithm for the two scenarios of 2s_vs_1sc and 5m_vs_6m. Thank you .
The text was updated successfully, but these errors were encountered:
It's a really nice work on benchmarking multi-agent RL algorithms. When I was running the code, I had some confusion about the parameters of the onpolicy algorithm. Especially when using MAPPO algorithm, the winning rate is always 0 during training 2s_vs_1sc and 5m_vs_6m. I tried to use mappo.yaml in the source code and the relevant parameters given in the paper, but there was no effect. Could you please provide the parameters for training the MAPPO algorithm for the two scenarios of 2s_vs_1sc and 5m_vs_6m. Thank you .
The text was updated successfully, but these errors were encountered: