|
actor_loss = agent.critic.forward(states, mu).flatten() |
Dear Phill,
First of all plenty of thanks and gratitude for your lessens, I've learned a lot from your lectures.
I've noticed a difference in the code at line 83 in MADDPG class while calculating actor-loss. It's running forward propagation of critic network instead of actor network. I believe this is typo, please correct me if I'm wrong.
Thanks and Regards
Viji