You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I noticed that in this implementation, when calculating the actor loss, DMC tasks use the dynamics loss and atari tasks use the reinforcement loss (w/o return normalization), which should be similar to the Dreamer v2
Thank you for bringing this to my attention!
I noticed that the paper and the official implementation were updated in April of this year as shown below. paper before update, paper after update
The difference in the actor loss calculation is due to this implementation being done prior to that update.
I'll look into the differences between the two approaches, but it may take some time, so I appreciate your patience.
Hi,
Thanks for the implementation!
I noticed that in this implementation, when calculating the actor loss, DMC tasks use the dynamics loss and atari tasks use the reinforcement loss (w/o return normalization), which should be similar to the Dreamer v2
dreamerv3-torch/models.py
Lines 406 to 412 in 4e50f30
Thanks a lot!
Best
The text was updated successfully, but these errors were encountered: