-
Notifications
You must be signed in to change notification settings - Fork 243
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Settings for Pong #138
Comments
As far as I know, the |
Hey, |
Thanks @IcarusWizard and @NonsansWD I'll try your suggestion! |
I have not run into that particular issue. I'd suggest that you install |
Alright good to know. In the end i was able to fix my issue and everything works fine. The only problem im left with is i just realized that the resulting folder called "replay" does not contain raw frames but instead a lot of data like rewards and so on. Do you by any chance know a way of obtaining a video of the agents steps or something so i can watch it do its stuff without too much effort? I feel like im missing something cause i also dont know where to get these wonderful score plots or do i have to construct that plot myself with matplotlib? sorry for going off topic |
@IcarusWizard @NonsansWD How did you know that the train_ratio should be 1024 for atari100k, instead of 256 as set by configs.yaml for "atari100k"? dreamerv3/dreamerv3/configs.yaml Lines 211 to 216 in 251910d
I wonder if the old version of DreamerV3 (in 2023) used train_ratio 1024, as specified in the old version of the paper (https://arxiv.org/pdf/2301.04104v1), whereas the new version of DreamerV3 (2024) uses a train_ratio of 256. However, when I attempted to reproduce Pong just now with the default config value of 256, my agent was stuck at -21 score as in #154. Old version of DreamerV3: dreamerv3/dreamerv3/configs.yaml Lines 135 to 146 in 84ecf19
|
@George614 I think I figured out the issue. The task config should be "atari100k_pong" instead of "atari_pong". #173 I am leaving the train_ratio as 256 as set by the atari100k config for now. |
The issue could also be that you're using |
@danijar Thanks so much for replying. I see, so the results reported in the paper are the final training scores after 400,000 steps? I noticed that script=train does not report a score exactly at step 400,000, since episodes are of varying length, so did you take the training score from the first step above 400,000? I also noticed that the default configs run scrip=train for 440,000 steps, so do you take the final training score after 440,000 steps instead? Also, does this mean that the scores reported in the paper are from a single training episode? Or is the reported score an average over multiple training episodes? |
The reported scores are the average episode returns within the last 10k steps, that is, all episodes that finished between 390k and 400k environment frames. The scores after that are not included in the results. |
Hi Danijar,
Thanks for sharing this amazing repo and creating a robust model-based RL algorithm! I've been playing with the replay buffer and trying to reproduce some of the results. I run the code on Pong with command
python dreamerv3/main.py --logdir ./logdir/uniform_pong --configs atari --task atari_pong --run.train_ratio 32
with the default configurations on a Ubuntu 22.04 LTS with a RTX 3090 GPU. Somehow, the agent does not work on the Pong task over 400K env steps (according to the first version of the paper). I'm not sure what went wrong. I've tried with the default uniform replay (cyan curve in figure), a mixed replay (gray curve) with ratio of (0.5, 0.3, 0.2) and uniform replay withcompute_dtype: float16
(magenta curve) since I've seen some warnings from CUDA and XLA.Here are the package versions that I installed:
Please let me know if anything was not set up properly. Thank you!
The text was updated successfully, but these errors were encountered: