-
Notifications
You must be signed in to change notification settings - Fork 96
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What is the default setting (e.g., total training steps, learning rate) of DQN? #14
Comments
Hi, If you want to know exactly where the default hyper-parameters come from, they are stored in an individual Python script as |
We will release new version of RLzoo with much more explicit hyper-parameters configuration process soon! |
Many thanks for the clarification. Now it's more clear for me. BTW, if you can provide some benchmark policies (just like what have been done in stablebaselines zoo) that are well tuned, that will be so great. In that way, the toolbox can be treated as a standard initialization for some research directions (e.g., offline RL and adversarial robustness). If you already have some policies well trained on Freeway, BankHeist, Boxing et al., could you please share it? |
Thanks for your suggestions. |
Hi,
In your code, the training parameter setting is imported from utils.
from rlzoo.common.utils import call_default_params
May I check is there any document that explains what is this default setting and how to you fix it?
The text was updated successfully, but these errors were encountered: