Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Tricks for fine-tuning the hyperparameters #32

Open
liushunyu opened this issue Mar 17, 2021 · 1 comment
Open

Tricks for fine-tuning the hyperparameters #32

liushunyu opened this issue Mar 17, 2021 · 1 comment

Comments

@liushunyu
Copy link

Hello! I am trying to implement other reinforcement learning method to deal with the l2rpn problem, but I find that my result cannot match the performance with the DQN implementation in l2rpn-baselines.

Even I use the other library to implement the DQN, when I change the hyperparameters a little, I cannot get the reasonable result.

So I want to know how you fine-tune the hyperparameters of DQN in the l2rpn problem. Is there any tricks?

@Tezirg
Copy link
Contributor

Tezirg commented Mar 17, 2021

Hello,

Fine tuning NN hyperparameters is a whole research area and there is no definitive answer.
For the DQN, it has been done by trial and error, alongside rigorous data tracking in both spreadsheets and tensorboard.
A good starting point is also to use the hyperparameters of the reference literature.

Finally, I would consider the DQN more like an example of how to get started with L2RPN+NN approach, not like a really good performing baseline. This is because the L2RPN challenge is more about solving the increasing complexity as the grid size grows from 10, to 100, 1000, and 10000 nodes; which a DQN cannot handle with its 1:1 mapping of output to actions.

Best

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants