-
Notifications
You must be signed in to change notification settings - Fork 66
Open
Description
I understand that for some reasons you might not have been able to release your complete code but I would highly appreciate if you could help me answering some questions about your implementation.
- The validation set on server, how much data it has and is it taken from original training (before partitioning) or test set?
- Do you train your DQN network with one optimization step after each communication round (after pushing the latest experience into replay memory) or multiple steps? Do you wait for the memory to collect some experience or train DQN even with 1 entry? What is the DQN training batch size?
- what is the optimization algorithm and learning rate used to train the DQN network?
- What is the frequency of updating the target network (from the learning DQN)?
- do you use learning rate decay as in FedAvg? Does it match their numbers?
- Do you use a discounting factor for reward (\gamma in your paper)?
Thank you in advance!
Metadata
Metadata
Assignees
Labels
No labels