Skip to content

Some doubts and questions #7

@fvarno

Description

@fvarno

I understand that for some reasons you might not have been able to release your complete code but I would highly appreciate if you could help me answering some questions about your implementation.

  1. The validation set on server, how much data it has and is it taken from original training (before partitioning) or test set?
  2. Do you train your DQN network with one optimization step after each communication round (after pushing the latest experience into replay memory) or multiple steps? Do you wait for the memory to collect some experience or train DQN even with 1 entry? What is the DQN training batch size?
  3. what is the optimization algorithm and learning rate used to train the DQN network?
  4. What is the frequency of updating the target network (from the learning DQN)?
  5. do you use learning rate decay as in FedAvg? Does it match their numbers?
  6. Do you use a discounting factor for reward (\gamma in your paper)?

Thank you in advance!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions