-
Notifications
You must be signed in to change notification settings - Fork 219
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add scripts for running gcn with dp & fix tensor under different device #313
Conversation
@DavdGao hi dawei, we are unfamiliar with this algorithm. Could you give an explanation for us about what is the (\epsilon, \delta) each single query (training round) satisfies, when we specify |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The shell script looks good.
The |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
please see the inline comments
Updated accordingly. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
@@ -63,8 +63,12 @@ def init_nbafl_ctx(base_trainer): | |||
ctx.regularizer = get_regularizer(cfg.regularizer.type) | |||
|
|||
# set noise scale during upload | |||
if cfg.trainer.type == 'nodefullbatch_trainer': | |||
num_train_data = sum(ctx.train_loader.dataset[0]['train_mask']) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe we should wrap all the dataset/dataloader with a unified attribute num_samples
in the future.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Agree.
As the title says.