Skip to content
Snippets Groups Projects
Commit cf970ee4 authored by Rishi Sharma's avatar Rishi Sharma
Browse files

Update random alpha

parents da846d2a e39822b4
No related branches found
No related tags found
No related merge requests found
Showing
with 252 additions and 26 deletions
[DATASET]
dataset_package = decentralizepy.datasets.Celeba
dataset_class = Celeba
model_class = CNN
images_dir = /mnt/nfs/shared/leaf/data/celeba/data/raw/img_align_celeba
train_dir = /mnt/nfs/shared/leaf/data/celeba/per_user_data/train
test_dir = /mnt/nfs/shared/leaf/data/celeba/data/test
; python list of fractions below
sizes =
[OPTIMIZER_PARAMS]
optimizer_package = torch.optim
optimizer_class = SGD
lr = 0.001
[TRAIN_PARAMS]
training_package = decentralizepy.training.Training
training_class = Training
rounds = 4
full_epochs = False
batch_size = 16
shuffle = True
loss_package = torch.nn
loss_class = CrossEntropyLoss
[COMMUNICATION]
comm_package = decentralizepy.communication.TCP
comm_class = TCP
addresses_filepath = ip_addr_6Machines.json
[SHARING]
sharing_package = decentralizepy.sharing.TopKNormalized
sharing_class = TopKNormalized
alpha = 0.1
accumulation = True
accumulate_averaging_changes = True
\ No newline at end of file
[DATASET]
dataset_package = decentralizepy.datasets.Celeba
dataset_class = Celeba
model_class = CNN
images_dir = /mnt/nfs/shared/leaf/data/celeba/data/raw/img_align_celeba
train_dir = /mnt/nfs/shared/leaf/data/celeba/per_user_data/train
test_dir = /mnt/nfs/shared/leaf/data/celeba/data/test
; python list of fractions below
sizes =
[OPTIMIZER_PARAMS]
optimizer_package = torch.optim
optimizer_class = SGD
lr = 0.001
[TRAIN_PARAMS]
training_package = decentralizepy.training.Training
training_class = Training
rounds = 4
full_epochs = False
batch_size = 16
shuffle = True
loss_package = torch.nn
loss_class = CrossEntropyLoss
[COMMUNICATION]
comm_package = decentralizepy.communication.TCP
comm_class = TCP
addresses_filepath = ip_addr_6Machines.json
[SHARING]
sharing_package = decentralizepy.sharing.LowerBoundTopK
sharing_class = LowerBoundTopK
lower_bound = 0.1
alpha = 0.1
metro_hastings = False
accumulation = True
accumulate_averaging_changes = True
\ No newline at end of file
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
...@@ -36,3 +36,4 @@ alpha = 0.1 ...@@ -36,3 +36,4 @@ alpha = 0.1
wavelet=sym2 wavelet=sym2
level= 4 level= 4
accumulation = True accumulation = True
accumulate_averaging_changes = True
[DATASET]
dataset_package = decentralizepy.datasets.CIFAR10
dataset_class = CIFAR10
model_class = LeNet
train_dir = /mnt/nfs/shared/CIFAR
test_dir = /mnt/nfs/shared/CIFAR
; python list of fractions below
sizes =
random_seed = 99
partition_niid = True
shards = 1
[OPTIMIZER_PARAMS]
optimizer_package = torch.optim
optimizer_class = SGD
lr = 0.001
[TRAIN_PARAMS]
training_package = decentralizepy.training.Training
training_class = Training
rounds = 65
full_epochs = False
batch_size = 8
shuffle = True
loss_package = torch.nn
loss_class = CrossEntropyLoss
[COMMUNICATION]
comm_package = decentralizepy.communication.TCPRandomWalkRouting
comm_class = TCPRandomWalkRouting
addresses_filepath = ip_addr_6Machines.json
sampler = equi
[SHARING]
sharing_package = decentralizepy.sharing.SharingWithRWAsyncDynamic
sharing_class = SharingWithRWAsyncDynamic
\ No newline at end of file
...@@ -5,19 +5,20 @@ model_class = LeNet ...@@ -5,19 +5,20 @@ model_class = LeNet
train_dir = /mnt/nfs/shared/CIFAR train_dir = /mnt/nfs/shared/CIFAR
test_dir = /mnt/nfs/shared/CIFAR test_dir = /mnt/nfs/shared/CIFAR
; python list of fractions below ; python list of fractions below
sizes = sizes =
random_seed = 91 random_seed = 99
partition_niid = True partition_niid = True
shards = 4
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
training_package = decentralizepy.training.Training training_package = decentralizepy.training.Training
training_class = Training training_class = Training
rounds = 5 rounds = 65
full_epochs = False full_epochs = False
batch_size = 8 batch_size = 8
shuffle = True shuffle = True
...@@ -32,5 +33,4 @@ addresses_filepath = ip_addr_6Machines.json ...@@ -32,5 +33,4 @@ addresses_filepath = ip_addr_6Machines.json
[SHARING] [SHARING]
sharing_package = decentralizepy.sharing.PartialModel sharing_package = decentralizepy.sharing.PartialModel
sharing_class = PartialModel sharing_class = PartialModel
alpha = 0.5 alpha=0.5
accumulation = True
...@@ -5,20 +5,20 @@ model_class = LeNet ...@@ -5,20 +5,20 @@ model_class = LeNet
train_dir = /mnt/nfs/shared/CIFAR train_dir = /mnt/nfs/shared/CIFAR
test_dir = /mnt/nfs/shared/CIFAR test_dir = /mnt/nfs/shared/CIFAR
; python list of fractions below ; python list of fractions below
sizes = sizes =
random_seed = 91 random_seed = 99
partition_niid = True partition_niid = True
shards = 4 shards = 4
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
training_package = decentralizepy.training.Training training_package = decentralizepy.training.Training
training_class = Training training_class = Training
rounds = 5 rounds = 65
full_epochs = False full_epochs = False
batch_size = 8 batch_size = 8
shuffle = True shuffle = True
......
[DATASET]
dataset_package = decentralizepy.datasets.CIFAR10
dataset_class = CIFAR10
model_class = LeNet
train_dir = /mnt/nfs/shared/CIFAR
test_dir = /mnt/nfs/shared/CIFAR
; python list of fractions below
sizes =
random_seed = 99
partition_niid = True
shards = 4
[OPTIMIZER_PARAMS]
optimizer_package = torch.optim
optimizer_class = SGD
lr = 0.001
[TRAIN_PARAMS]
training_package = decentralizepy.training.Training
training_class = Training
rounds = 65
full_epochs = False
batch_size = 8
shuffle = True
loss_package = torch.nn
loss_class = CrossEntropyLoss
[COMMUNICATION]
comm_package = decentralizepy.communication.TCP
comm_class = TCP
addresses_filepath = ip_addr_6Machines.json
[SHARING]
sharing_package = decentralizepy.sharing.SubSampling
sharing_class = SubSampling
alpha = 0.5
[DATASET]
dataset_package = decentralizepy.datasets.CIFAR10
dataset_class = CIFAR10
model_class = LeNet
train_dir = /mnt/nfs/shared/CIFAR
test_dir = /mnt/nfs/shared/CIFAR
; python list of fractions below
sizes =
random_seed = 99
partition_niid = True
shards = 4
[OPTIMIZER_PARAMS]
optimizer_package = torch.optim
optimizer_class = SGD
lr = 0.001
[TRAIN_PARAMS]
training_package = decentralizepy.training.Training
training_class = Training
rounds = 65
full_epochs = False
batch_size = 8
shuffle = True
loss_package = torch.nn
loss_class = CrossEntropyLoss
[COMMUNICATION]
comm_package = decentralizepy.communication.TCP
comm_class = TCP
addresses_filepath = ip_addr_6Machines.json
[SHARING]
sharing_package = decentralizepy.sharing.PartialModel
sharing_class = PartialModel
alpha = 0.5
accumulation = True
accumulate_averaging_changes = True
\ No newline at end of file
...@@ -5,20 +5,20 @@ model_class = LeNet ...@@ -5,20 +5,20 @@ model_class = LeNet
train_dir = /mnt/nfs/shared/CIFAR train_dir = /mnt/nfs/shared/CIFAR
test_dir = /mnt/nfs/shared/CIFAR test_dir = /mnt/nfs/shared/CIFAR
; python list of fractions below ; python list of fractions below
sizes = sizes =
random_seed = 91 random_seed = 99
partition_niid = True partition_niid = True
shards = 4 shards = 4
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
training_package = decentralizepy.training.Training training_package = decentralizepy.training.Training
training_class = Training training_class = Training
rounds = 5 rounds = 65
full_epochs = False full_epochs = False
batch_size = 8 batch_size = 8
shuffle = True shuffle = True
...@@ -38,3 +38,4 @@ alpha = 0.5 ...@@ -38,3 +38,4 @@ alpha = 0.5
wavelet=sym2 wavelet=sym2
level= 4 level= 4
accumulation = True accumulation = True
accumulate_averaging_changes = True
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
# There are 734463 femnist samples # There are 734463 femnist samples
......
...@@ -9,7 +9,7 @@ sizes = ...@@ -9,7 +9,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
[TRAIN_PARAMS] [TRAIN_PARAMS]
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
# There are 734463 femnist samples # There are 734463 femnist samples
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
# There are 734463 femnist samples # There are 734463 femnist samples
...@@ -33,4 +33,5 @@ addresses_filepath = ip_addr_6Machines.json ...@@ -33,4 +33,5 @@ addresses_filepath = ip_addr_6Machines.json
sharing_package = decentralizepy.sharing.PartialModel sharing_package = decentralizepy.sharing.PartialModel
sharing_class = PartialModel sharing_class = PartialModel
alpha = 0.1 alpha = 0.1
accumulation = True accumulation = True
\ No newline at end of file accumulate_averaging_changes = True
\ No newline at end of file
[DATASET]
dataset_package = decentralizepy.datasets.Femnist
dataset_class = Femnist
random_seed = 97
model_class = CNN
train_dir = /mnt/nfs/shared/leaf/data/femnist/per_user_data/train
test_dir = /mnt/nfs/shared/leaf/data/femnist/data/test
; python list of fractions below
sizes =
[OPTIMIZER_PARAMS]
optimizer_package = torch.optim
optimizer_class = SGD
lr = 0.001
# There are 734463 femnist samples
[TRAIN_PARAMS]
training_package = decentralizepy.training.Training
training_class = Training
rounds = 47
full_epochs = False
batch_size = 16
shuffle = True
loss_package = torch.nn
loss_class = CrossEntropyLoss
[COMMUNICATION]
comm_package = decentralizepy.communication.TCP
comm_class = TCP
addresses_filepath = ip_addr_6Machines.json
[SHARING]
sharing_package = decentralizepy.sharing.LowerBoundTopK
sharing_class = LowerBoundTopK
lower_bound = 0.1
alpha = 0.1
accumulation = True
accumulate_averaging_changes = True
\ No newline at end of file
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
# There are 734463 femnist samples # There are 734463 femnist samples
......
...@@ -10,7 +10,7 @@ sizes = ...@@ -10,7 +10,7 @@ sizes =
[OPTIMIZER_PARAMS] [OPTIMIZER_PARAMS]
optimizer_package = torch.optim optimizer_package = torch.optim
optimizer_class = Adam optimizer_class = SGD
lr = 0.001 lr = 0.001
# There are 734463 femnist samples # There are 734463 femnist samples
...@@ -37,3 +37,4 @@ alpha = 0.1 ...@@ -37,3 +37,4 @@ alpha = 0.1
wavelet=sym2 wavelet=sym2
level= 4 level= 4
accumulation = True accumulation = True
accumulate_averaging_changes = True
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment