Skip to content

We utilize the Adversarial Model Perturbations (AMP) regularizer to regularize clients’ models. The AMP regulzaizer is based on perturbing the model parameters so as to get a more generalized model. The claim of AMP regularizer is to reach flat minima and therefore is expected to reach flat minima in FL settings as well.

Notifications You must be signed in to change notification settings

gautamHCSCV/Federated-Learning-Methods-Comparison

Repository files navigation

Regularizing Federated Learning via adversarial model perturbations

Procedure

We regularize FL algorithms using Adversarial Model Perturbations (AMP) regularizer

We implement 4 algorithm for FL from scratch

  • FedAVG
  • FedProx
  • FedNTD
  • SCAFFOLD

Dependencies are given in requirement.txt file

pip install -r requirements.txt

How to RUN

Choose the required YAML file and run the following commands

python scaffold.py configs/scaffold_c100_amp.yaml > logfiles/scaffold_c100_amp.log
python fedntd.py configs/fedntd_c100_amp.yaml > logfiles/fedntd_c100_amp.log
python fedprox.py configs/fedprox_c100_amp.yaml > logfiles/fedprox_c100_amp.log
python fedavg.py configs/fedavg_c100_amp.yaml > logfiles/fedavg_c100_amp.log

Results

Comparsion of FL algorithms on CIFAR10 dataset

Comparison without AMP regularizer

img

Comparison with AMP regularizer

img

Comparsion of FL algorithms on CIFAR100 dataset

Comparison without AMP regularizer

img

Comparison with AMP regularizer

img

Members

  • Gautam Kumar (B19EE031)
  • Nirbhay Sharma (B19CSE114)

About

We utilize the Adversarial Model Perturbations (AMP) regularizer to regularize clients’ models. The AMP regulzaizer is based on perturbing the model parameters so as to get a more generalized model. The claim of AMP regularizer is to reach flat minima and therefore is expected to reach flat minima in FL settings as well.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published