Skip to content

fclesio/pycon-africa-2020-security-ml

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 

Repository files navigation

PyCon Africa 2020 - Security in Machine Learning: Adversarial attacks, model failures and countermeasures

This repo brings the code related with the talk Security in Machine Learning: Adversarial attacks, model failures and countermeasures that was given in the PyCon Africa 2020.

The slide deck contains all the references and more cases

Demo Summary

  • Attacking ML Supply Chain: Owing to large resources (data + computation) required to train algorithms, the current practice is to reuse models trained by large corporations, and modify them slightly for task at hand (e.g: ResNet is a popular image recognition model from Microsoft). These models are curated ina Model Zoo (Caffe hosts popular image recognition models). In this attack,the adversary attacks the models hosted in Caffe, thereby poisoning the well for anyone else[1].

  • Membership Inference: The attacker can determine whether a given data record was part of the model’s training dataset or not[1]

  • Model Backdooring: Like in the “Attacking the ML Supply Chain”, In this attack scenario,the training process is either fully or partially outsourced to a malicious party who wants to provide the user with a trained model that contains a backdoor. The backdoored model would perform well on most inputs (including inputs that the end user may hold out as a validation set) but cause targeted misclassifications or degrade the accuracy of the model for inputs that satisfy some secret, attacker-chosen property, which we will refer to as the backdoor trigger[1]

  • Model Inversion: Attacker recovers the secret features used in the model by through careful queries[1]

  • Model Poisoning: The goal of the attacker is to contaminate the machine model generated in the training phase, so that predictions on new data will be modified in the testing phase[1]

  • Model Stealing: The attackers recreate the underlying model by legitimately querying the model. The functionality of the new model is same as that of the underlying model.[1]

References

[1] - Kumar, Ram Shankar Siva, et al. "Failure modes in machine learning systems." arXiv preprint arXiv:1911.11034 (2019)..

[2] - Adversatial.io

[3] - AI Myths