Back

Disclaimer: These are my personal notes on this paper. I am in no way related to this paper. All credits go towards the authors.



Evading Deepfake-Image Detectors with White- and Black-Box Attacks

April 1, 2020 - Paper Link - Tags: Adversarial, Black-Box, Deepfake, Detection, Perturbation

Summary

Created 5 different adversarial perturbation methods to attack deepfake detectors, primarily Frank et al's detector. The first two attacks focused on only changing the least significant bit in every pixel, the third created a universal patch, the fourth used a universal (single) low-level attribute vector of ProGAN to generate adversarial deepfakes, the fifth was a black-box method where adversarial perturbation was learned from one network, and those samples were tested on the goal network. Figure 3 shows the last 4 attack's results using ROC curves.

Notes

Interesting References

Citation: Carlini, Nicholas, and Hany Farid. "Evading Deepfake-Image Detectors with White-and Black-Box Attacks." Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. 2020.