Improving Robustness to Model Inversion Attacks via Sparse Coding Architectures
Robustness
DOI:
10.48550/arxiv.2403.14772
Publication Date:
2024-03-21
AUTHORS (4)
ABSTRACT
Recent model inversion attack algorithms permit adversaries to reconstruct a neural network's private training data just by repeatedly querying the network and inspecting its outputs. In this work, we develop novel architecture that leverages sparse-coding layers obtain superior robustness class of attacks. Three decades computer science research has studied sparse coding in context image denoising, object recognition, adversarial misclassification settings, but best our knowledge, connection state-of-the-art privacy vulnerabilities remains unstudied. However, architectures suggest an advantageous means defend against attacks because they allow us control amount irrelevant information encoded intermediate representations manner can be computed efficiently during is known have little effect on classification accuracy. Specifically, compared networks trained with variety defenses, maintain comparable or higher accuracy while degrading reconstructions factors 1.1 18.3 across reconstruction quality metrics (PSNR, SSIM, FID). This performance advantage holds 5 datasets ranging from CelebA faces medical images CIFAR-10, various SGD-based GAN-based attacks, including Plug-&-Play We provide cluster-ready PyTorch codebase promote standardize defense evaluations.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES ()
CITATIONS ()
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....