[ { "Accuracy (PGD, eps=8/255)": 58.56, "Accuracy (clean data)": 81.07, "code_links": [ { "title": "yanemcovsky/SIAM", "url": "https://github.com/yanemcovsky/SIAM" } ], "date": "2019-11-17", "date2": 20191117, "model": "Smoothed inference", "paper": { "title": "Smoothed Inference for Adversarially-Trained Models", "url": "https://cknow.io/lib/153b44e4ebcb0379" }, "paper_data_uoa": "153b44e4ebcb0379" }, { "Accuracy (PGD, eps=8/255)": 54.77, "Accuracy (clean data)": 87.65, "code_links": [], "date": "2019-10-17", "date2": 20191017, "model": "CSAT", "paper": { "title": "Enforcing Linearity in DNN succours Robustness and Adversarial Image Generation", "url": "https://cknow.io/lib/4d14574cb2bcf0a6" }, "paper_data_uoa": "4d14574cb2bcf0a6" }, { "Accuracy (PGD, eps=8/255)": 53.42, "Accuracy (clean data)": 85.31, "code_links": [], "date": "2018-11-03", "date2": 20181103, "model": "L2L", "paper": { "title": "Learning to Defense by Learning to Attack", "url": "https://cknow.io/lib/03c1b42639911d95" }, "paper_data_uoa": "03c1b42639911d95" }, { "Accuracy (PGD, eps=8/255)": 48.53, "Accuracy (clean data)": 91.34, "code_links": [], "date": "2020-01-01", "date2": 20200101, "model": "IAAT", "paper": { "title": "Instance adaptive adversarial training: Improved accuracy tradeoffs in neural nets", "url": "https://cknow.io/lib/b08eb77b911187ad" }, "paper_data_uoa": "b08eb77b911187ad" }, { "Accuracy (PGD, eps=8/255)": 46.7, "code_links": [ { "title": "aamir-mustafa/pcl-adversarial-defense", "url": "https://github.com/aamir-mustafa/pcl-adversarial-defense" } ], "date": "2019-04-01", "date2": 20190401, "model": "PCL", "paper": { "title": "Adversarial Defense by Restricting the Hidden Space of Deep Neural Networks", "url": "https://cknow.io/lib/7044ec2a4c78b415" }, "paper_data_uoa": "7044ec2a4c78b415" }, { "Accuracy (PGD, eps=8/255)": 45.94, "Accuracy (clean data)": 84.89, "code_links": [], "date": "2018-11-22", "date2": 20181122, "model": "PNI-W", "paper": { "title": "Parametric Noise Injection: Trainable Randomness to Improve Deep Neural Network Robustness against Adversarial Attack", "url": "https://cknow.io/lib/62067e97d64a8e8c" }, "paper_data_uoa": "62067e97d64a8e8c" }, { "Accuracy (PGD, eps=8/255)": 39.14, "Accuracy (clean data)": 83.84, "code_links": [ { "title": "MadryLab/mnist_challenge", "url": "https://github.com/MadryLab/mnist_challenge" }, { "title": "locuslab/convex_adversarial", "url": "https://github.com/locuslab/convex_adversarial" }, { "title": "MadryLab/cifar10_challenge", "url": "https://github.com/MadryLab/cifar10_challenge" }, { "title": "revbucket/mister_ed", "url": "https://github.com/revbucket/mister_ed" }, { "title": "tianzheng4/Distributionally-Adversarial-Attack", "url": "https://github.com/tianzheng4/Distributionally-Adversarial-Attack" }, { "title": "P2333/Max-Mahalanobis-Training", "url": "https://github.com/P2333/Max-Mahalanobis-Training" }, { "title": "henry8527/GCE", "url": "https://github.com/henry8527/GCE" }, { "title": "Hadisalman/robust-verify-benchmark", "url": "https://github.com/Hadisalman/robust-verify-benchmark" }, { "title": "locuslab/robust_overfitting", "url": "https://github.com/locuslab/robust_overfitting" }, { "title": "bingcheng45/hnr-extension", "url": "https://github.com/bingcheng45/hnr-extension" }, { "title": "cs-giung/course-dl-TP", "url": "https://github.com/cs-giung/course-dl-TP" }, { "title": "amerch/CIFAR100-Training", "url": "https://github.com/amerch/CIFAR100-Training" }, { "title": "tianzheng4/Noisy-Training-Soft-Quantization", "url": "https://github.com/tianzheng4/Noisy-Training-Soft-Quantization" }, { "title": "snaka0213/PyTorch-AdvAttacks", "url": "https://github.com/snaka0213/PyTorch-AdvAttacks" }, { "title": "hope-yao/robust_attention_cifar", "url": "https://github.com/hope-yao/robust_attention_cifar" }, { "title": "abahram77/mnistChallenge", "url": "https://github.com/abahram77/mnistChallenge" }, { "title": "Harry24k/adversairal-defenses-pytorch", "url": "https://github.com/Harry24k/adversairal-defenses-pytorch" }, { "title": "jxnding/dsc531_bayes", "url": "https://github.com/jxnding/dsc531_bayes" } ], "date": "2017-06-19", "date2": 20170619, "model": "Adversarial training", "paper": { "title": "Towards Deep Learning Models Resistant to Adversarial Attacks", "url": "https://cknow.io/lib/2e16f82f424bbba9" }, "paper_data_uoa": "2e16f82f424bbba9" } ]