Inspired by this repo and ML Writing Month. Questions and discussions are most welcome!
Lil-log is the best blog I have ever read!
TNNLS 2019Adversarial Examples: Attacks and Defenses for Deep LearningIEEE ACCESS 2018Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey2019Adversarial Attacks and Defenses in Images, Graphs and Text: A Review2019A Study of Black Box Adversarial Attacks in Computer Vision2019Adversarial Examples in Modern Machine Learning: A Review2020Opportunities and Challenges in Deep Learning Adversarial Robustness: A SurveyTPAMI 2021Knowledge Distillation and Student-Teacher Learning for Visual Intelligence: A Review and New Outlooks2019Adversarial attack and defense in reinforcement learning-from AI security view2020A Survey of Privacy Attacks in Machine Learning2020Learning from Noisy Labels with Deep Neural Networks: A Survey2020Optimization for Deep Learning: An Overview2020Backdoor Attacks and Countermeasures on Deep Learning: A Comprehensive Review2020Learning from Noisy Labels with Deep Neural Networks: A Survey2020Adversarial Machine Learning in Image Classification: A Survey Towards the Defender's Perspective2020Efficient Transformers: A Survey2019A Survey of Black-Box Adversarial Attacks on Computer Vision Models2020Backdoor Learning: A Survey2020Transformers in Vision: A Survey2020A Survey on Neural Network Interpretability2020A Survey of Privacy Attacks in Machine Learning2020Dataset Security for Machine Learning: Data Poisoning, Backdoor Attacks, and Defenses2021Recent Advances in Adversarial Training for Adversarial Robustness (Our work, accepted by IJCAI 2021)2021Explainable Artificial Intelligence Approaches: A Survey2021A Survey on Understanding, Visualizations, and Explanation of Deep Neural Networks2020A survey on Semi-, Self- and Unsupervised Learning for Image Classification2021Model Complexity of Deep Learning: A Survey2021Deep Generative Modelling: A Comparative Review of VAEs, GANs, Normalizing Flows, Energy-Based and Autoregressive Models2021Dataset Security for Machine Learning: Data Poisoning, Backdoor Attacks, and Defenses2019Advances and Open Problems in Federated Learning2021Countering Malicious DeepFakes: Survey, Battleground, and Horizon
ICLRIntriguing properties of neural networksARXIV[Identifying and attacking the saddle point problem in high-dimensional non-convex optimization]
EuroS&PThe limitations of deep learning in adversarial settingsCVPRDeepfoolSPC&W Towards evaluating the robustness of neural networksArxivTransferability in machine learning: from phenomena to black-box attacks using adversarial samplesNIPS[Adversarial Images for Variational Autoencoders]ARXIV[A boundary tilting persepective on the phenomenon of adversarial examples]ARXIV[Adversarial examples in the physical world]
ICLRDelving into Transferable Adversarial Examples and Black-box AttacksCVPRUniversal Adversarial PerturbationsICCVAdversarial Examples for Semantic Segmentation and Object DetectionARXIVAdversarial Examples that Fool DetectorsCVPRA-Fast-RCNN: Hard Positive Generation via Adversary for Object DetectionICCVAdversarial Examples Detection in Deep Networks with Convolutional Filter StatisticsAIS[Adversarial examples are not easily detected: Bypassing ten detection methods]ICCVUNIVERSAL[Universal Adversarial Perturbations Against Semantic Image Segmentation]ICLR[Adversarial Machine Learning at Scale]ARXIV[The space of transferable adversarial examples]ARXIV[Adversarial attacks on neural network policies]
ICLRGenerating Natural Adversarial ExamplesNeurlPSConstructing Unrestricted Adversarial Examples with Generative ModelsIJCAIGenerating Adversarial Examples with Adversarial NetworksCVPRGenerative Adversarial PerturbationsAAAILearning to Attack: Adversarial transformation networksS&PLearning Universal Adversarial Perturbations with Generative ModelsCVPRRobust physical-world attacks on deep learning visual classificationICLRSpatially Transformed Adversarial ExamplesCVPRBoosting Adversarial Attacks With MomentumICMLObfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples πCVPRUNIVERSAL[Art of Singular Vectors and Universal Adversarial Perturbations]ARXIV[Adversarial Spheres]ECCV[Characterizing adversarial examples based on spatial consistency information for semantic segmentation]ARXIV[Generating natural language adversarial examples]SP[Audio adversarial examples: Targeted attacks on speech-to-text]ARXIV[Adversarial attack on graph structured data]ARXIV[Maximal Jacobian-based Saliency Map Attack (Variants of JAMA)]SP[Exploiting Unintended Feature Leakage in Collaborative Learning]
CVPRFeature Space Perturbations Yield More Transferable Adversarial ExamplesICLRThe Limitations of Adversarial Training and the Blind-Spot AttackICLRAre adversarial examples inevitable? πIEEE TECOne pixel attack for fooling deep neural networksARXIVGeneralizable Adversarial Attacks Using Generative ModelsICMLNATTACK: Learning the Distributions of Adversarial Examples for an Improved Black-Box Attack on Deep Neural NetworksπARXIVSemanticAdv: Generating Adversarial Examples via Attribute-conditional Image EditingCVPRRob-GAN: Generator, Discriminator, and Adversarial AttackerARXIVCycle-Consistent Adversarial {GAN:} the integration of adversarial attack and defenseARXIVGenerating Realistic Unrestricted Adversarial Inputs using Dual-Objective {GAN} Training πICCVSparse and Imperceivable Adversarial AttacksπARXIVPerturbations are not Enough: Generating Adversarial Examples with Spatial DistortionsARXIVJoint Adversarial Training: Incorporating both Spatial and Pixel AttacksIJCAITransferable Adversarial Attacks for Image and Video Object DetectionTPAMIGeneralizable Data-Free Objective for Crafting Universal Adversarial PerturbationsCVPRDecoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and DefensesCVPR[FDA: Feature Disruptive Attack]ARXIV[SmoothFool: An Efficient Framework for Computing Smooth Adversarial Perturbations]CVPR[SparseFool: a few pixels make a big difference]ICLR[Adversarial Attacks on Graph Neural Networks via Meta Learning]NeurIPS[Deep Leakage from Gradients]CCS[Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning]ICCV[Universal Perturbation Attack Against Image Retrieval]ICCV[Enhancing Adversarial Example Transferability with an Intermediate Level Attack]CVPR[Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks]ICLR[ADef: an Iterative Algorithm to Construct Adversarial Deformations]Neurips[iDLG: Improved deep leakage from gradients.]ARXIV[Reversible Adversarial Attack based on Reversible Image Transformation]CCS[Seeing isnβt Believing: Towards More Robust Adversarial Attack Against Real World Object Detectors]NeurIPS[Learning to Confuse: Generating Training Time Adversarial Data with Auto-Encoder]
ICLRFooling Detection Alone is Not Enough: Adversarial Attack against Multiple Object TrackingπARXIV[Sponge Examples: Energy-Latency Attacks on Neural Networks]ICML[Minimally distorted Adversarial Examples with a Fast Adaptive Boundary Attack]ICML[Stronger and Faster Wasserstein Adversarial Attacks]CVPR[QEBA: Query-Efο¬cient Boundary-Based Blackbox Attack]ECCV[New Threats Against Object Detector with Non-local Block]ARXIV[Towards Imperceptible Universal Attacks on Texture Recognition]ECCV[Frequency-Tuned Universal Adversarial Attacks]AAAI[Learning Transferable Adversarial Examples via Ghost Networks]ECCV[SPARK: Spatial-aware Online Incremental Attack Against Visual Tracking]Neurips[Inverting Gradients - How easy is it to break privacy in federated learning?]ICLR[Nesterov Accelerated Gradient and Scale Invariance for Adversarial Attacks]NeurIPS[On Adaptive Attacks to Adversarial Example Defenses]AAAI[Beyond Digital Domain: Fooling Deep Learning Based Recognition System in Physical World]ARXIV[Adversarial Color Enhancement: Generating Unrestricted Adversarial Images by Optimizing a Color Filter]CVPR[Adversarial Camouflage: Hiding Physical-World Attacks With Natural Styles]CVPR[Universal Physical Camouflage Attacks on Object Detectors] codeARXIV[Understanding Object Detection Through An Adversarial Lens]CIKM[Can Adversarial Weight Perturbations Inject Neural Backdoors?]ICCV[Semantic Adversarial Attacks: Parametric Transformations That Fool Deep Classifiers]
ARXIV[On Generating Transferable Targeted Perturbations]CVPR[See through Gradients: Image Batch Recovery via GradInversion] πARXIV[Admix: Enhancing the Transferability of Adversarial Attacks]ARXIV[Deep Image Destruction: A Comprehensive Study on Vulnerability of Deep Image-to-Image Models against Adversarial Attacks]ARXIV[Poisoning the Unlabeled Dataset of Semi-Supervised Learning] CarliniARXIV[AdvHaze: Adversarial Haze Attack]CVPRLAFEAT : Piercing Through Adversarial Defenses with Latent FeaturesARXIV[IMPERCEPTIBLE ADVERSARIAL EXAMPLES FOR FAKE IMAGE DETECTION]ICME[TRANSFERABLE ADVERSARIAL EXAMPLES FOR ANCHOR FREE OBJECT DETECTION]ICLR[Unlearnable Examples: Making Personal Data Unexploitable]ICMLW[Detecting Adversarial Examples Is (Nearly) As Hard As Classifying Them]ARXIV[Mischief: A Simple Black-Box Attack Against Transformer Architectures]ECCV[Patch-wise Attack for Fooling Deep Neural Network]ICCV[Naturalistic Physical Adversarial Patch for Object Detectors]CVPR[Natural Adversarial Examples]ICLR[WaNet - Imperceptible Warping-based Backdoor Attack]
ICLR[ON IMPROVING ADVERSARIAL TRANSFERABILITY OF VISION TRANSFORMERS]TIFS[Decision-based Adversarial Attack with Frequency Mixup]
- [Learning with a strong adversary]
- [IMPROVING BACK-PROPAGATION BY ADDING AN ADVERSARIAL GRADIENT]
- [Distributional Smoothing with Virtual Adversarial Training]
ARXIVCountering Adversarial Images using Input TransformationsICCV[SafetyNet: Detecting and Rejecting Adversarial Examples Robustly]ArxivDetecting adversarial samples from artifactsICLROn Detecting Adversarial Perturbations πASIA CCS[Practical black-box attacks against machine learning]ARXIV[The space of transferable adversarial examples]ICCV[Adversarial Examples for Semantic Segmentation and Object Detection]
ICLRDefense-{GAN}: Protecting Classifiers Against Adversarial Attacks Using Generative Models- .
ICLREnsemble Adversarial Training: Attacks and Defences CVPRDefense Against Universal Adversarial PerturbationsCVPRDeflecting Adversarial Attacks With Pixel DeflectionTPAMIVirtual adversarial training: a regularization method for supervised and semi-supervised learning πARXIVAdversarial Logit PairingCVPRDefense Against Adversarial Attacks Using High-Level Representation Guided DenoiserARXIVEvaluating and understanding the robustness of adversarial logit pairingCCSMachine Learning with Membership Privacy Using Adversarial RegularizationARXIV[On the robustness of the cvpr 2018 white-box adversarial example defenses]ICLR[Thermometer Encoding: One Hot Way To Resist Adversarial Examples]IJCAI[Curriculum Adversarial Training]ICLR[Countering Adversarial Images using Input Transformations]CVPR[Defense Against Adversarial Attacks Using High-Level Representation Guided Denoiser]ICLR[Towards Deep Learning Models Resistant to Adversarial Attacks]AAAI[Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing Their Input Gradients]NIPS[Adversarially robust generalization requires more data]ARXIV[Is robustness the cost of accuracy? - {A} comprehensive study on the robustness of 18 deep image classification models.]ARXIV[Robustness may be at odds with accuracy]ICLR[PIXELDEFEND: LEVERAGING GENERATIVE MODELS TO UNDERSTAND AND DEFEND AGAINST ADVERSARIAL EXAMPLES]
NIPSAdversarial Training and Robustness for Multiple PerturbationsNIPSAdversarial Robustness through Local LinearizationCVPRRetrieval-Augmented Convolutional Neural Networks against Adversarial ExamplesCVPRFeature Denoising for Improving Adversarial RobustnessNEURIPSA New Defense Against Adversarial Images: Turning a Weakness into a StrengthICMLInterpreting Adversarially Trained Convolutional Neural NetworksICLRRobustness May Be at Odds with AccuracyπIJCAIImproving the Robustness of Deep Neural Networks via Adversarial Training with Triplet LossICMLAdversarial Examples Are a Natural Consequence of Test Error in NoiseπICMLOn the Connection Between Adversarial Robustness and Saliency Map InterpretabilityNeurIPSMetric Learning for Adversarial RobustnessARXIVDefending Adversarial Attacks by Correcting logitsICCVAdversarial Learning With Margin-Based Triplet Embedding RegularizationICCVCIIDefence: Defeating Adversarial Attacks by Fusing Class-Specific Image Inpainting and Image DenoisingNIPSAdversarial Examples Are Not Bugs, They Are FeaturesICMLUsing Pre-Training Can Improve Model Robustness and UncertaintyNIPSDefense Against Adversarial Attacks Using Feature Scattering-based Adversarial TrainingπICCVImproving Adversarial Robustness via Guided Complement EntropyNIPSRobust Attribution Regularization πNIPSAre Labels Required for Improving Adversarial Robustness?ICLRTheoretically Principled Trade-off between Robustness and AccuracyCVPR[Adversarial defense by stratified convolutional sparse coding]ICML[On the Convergence and Robustness of Adversarial Training]CVPR[Robustness via Curvature Regularization, and Vice Versa]CVPR[ComDefend: An Efficient Image Compression Model to Defend Adversarial Examples]ICML[Improving Adversarial Robustness via Promoting Ensemble Diversity]ICML[Towards the first adversarially robust neural network model on {MNIST}]NIPS[Unlabeled Data Improves Adversarial Robustness]ICCV[Evaluating Robustness of Deep Image Super-Resolution Against Adversarial Attacks]ICML[Using Pre-Training Can Improve Model Robustness and Uncertainty]ARXIV[Improving adversarial robustness of ensembles with diversity training]ICML[Adversarial Robustness Against the Union of Multiple Perturbation Models]CVPR[Robustness via Curvature Regularization, and Vice Versa]NIPS[Robustness to Adversarial Perturbations in Learning from Incomplete Data]ICML[Improving Adversarial Robustness via Promoting Ensemble Diversity]NIPS[Adversarial Robustness through Local Linearization]ARXIV[Adversarial training can hurt generalization]NIPS[Adversarial training for free!]ICLR[Improving the generalization of adversarial training with domain adaptation]CVPR[Disentangling Adversarial Robustness and Generalization]NIPS[Adversarial Training and Robustness for Multiple Perturbations]ICCV[Bilateral Adversarial Training: Towards Fast Training of More Robust Models Against Adversarial Attacks]ICML[On the Convergence and Robustness of Adversarial Training]ICML[Rademacher Complexity for Adversarially Robust Generalization]ARXIV[Adversarially Robust Generalization Just Requires More Unlabeled Data]ARXIV[You only propagate once: Accelerating adversarial training via maximal principle]NIPSCross-Domain Transferability of Adversarial PerturbationsARXIV[Adversarial Robustness as a Prior for Learned Representations]ICLR[Structured Adversarial Attack: Towards General Implementation and Better Interpretability]ICLR[Defensive Quantization: When Efficiency Meets Robustness]NeurIPS[A New Defense Against Adversarial Images: Turning a Weakness into a Strength]
ICLRJacobian Adversarially Regularized Networks for RobustnessCVPRWhat it Thinks is Important is Important: Robustness Transfers through Input GradientsICLRAdversarially Robust Representations with Smooth Encoders πARXIVHeat and Blur: An Effective and Fast Defense Against Adversarial ExamplesICMLTriple Wins: Boosting Accuracy, Robustness and Efficiency Together by Enabling Input-Adaptive InferenceCVPRWavelet Integrated CNNs for Noise-Robust Image ClassificationARXIVDeflecting Adversarial AttacksICLRRobust Local Features for Improving the Generalization of Adversarial TrainingICLREnhancing Transformation-Based Defenses Against Adversarial Attacks with a Distribution ClassifierCVPRA Self-supervised Approach for Adversarial RobustnessICLRImproving Adversarial Robustness Requires Revisiting Misclassified Examples πARXIVManifold regularization for adversarial robustnessNeurIPSDVERGE: Diversifying Vulnerabilities for Enhanced Robust Generation of EnsemblesARXIVA Closer Look at Accuracy vs. RobustnessNeurIPSEnergy-based Out-of-distribution DetectionARXIVOut-of-Distribution Generalization via Risk Extrapolation (REx)CVPRAdversarial Examples Improve Image RecognitionICML[Confidence-Calibrated Adversarial Training: Generalizing to Unseen Attacks] πICML[Efficiently Learning Adversarially Robust Halfspaces with Noise]ICML[Implicit Euler Skip Connections: Enhancing Adversarial Robustness via Numerical Stability]ICML[Friendly Adversarial Training: Attacks Which Do Not Kill Training Make Adversarial Learning Stronger]ICML[Learning Adversarially Robust Representations via Worst-Case Mutual Information Maximization] πICML[Overfitting in adversarially robust deep learning] πICML[Proper Network Interpretability Helps Adversarial Robustness in Classification]ICML[Randomization matters How to defend against strong adversarial attacks]ICML[Reliable Evaluation of Adversarial Robustness with an Ensemble of Diverse Parameter-free Attacks]ICML[Towards Understanding the Regularization of Adversarial Robustness on Neural Networks]CVPR[Defending Against Universal Attacks Through Selective Feature Regeneration]ARXIV[Understanding and improving fast adversarial training]ARXIV[Cat: Customized adversarial training for improved robustness]ICLR[MMA Training: Direct Input Space Margin Maximization through Adversarial Training]ARXIV[Bridging the performance gap between fgsm and pgd adversarial training]CVPR[Adversarial Vertex Mixup: Toward Better Adversarially Robust Generalization]ARXIV[Towards understanding fast adversarial training]ARXIV[Overfitting in adversarially robust deep learning]ICLR[Robust local features for improving the generalization of adversarial training]ICML[Confidence-Calibrated Adversarial Training: Generalizing to Unseen Attacks]ARXIV[Regularizers for single-step adversarial training]CVPR[Single-step adversarial training with dropout scheduling]ICLR[Improving Adversarial Robustness Requires Revisiting Misclassified Examples]ARXIV[Fast is better than free: Revisiting adversarial training.]ARXIV[On the Generalization Properties of Adversarial Training]ARXIV[A closer look at accuracy vs. robustness]ICLR[Adversarially robust transfer learning]ARXIV[On Saliency Maps and Adversarial Robustness]ARXIV[On Detecting Adversarial Inputs with Entropy of Saliency Maps]ARXIV[Detecting Adversarial Perturbations with Saliency]ARXIV[Detection Defense Against Adversarial Attacks with Saliency Map]ARXIV[Model-based Saliency for the Detection of Adversarial Examples]CVPR[Auxiliary Training: Towards Accurate and Robust Models]CVPR[Single-step Adversarial training with Dropout Scheduling]CVPR[Achieving Robustness in the Wild via Adversarial Mixing With Disentangled Representations]ICMLTest-Time Training with Self-Supervision for Generalization under Distribution ShiftsNeurIPS[Improving robustness against common corruptions by covariate shift adaptation]CCS[Gotta Catch'Em All: Using Honeypots to Catch Adversarial Attacks on Neural Networks]ECCV[A simple way to make neural networks robust against diverse image corruptions]CVPRW[Role of Spatial Context in Adversarial Robustness for Object Detection]WACV[Local Gradients Smoothing: Defense against localized adversarial attacks]NeurIPS[Adversarial Weight Perturbation Helps Robust Generalization]MM[DIPDefend: Deep Image Prior Driven Defense against Adversarial Examples]ECCV[Adversarial Data Augmentation viaDeformation Statistics]
ARXIVOn the Limitations of Denoising Strategies as Adversarial DefensesAAAI[Understanding catastrophic overfitting in single-step adversarial training]ICLR[Bag of tricks for adversarial training]ARXIV[Bridging the Gap Between Adversarial Robustness and Optimization Bias]ICLR[Perceptual Adversarial Robustness: Defense Against Unseen Threat Models]AAAI[Adversarial Robustness through Disentangled Representations]ARXIV[Understanding Robustness of Transformers for Image Classification]CVPR[Adversarial Robustness under Long-Tailed Distribution]ARXIV[Adversarial Attacks are Reversible with Natural Supervision]AAAI[Attribute-Guided Adversarial Training for Robustness to Natural Perturbations]ICLR[LEARNING PERTURBATION SETS FOR ROBUST MACHINE LEARNING]ICLR[Improving Adversarial Robustness via Channel-wise Activation Suppressing]AAAI[Efficient Certification of Spatial Robustness]ARXIV[Domain Invariant Adversarial Learning]ARXIV[Learning Defense Transformers for Counterattacking Adversarial Examples]ICLR[ONLINE ADVERSARIAL PURIFICATION BASED ON SELF-SUPERVISED LEARNING]ARXIV[Removing Adversarial Noise in Class Activation Feature Space]ARXIV[Improving Adversarial Robustness Using Proxy Distributions]ARXIV[Decoder-free Robustness Disentanglement without (Additional) Supervision]ARXIV[Fighting Gradients with Gradients: Dynamic Defenses against Adversarial Attacks]ARXIV[Reversible Adversarial Attack based on Reversible Image Transformation]ICLR[ONLINE ADVERSARIAL PURIFICATION BASED ON SELF-SUPERVISED LEARNING]ARXIV[Towards Corruption-Agnostic Robust Domain Adaptation]ARXIV[Adversarially Trained Models with Test-Time Covariate Shift Adaptation]ICLR workshop[COVARIATE SHIFT ADAPTATION FOR ADVERSARIALLY ROBUST CLASSIFIER]ARXIV[Self-Supervised Adversarial Example Detection by Disentangled Representation]AAAI[Adversarial Defence by Diversified Simultaneous Training of Deep Ensembles]ARXIV[Understanding Catastrophic Overfitting in Adversarial Training]ACM Trans. Multimedia Comput. Commun. Appl[Towards Corruption-Agnostic Robust Domain Adaptation]ICLR[TENT: FULLY TEST-TIME ADAPTATION BY ENTROPY MINIMIZATION]ARXIV[Attacking Adversarial Attacks as A Defense]ICML[Adversarial purification with Score-based generative models]ARXIV[Adversarial Visual Robustness by Causal Intervention]CVPR[MaxUp: Lightweight Adversarial Training With Data Augmentation Improves Neural Network Training]MM[AdvFilter: Predictive Perturbation-aware Filtering against Adversarial Attack via Multi-domain Learning]CVPR[Robust and Accurate Object Detection via Adversarial Learning]ARXIV[Markpainting: Adversarial Machine Learning meets Inpainting]ICLR[EFFICIENT CERTIFIED DEFENSES AGAINST PATCH ATTACKS ON IMAGE CLASSIFIERS]ARXIV[Learning Defense Transformers for Counterattacking Adversarial Examples]ARXIV[Towards Robust Vision Transformer]ARXIV[Reveal of Vision Transformers Robustness against Adversarial Attacks]ARXIV[Intriguing Properties of Vision Transformers]ARXIV[Vision transformers are robust learners]ARXIV[On Improving Adversarial Transferability of Vision Transformers]ARXIV[On the adversarial robustness of visual transformers]ARXIV[On the robustness of vision transformers to adversarial examples]ARXIV[Understanding Robustness of Transformers for Image Classification]ARXIV[Regional Adversarial Training for Better Robust Generalization]CCS[DetectorGuard: Provably Securing Object Detectors against Localized Patch Hiding Attacks]ARXIV[MODELLING ADVERSARIAL NOISE FOR ADVERSARIAL DEFENSE]ICCV[Adversarial Example Detection Using Latent Neighborhood Graph]ARXIV[Identification of Attack-Specific Signatures in Adversarial Examples]Neurips[How Should Pre-Trained Language Models Be Fine-Tuned Towards Adversarial Robustness?]ARXIV[Adversarial Robustness Comparison of Vision Transformer and MLP-Mixer to CNNs]ARXIV[Learning Defense Transformers for Counterattacking Adversarial Examples]ADVM[Detecting Adversarial Patch Attacks through Global-local Consistency]ICCV[Can Shape Structure Features Improve Model Robustness under Diverse Adversarial Settings?]ICLR[Undistillable: Making A Nasty Teacher That CANNOT teach students]ICCV[Revisiting Adversarial Robustness Distillation: Robust Soft Labels Make Student Better]ARXIV[Two Coupled Rejection Metrics Can Tell Adversarial Examples Apart]ARXIV[Consistency Regularization for Adversarial Robustness]ICML[CIFS: Improving Adversarial Robustness of CNNs via Channel-wise Importance-based Feature Selection]NeurIPS[Adversarial Neuron Pruning Purifies Backdoored Deep Models]ICCV[Towards Understanding the Generative Capability of Adversarially Robust Classifiers]NeurIPS[Better Safe Than Sorry: Preventing Delusive Adversaries with Adversarial Training]NeurIPS[Data Augmentation Can Improve Robustness]NeurIPS[When does Contrastive Learning Preserve Adversarial Robustness from Pretraining to Finetuning?]
ARXIV[$\alpha$ Weighted Federated Adversarial Training]AAAI[Safe Distillation Box]USENIX[Transferring Adversarial Robustness Through Robust Representation Matching]ARXIV[Robustness and Accuracy Could Be Reconcilable by (Proper) Definition]ARXIV[IMPROVING ADVERSARIAL DEFENSE WITH SELF SUPERVISED TEST-TIME FINE-TUNING]ARXIV[Exploring Memorization in Adversarial Training]IJCV[Open-Set Adversarial Defense with Clean-Adversarial Mutual Learning]ARXIV[Adversarial Detection and Correction by Matching Prediction Distribution]ARXIV[Enhancing Adversarial Training with Feature Separability]ARXIV[An Eye for an Eye: Defending against Gradient-based Attacks with Gradients]
ICCV 2017CVAE-GAN: Fine-Grained Image Generation Through Asymmetric TrainingICML 2016Autoencoding beyond pixels using a learned similarity metricARXIV 2019Natural Adversarial ExamplesICML 2017Conditional Image Synthesis with Auxiliary Classifier {GAN}sICCV 2019SinGAN: Learning a Generative Model From a Single Natural ImageICLR 2020Robust And Interpretable Blind Image Denoising Via Bias-Free Convolutional Neural NetworksICLR 2020Pay Attention to Features, Transfer Learn Faster CNNsICLR 2020On Robustness of Neural Ordinary Differential EquationsICCV 2019Real Image Denoising With Feature AttentionICLR 2018Multi-Scale Dense Networks for Resource Efficient Image ClassificationARXIV 2019Rethinking Data Augmentation: Self-Supervision and Self-DistillationICCV 2019Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self DistillationARXIV 2019Adversarially Robust DistillationARXIV 2019Knowledge Distillation from Internal RepresentationsICLR 2020Contrastive Representation Distillation πNIPS 2018Faster Neural Networks Straight from JPEGARXIV 2019A Closer Look at Double BackpropagationπCVPR 2016Learning Deep Features for Discriminative LocalizationICML 2019Noise2Self: Blind Denoising by Self-SupervisionARXIV 2020Supervised Contrastive LearningCVPR 2020High-Frequency Component Helps Explain the Generalization of Convolutional Neural NetworksNIPS 2017[Counterfactual Fairness]ARXIV 2020[An Adversarial Approach for Explaining the Predictions of Deep Neural Networks]CVPR 2014[Rich feature hierarchies for accurate object detection and semantic segmentation]ICLR 2018[Spectral Normalization for Generative Adversarial Networks]NIPS 2018[MetaGAN: An Adversarial Approach to Few-Shot Learning]ARXIV 2019[Breaking the cycle -- Colleagues are all you need]ARXIV 2019[LOGAN: Latent Optimisation for Generative Adversarial Networks]ICML 2020[Margin-aware Adversarial Domain Adaptation with Optimal Transport]ICML 2020[Representation Learning Using Adversarially-Contrastive Optimal Transport]ICLR 2021[Free Lunch for Few-shot Learning: Distribution Calibration]CVPR 2019[Unprocessing Images for Learned Raw Denoising]TPAMI 2020[Image Quality Assessment: Unifying Structure and Texture Similarity]CVPR 2020[Dreaming to Distill: Data-free Knowledge Transfer via DeepInversion]ICLR 2021[WHAT SHOULD NOT BE CONTRASTIVE IN CONTRASTIVE LEARNING]ARXIV[MT3: Meta Test-Time Training for Self-Supervised Test-Time Adaption]ARXIV[UNSUPERVISED DOMAIN ADAPTATION THROUGH SELF-SUPERVISION]ARXIV[Estimating Example Difficulty using Variance of Gradients]ICML 2020[Transfer Learning without Knowing: Reprogramming Black-box Machine Learning Models with Scarce Data and Limited Resources]ARXIV[DATASET DISTILLATION]ARXIV 2022[Debugging Differential Privacy: A Case Study for Privacy Auditing]ARXIV[Adversarial Robustness and Catastrophic Forgetting]