TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Enhancing Robust Representation in Adversarial Training: A...

Enhancing Robust Representation in Adversarial Training: Alignment and Exclusion Criteria

Nuoyan Zhou, Nannan Wang, Decheng Liu, Dawei Zhou, Xinbo Gao

2023-10-05Adversarial RobustnessRepresentation LearningAdversarial DefenseAdversarial Attack
PaperPDFCode(official)

Abstract

Deep neural networks are vulnerable to adversarial noise. Adversarial Training (AT) has been demonstrated to be the most effective defense strategy to protect neural networks from being fooled. However, we find AT omits to learning robust features, resulting in poor performance of adversarial robustness. To address this issue, we highlight two criteria of robust representation: (1) Exclusion: \emph{the feature of examples keeps away from that of other classes}; (2) Alignment: \emph{the feature of natural and corresponding adversarial examples is close to each other}. These motivate us to propose a generic framework of AT to gain robust representation, by the asymmetric negative contrast and reverse attention. Specifically, we design an asymmetric negative contrast based on predicted probabilities, to push away examples of different classes in the feature space. Moreover, we propose to weight feature by parameters of the linear classifier as the reverse attention, to obtain class-aware feature and pull close the feature of the same class. Empirical evaluations on three benchmark datasets show our methods greatly advance the robustness of AT and achieve state-of-the-art performance.

Results

TaskDatasetMetricValueModel
Adversarial AttackCIFAR-10Attack: AutoAttack59.7TRADES-ANCRA/ResNet18
Adversarial DefenseCIFAR-10Accuracy81.7ResNet18 (TRADES-ANCRA/PGD-40)
Adversarial DefenseCIFAR-10Attack: AutoAttack59.7ResNet18 (TRADES-ANCRA/PGD-40)
Adversarial DefenseCIFAR-10Robust Accuracy82.96ResNet18 (TRADES-ANCRA/PGD-40)
Adversarial RobustnessCIFAR-100AutoAttacked Accuracy35.05ResNet18/MART-ANCRA
Adversarial RobustnessCIFAR-100Clean Accuracy60.1ResNet18/MART-ANCRA
Adversarial RobustnessCIFAR-10Accuracy81.7TRADES-ANCRA/ResNet18
Adversarial RobustnessCIFAR-10Attack: AutoAttack59.7TRADES-ANCRA/ResNet18

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16Language-Guided Contrastive Audio-Visual Masked Autoencoder with Automatically Generated Audio-Visual-Text Triplets from Videos2025-07-16A Mixed-Primitive-based Gaussian Splatting Method for Surface Reconstruction2025-07-15Bridging Robustness and Generalization Against Word Substitution Attacks in NLP via the Growth Bound Matrix Approach2025-07-14