Adversarial Attacks and Defenses in Neural Networks

The recent advancements in adversarial attacks and defenses for neural networks, particularly in the context of Graph Neural Networks (GNNs) and Deep Reinforcement Learning (DRL), have seen significant innovations. Researchers are increasingly focusing on developing methods to protect models from backdoor attacks, where malicious triggers can manipulate model behavior without altering the model's overall performance. Notably, there is a shift towards creating defenses that are robust against both out-of-distribution (OOD) and in-distribution (ID) backdoor attacks, with a particular emphasis on maintaining high model accuracy while effectively mitigating these threats. Additionally, the field is witnessing the emergence of novel attack strategies that minimize the need for large-scale data poisoning or extensive computational resources, making them more practical and dangerous. These developments highlight the ongoing arms race between attackers and defenders, pushing the boundaries of what is possible in both securing and compromising neural network models.

Noteworthy papers include one that introduces a backdoor attack framework specifically designed for Graph Prompt Learning (GPL), demonstrating high attack success rates without modifying pre-trained GNN encoders. Another paper proposes a new class of backdoor attacks against DRL that achieve state-of-the-art performance by minimally altering the agent's rewards, showcasing a sophisticated approach to adversarial behavior induction.

Sources

Trojan Prompt Attacks on Graph Neural Networks

Adversarial Inception for Bounded Backdoor Poisoning in Deep Reinforcement Learning

DMGNN: Detecting and Mitigating Backdoor Attacks in Graph Neural Networks

BeniFul: Backdoor Defense via Middle Feature Analysis for Deep Neural Networks

Attack as Defense: Run-time Backdoor Implantation for Image Content Protection

LEGO-Learn: Label-Efficient Graph Open-Set Learning

Hacking the Fabric: Targeting Partial Reconfiguration for Fault Injection in FPGA Fabrics

Backdoor in Seconds: Unlocking Vulnerabilities in Large Pre-trained Models via Model Editing

Built with on top of