The field of audio and image security is rapidly evolving, with a focus on developing innovative methods to detect and prevent threats such as deepfakes, adversarial attacks, and watermarking. Recent research has explored the use of machine learning and deep learning techniques to improve the robustness of audio and image classification models. One notable direction is the development of anomaly detection frameworks that can identify out-of-distribution samples, such as speech deepfakes. Additionally, there is a growing interest in designing responsible AI systems that prioritize transparency, explainability, and fairness. Noteworthy papers in this area include CAARMA, which introduces a class augmentation framework to improve speaker verification, and SITA, which proposes a structurally imperceptible and transferable adversarial attack method for stylized image generation. Furthermore, the Imperceptible but Forgeable paper highlights the vulnerability of existing watermarking schemes to forgery attacks, emphasizing the need for more robust security measures.Overall, the field is moving towards developing more sophisticated and robust methods to address the increasingly complex threats in audio and image security.
Advances in Audio and Image Security
Sources
Adoption of Watermarking for Generative AI Systems in Practice and Implications under the new EU AI Act
Towards Imperceptible Adversarial Attacks for Time Series Classification with Local Perturbations and Frequency Analysis