Enhancing Machine Learning Security and Privacy

The recent developments in the field of machine learning security and privacy are significantly advancing the protection and verification of models and datasets. There is a notable shift towards ensuring the integrity and authenticity of both the hardware and software platforms used for model inference, as well as the datasets upon which models are trained. Innovations in verifiable fine-tuning methods are emerging to address the transparency issues in third-party fine-tuning services, ensuring that models are genuinely customized for individual users. Additionally, active privacy auditing frameworks are being developed to monitor and mitigate privacy risks during the fine-tuning process of large language models. The field is also witnessing a critical re-evaluation of data forging attacks, highlighting their practical limitations and the need for more robust detection mechanisms. Furthermore, there is a growing emphasis on benchmarking the adversarial robustness of dataset distillation methods, ensuring that these techniques are secure against various adversarial attacks. Overall, the research landscape is evolving towards more secure, transparent, and privacy-conscious practices in machine learning, with a focus on both proactive and reactive measures to safeguard intellectual property and user data.

Noteworthy papers include one that introduces a method for identifying the underlying GPU architecture and software stack of a machine learning model solely based on its input-output behavior, and another that proposes a novel active privacy auditing framework designed to identify and quantify privacy leakage risks during the supervised fine-tuning of language models.

Sources

Intellectual Property Protection for Deep Learning Model and Dataset Intelligence

Hardware and Software Platform Inference

Towards a Re-evaluation of Data Forging Attacks in Practice

vTune: Verifiable Fine-Tuning for LLMs Through Backdooring

On Active Privacy Auditing in Supervised Fine-tuning for White-Box Language Models

New Emerged Security and Privacy of Pre-trained Model: a Survey and Outlook

BEARD: Benchmarking the Adversarial Robustness for Dataset Distillation

Built with on top of