Home > Research > Publications & Outputs > Unlearnable Examples Detection via Iterative Fi...

Links

Text available via DOI:

View graph of relations

Unlearnable Examples Detection via Iterative Filtering

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
  • Yi Yu
  • Qichen Zheng
  • Siyuan Yang
  • Wenhan Yang
  • Jun Liu
  • Shijian Lu
  • Yap Peng Tan
  • Kwok Yan Lam
  • Alex Kot
Close
Publication date18/09/2024
Host publicationArtificial Neural Networks and Machine Learning – ICANN 2024 - 33rd International Conference on Artificial Neural Networks, Proceedings
EditorsMichael Wand, Kristína Malinovská, Jürgen Schmidhuber, Igor V. Tetko
Place of PublicationCham
PublisherSpringer Science and Business Media Deutschland GmbH
Pages241-256
Number of pages16
ISBN (electronic)9783031723599
ISBN (print)9783031723582
<mark>Original language</mark>English
Event33rd International Conference on Artificial Neural Networks, ICANN 2024 - Lugano, Switzerland
Duration: 17/09/202420/09/2024

Conference

Conference33rd International Conference on Artificial Neural Networks, ICANN 2024
Country/TerritorySwitzerland
CityLugano
Period17/09/2420/09/24

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume15025 LNCS
ISSN (Print)0302-9743
ISSN (electronic)1611-3349

Conference

Conference33rd International Conference on Artificial Neural Networks, ICANN 2024
Country/TerritorySwitzerland
CityLugano
Period17/09/2420/09/24

Abstract

Deep neural networks are proven to be vulnerable to data poisoning attacks. Recently, a specific type of data poisoning attack known as availability attacks, has led to the failure of data utilization for model learning by adding imperceptible perturbations to images. Consequently, it is quite beneficial and challenging to detect poisoned samples, also known as Unlearnable Examples (UEs), from a mixed dataset. In response, we propose an Iterative Filtering approach for UEs identification. This method leverages the distinction between the inherent semantic mapping rules and shortcuts, without the need for any additional information. We verify that when training a classifier on a mixed dataset containing both UEs and clean data, the model tends to quickly adapt to the UEs compared to the clean data. Due to the accuracy gaps between training with clean/poisoned samples, we employ a model to misclassify clean samples while correctly identifying the poisoned ones. The incorporation of additional classes and iterative refinement enhances the model’s ability to differentiate between clean and poisoned samples. Extensive experiments demonstrate the superiority of our method over state-of-the-art detection approaches across various attacks, datasets, and poison ratios, significantly reducing the Half Total Error Rate (HTER) compared to existing methods.