Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3488932.3527283acmconferencesArticle/Chapter ViewAbstractPublication Pagesasia-ccsConference Proceedingsconference-collections
poster

Black-box and Target-specific Attack Against Interpretable Deep Learning Systems

Published: 30 May 2022 Publication History

Abstract

Deep neural network models are susceptible to malicious manipulations even in the black-box settings. Providing explanations for DNN models offers a sense of security by human involvement, which reveals whether the sample is benign or adversarial even though previous studies achieved a high attack success rate. However, interpretable deep learning systems (IDLSes) are shown to be susceptible to adversarial manipulations in white-box settings. Attacking IDLSes in black-box settings is challenging and remains an open research domain. In this work, we propose a black-box version of the white-box AdvEdge approach against IDLSes, which is query-efficient and gradient-free without obtaining any knowledge of the target DNN model and its coupled interpreter. Our approach takes advantage of transfer-based and score-based techniques using the effective microbial genetic algorithm (MGA). We achieve a high attack success rate with a small number of queries and high similarity in interpretations between adversarial and benign samples.

Supplementary Material

MP4 File (poster_video_black-box advedge_Eldor.mp4)
Eldor Abdukhamidov, Firuz Juraev, Mohammed Abuhamad, and Tamer ABUHMED. 2022. POSTER: Black-box and Target-specific Attack Against Interpretable Deep Learning Systems. In Proceedings of the 2022 ACM Asia Conference on Computer and Communications Security (ASIA CCS ?22)

References

[1]
Eldor Abdukhamidov, Mohammed Abuhamad, Firuz Juraev, Eric Chan-Tin, and Tamer AbuHmed. 2021. AdvEdge: Optimizing Adversarial Perturbations Against Interpretable Deep Learning. In International Conference on Computational Data and Social Networks. Springer, 93--105.
[2]
Inman Harvey. 2009. The microbial genetic algorithm. In European conference on artificial life. Springer, 126--133.
[3]
Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. 2017. Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083 (2017).
[4]
Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. 2013. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034 (2013).
[5]
Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. 2016. Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition. 2921--2929.

Cited By

View all
  • (2025)Untargeted Evasion Attacks on Deep Neural Networks Using StyleGANElectronics10.3390/electronics1403057414:3(574)Online publication date: 31-Jan-2025
  • (2024)Machine learning security and privacy: a review of threats and countermeasuresEURASIP Journal on Information Security10.1186/s13635-024-00158-32024:1Online publication date: 23-Apr-2024
  • (2024)SingleADV: Single-Class Target-Specific Attack Against Interpretable Deep Learning SystemsIEEE Transactions on Information Forensics and Security10.1109/TIFS.2024.340765219(5985-5998)Online publication date: 2024
  • Show More Cited By

Index Terms

  1. Black-box and Target-specific Attack Against Interpretable Deep Learning Systems

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    ASIA CCS '22: Proceedings of the 2022 ACM on Asia Conference on Computer and Communications Security
    May 2022
    1291 pages
    ISBN:9781450391405
    DOI:10.1145/3488932
    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 30 May 2022

    Check for updates

    Author Tags

    1. adversarial machine learning
    2. genetic algorithm
    3. interpretable machine learning
    4. single-class attack
    5. target-specific attack

    Qualifiers

    • Poster

    Funding Sources

    Conference

    ASIA CCS '22
    Sponsor:

    Acceptance Rates

    Overall Acceptance Rate 418 of 2,322 submissions, 18%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)16
    • Downloads (Last 6 weeks)2
    Reflects downloads up to 05 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2025)Untargeted Evasion Attacks on Deep Neural Networks Using StyleGANElectronics10.3390/electronics1403057414:3(574)Online publication date: 31-Jan-2025
    • (2024)Machine learning security and privacy: a review of threats and countermeasuresEURASIP Journal on Information Security10.1186/s13635-024-00158-32024:1Online publication date: 23-Apr-2024
    • (2024)SingleADV: Single-Class Target-Specific Attack Against Interpretable Deep Learning SystemsIEEE Transactions on Information Forensics and Security10.1109/TIFS.2024.340765219(5985-5998)Online publication date: 2024
    • (2024)The Impact of Model Variations on the Robustness of Deep Learning Models in Adversarial Settings2024 Silicon Valley Cybersecurity Conference (SVCC)10.1109/SVCC61185.2024.10637362(1-7)Online publication date: 17-Jun-2024
    • (2023)Black-box Adversarial Attack against Visual Interpreters for Deep Neural Networks2023 18th International Conference on Machine Vision and Applications (MVA)10.23919/MVA57639.2023.10215758(1-6)Online publication date: 23-Jul-2023

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media