Among Machine Learning (ML) models, Graph Neural Networks (GNN) have been shown to improve the performance of modern Network Intrusion Detection Systems (NIDS). However, their black-box nature poses a significant challenge to their practical deployment in the real world. In this context, researchers have developed eXplainable Artificial Intelligence (XAI) methods that reveal the inner workings of GNN models. Despite this, determining the most effective explainer is complex because different methods yield different explanations, and there are no standardized strategies. In this paper, we present an innovative approach for evaluating XAI methods in GNN-based NIDS. We evaluate explainers based on their capability to identify key graph components that an attacker can exploit to bypass detection. More accurate XAI algorithms can identify topological vulnerabilities, resulting in more effective attacks. We assess the effectiveness of different explainers by measuring the severity of structural attacks guided by the corresponding explanations. Our case study compares five XAI techniques on two publicly available datasets containing real-world network traffic. Results show that the explainer based on Integrated Gradients (IG) generates the most accurate explanations, allowing attackers to refine their strategies.

Evaluating Explainability of Graph Neural Networks for Network Intrusion Detection with Structural Attacks / Galli, D.; Venturi, A.; Marasco, I.; Marchetti, M.. - 3962:(2025). ( 2025 Joint National Conference on Cybersecurity, ITASEC and SERICS 2025 Alma Mater Studiorum University, ita 2025).

Evaluating Explainability of Graph Neural Networks for Network Intrusion Detection with Structural Attacks

Galli D.;Marchetti M.
2025

Abstract

Among Machine Learning (ML) models, Graph Neural Networks (GNN) have been shown to improve the performance of modern Network Intrusion Detection Systems (NIDS). However, their black-box nature poses a significant challenge to their practical deployment in the real world. In this context, researchers have developed eXplainable Artificial Intelligence (XAI) methods that reveal the inner workings of GNN models. Despite this, determining the most effective explainer is complex because different methods yield different explanations, and there are no standardized strategies. In this paper, we present an innovative approach for evaluating XAI methods in GNN-based NIDS. We evaluate explainers based on their capability to identify key graph components that an attacker can exploit to bypass detection. More accurate XAI algorithms can identify topological vulnerabilities, resulting in more effective attacks. We assess the effectiveness of different explainers by measuring the severity of structural attacks guided by the corresponding explanations. Our case study compares five XAI techniques on two publicly available datasets containing real-world network traffic. Results show that the explainer based on Integrated Gradients (IG) generates the most accurate explanations, allowing attackers to refine their strategies.
2025
2025 Joint National Conference on Cybersecurity, ITASEC and SERICS 2025
Alma Mater Studiorum University, ita
2025
3962
Galli, D.; Venturi, A.; Marasco, I.; Marchetti, M.
Evaluating Explainability of Graph Neural Networks for Network Intrusion Detection with Structural Attacks / Galli, D.; Venturi, A.; Marasco, I.; Marchetti, M.. - 3962:(2025). ( 2025 Joint National Conference on Cybersecurity, ITASEC and SERICS 2025 Alma Mater Studiorum University, ita 2025).
File in questo prodotto:
File Dimensione Formato  
paper50.pdf

Open access

Tipologia: VOR - Versione pubblicata dall'editore
Dimensione 355.56 kB
Formato Adobe PDF
355.56 kB Adobe PDF Visualizza/Apri
Pubblicazioni consigliate

Licenza Creative Commons
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11380/1379588
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact