Skip to Main Content (Press Enter)

Logo UNIMORE
  • ×
  • Home
  • Corsi
  • Insegnamenti
  • Professioni
  • Persone
  • Pubblicazioni
  • Strutture
  • Terza Missione
  • Attività
  • Competenze

UNI-FIND
Logo UNIMORE

|

UNI-FIND

unimore.it
  • ×
  • Home
  • Corsi
  • Insegnamenti
  • Professioni
  • Persone
  • Pubblicazioni
  • Strutture
  • Terza Missione
  • Attività
  • Competenze
  1. Pubblicazioni

Evaluating Explainability of Graph Neural Networks for Network Intrusion Detection with Structural Attacks

Contributo in Atti di convegno
Data di Pubblicazione:
2025
Citazione:
Evaluating Explainability of Graph Neural Networks for Network Intrusion Detection with Structural Attacks / Galli, D.; Venturi, A.; Marasco, I.; Marchetti, M.. - 3962:(2025). ( 2025 Joint National Conference on Cybersecurity, ITASEC and SERICS 2025 Alma Mater Studiorum University, ita 2025).
Abstract:
Among Machine Learning (ML) models, Graph Neural Networks (GNN) have been shown to improve the performance of modern Network Intrusion Detection Systems (NIDS). However, their black-box nature poses a significant challenge to their practical deployment in the real world. In this context, researchers have developed eXplainable Artificial Intelligence (XAI) methods that reveal the inner workings of GNN models. Despite this, determining the most effective explainer is complex because different methods yield different explanations, and there are no standardized strategies. In this paper, we present an innovative approach for evaluating XAI methods in GNN-based NIDS. We evaluate explainers based on their capability to identify key graph components that an attacker can exploit to bypass detection. More accurate XAI algorithms can identify topological vulnerabilities, resulting in more effective attacks. We assess the effectiveness of different explainers by measuring the severity of structural attacks guided by the corresponding explanations. Our case study compares five XAI techniques on two publicly available datasets containing real-world network traffic. Results show that the explainer based on Integrated Gradients (IG) generates the most accurate explanations, allowing attackers to refine their strategies.
Tipologia CRIS:
Relazione in Atti di Convegno
Keywords:
Explainable Artificial Intelligence; Graph Neural Network; Network Intrusion Detection
Elenco autori:
Galli, D.; Venturi, A.; Marasco, I.; Marchetti, M.
Autori di Ateneo:
GALLI DIMITRI
MARCHETTI Mirco
Link alla scheda completa:
https://iris.unimore.it/handle/11380/1379588
Link al Full Text:
https://iris.unimore.it//retrieve/handle/11380/1379588/899898/paper50.pdf
Titolo del libro:
CEUR Workshop Proceedings
Pubblicato in:
CEUR WORKSHOP PROCEEDINGS
Journal
CEUR WORKSHOP PROCEEDINGS
Series
  • Utilizzo dei cookie

Realizzato con VIVO | Designed by Cineca | 26.5.0.0