adversarial samples. In addition, eXplainable Artificial Intelligence (XAI) has been recently investigated to improve the interpretability and explainability of black-box artificial systems such as deep neural models. In this study, we propose a methodology that combines adversarial training and XAI, in order to increase the accuracy of deep neural models trained for cyber-threat detection. In particular, we use the FGSM technique to generate the adversarial samples for the adversarial training stage, and SHAP to produce the local explanations of decisions made during the adversarial training stage. These local explanations are, subsequently, used to produce a new feature set that describes the effect of the original cyber-data characteristics on the classifications of the examples processed during the adversarial training stage. Leveraging this XAI-based information, we apply a transfer learning strategy, namely fine-tuning, to improve the accuracy performance of the deep neural model. Experiments conducted on two benchmark cybersecurity datasets prove the effectiveness of the proposed methodology in the multi-class classification of cyber-data.

An XAI-based adversarial training approach for cyber-threat detection

Andresini G.;Appice A.;Malerba D.
2022-01-01

Abstract

adversarial samples. In addition, eXplainable Artificial Intelligence (XAI) has been recently investigated to improve the interpretability and explainability of black-box artificial systems such as deep neural models. In this study, we propose a methodology that combines adversarial training and XAI, in order to increase the accuracy of deep neural models trained for cyber-threat detection. In particular, we use the FGSM technique to generate the adversarial samples for the adversarial training stage, and SHAP to produce the local explanations of decisions made during the adversarial training stage. These local explanations are, subsequently, used to produce a new feature set that describes the effect of the original cyber-data characteristics on the classifications of the examples processed during the adversarial training stage. Leveraging this XAI-based information, we apply a transfer learning strategy, namely fine-tuning, to improve the accuracy performance of the deep neural model. Experiments conducted on two benchmark cybersecurity datasets prove the effectiveness of the proposed methodology in the multi-class classification of cyber-data.
2022
978-1-6654-6297-6
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11586/429193
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
social impact