Skip to Main Content (Press Enter)

Logo UNIPV
  • ×
  • Home
  • Degrees
  • Courses
  • Jobs
  • People
  • Outputs
  • Organizations

UNIFIND
Logo UNIPV

|

UNIFIND

unipv.it
  • ×
  • Home
  • Degrees
  • Courses
  • Jobs
  • People
  • Outputs
  • Organizations
  1. Outputs

How robust are ensemble machine learning explanations?

Academic Article
Publication Date:
2025
abstract:
To date, several explainable AI methods are available. The variability of the resulting explanations can be high, especially when many input features are considered. This lack of robustness may limit their usability. In this paper we try to fill this gap, by contributing a methodology that: i) is able to measure the robustness of a given set of explanations; ii) suggests how to improve robustness, by tuning the model parameters. Without loss of generality, we exemplify our proposal for ensemble tree models, which typically reach a high predictive performance in classification problems. We consider a toy case study with artificially generated data as well as two real case studies whose application domain is cybersecurity and more precisely the models used for detecting phishing websites.
Iris type:
1.1 Articolo in rivista
Keywords:
Concentration; Cybersecurity; Ensemble tree models; Explainable AI; Machine learning; Phishing detection; Robustness
List of contributors:
Calzarossa, Maria Carla; Giudici, Paolo; Zieni, Rasha
Authors of the University:
CALZAROSSA MARIA
GIUDICI PAOLO STEFANO
ZIENI Rasha
Handle:
https://iris.unipv.it/handle/11571/1520956
Published in:
NEUROCOMPUTING
Journal
  • Use of cookies

Powered by VIVO | Designed by Cineca | 26.4.0.0