Substantial progress in spoofing and deepfake detection has been made in recent years. Nonetheless, the community has yet to make notable inroads in providing an explanation for how a classifier produces its output. The dominance of black box spoofing detection solutions is at further odds with the drive toward trustworthy, explainable artificial intelligence. This paper describes our use of SHapley Additive exPlanations (SHAP) to gain new insights in spoofing detection. We demonstrate use of the tool in revealing unexpected classifier behaviour, the artefacts that contribute most to classifier outputs and differences in the behaviour of competing spoofing detection models. The tool is both efficient and flexible, being readily applicable to a host of different architecture models in addition to related, different applications. All results reported in the paper are reproducible using open-source software.
Explaining deep learning models for spoofing and deepfake detection with shapley additive explanations
Submitted to ICASSP 2022, 7 October 2021
© EURECOM. Personal use of this material is permitted. The definitive version of this paper was published in Submitted to ICASSP 2022, 7 October 2021 and is available at :
PERMALINK : https://www.eurecom.fr/publication/6699