Artículo
Acceso Abierto

Rule Extraction in Trained Feedforward Deep Neural Networks: Integrating Cosine Similarity and Logic for Explainability

Resumen

Explainability is a key aspect of machine learning, necessary for ensuring transparency and trust in decision-making processes. As machine learning models become more complex, the integration of neural and symbolic approaches has emerged as a promising solution to the explainability problem. One effective solution involves using search techniques to extract rules from trained deep neural networks by examining weight and bias values and calculating their correlation with outputs. This article proposes incorporating cosine similarity in this process to narrow down the search space and identify the critical path connecting inputs to final results. Additionally, the integration of first-order logic (FOL) is suggested to provide a more comprehensive and interpretable understanding of the decision-making process. By leveraging cosine similarity and FOL, an innovative algorithm capable of extracting and explaining rule patterns learned by a feedforward trained neural network was developed and tested in two use cases, demonstrating its effectiveness in providing insights into model behavior.

Palabras clave
Artificial Intelligence
Black Box Models
Cosine Similarity
Deep Learning
Distance Function
Entropy
Explainability
Feedforward Neural Network
Logic
Regularization
Rule Extraction
http://creativecommons.org/licenses/by/4.0/

Esta obra se publica con la licencia Creative Commons Attribution 4.0 International (BY 4.0)

item.page.license
Cargando...
Miniatura