Rule Extraction in Trained Feedforward Deep Neural Networks - Integrating Cosine Similarity and Logic for Explainability
Rule Extraction in Trained Feedforward Deep Neural Networks - Integrating Cosine Similarity and Logic for Explainability
| dc.contributor.author | Negro, Pablo Ariel | |
| dc.contributor.author | Pons, Claudia Fabiana | |
| dc.date.accessioned | 2026-03-24T13:41:40Z | |
| dc.date.available | 2026-03-24T13:41:40Z | |
| dc.date.issued | 2024-12-30 | |
| dc.description.abstract | Explainability is a fundamental aspect in the field of machine learning, particularly in ensuring transparency and trust in decision-making processes. As the complexity of machine learning models increases, the integration of neural and symbolic approaches has emerged as a promising solution to the explainability problem. In this context, the utilization of search methods for rule extraction in trained deep neural networks has been proven effective. This involves the examination of weight and bias values generated by the network, typically through calculating the correlation between weight vectors and outputs. The hypothesis developed in this article states that by incorporating cosine similarity in this process, the search space can be efficiently narrowed down to identify the critical path connecting inputs to results. Furthermore, to provide a more comprehensive and interpretable understanding of the decision making process, this article proposes the integration of first-order logic (FOL) in the rule extraction process. By leveraging cosine similarity and FOL, a groundbreaking algorithm that is capable of extracting and explaining the rule patterns learned by a feedforward trained neural network was designed and implemented. The algorithm was tested in three use cases showing effectiveness in providing insights into the model’s behavior. | |
| dc.identifier.citation | Negro, P. A., & Pons, C. (2024). Rule Extraction in Trained Feedforward Deep Neural Networks: Integrating Cosine Similarity and Logic for Explainability In: International Journal Of Artificial Intelligence And Machine Learning, 13(1), 1-22. | |
| dc.identifier.other | https://doi.org/10.59471/raia2024203 | |
| dc.identifier.uri | https://repositorio.uai.edu.ar/handle/123456789/4745 | |
| dc.language.iso | en | |
| dc.publisher | Universidad Abierta Interamericana. Facultad de Tecnología Informática | |
| dc.subject | artificial intelligence | |
| dc.subject | cosine similarity | |
| dc.subject | deep learning | |
| dc.subject | explainability | |
| dc.subject | logic | |
| dc.subject | rule extraction | |
| dc.title | Rule Extraction in Trained Feedforward Deep Neural Networks - Integrating Cosine Similarity and Logic for Explainability | |
| dc.type | ARTICULO |
Files
Original bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- 0000755743.pdf
- Size:
- 446.12 KB
- Format:
- Adobe Portable Document Format
- Description:
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 1.71 KB
- Format:
- Item-specific license agreed to upon submission
- Description: