Have a personal or library account? Click to login

DeConvolve: Towards Textually Explainable and Human Cognizable Convolutional Networks

Open Access
|Sep 2025

Abstract

Convolutional Neural Networks (CNNs) have demonstrated remarkable accuracy and are employed in different applications. However, adding existing CNNs to physics-aware frameworks can distort image features, reducing classification accuracy. To overcome this, a new term is added to the loss function to reduce distortions and highlight human-recognizable structures in the feature maps. The proposed DeConvolve is an explainability methodology for multimodal Large Language Models (LLM) on feature maps to extract human-understandable sub-steps and provide textual explanations for model inference. DeConvolve recognizes three major impediments when using LLMs to describe feature maps: scattered regions of interest within the feature map, large areas of interest, and conflicting learning across filters in each convolutional layer. Finally, explanations for specific toy examples are derived through weighted semantic averaging. The data is curated in the format of images, classes, and the rationale behind a professional’s classification to train a Contrastive Language–Image Pre-training (CLIP)-based model for generating robust explanations.

DOI: https://doi.org/10.2478/cait-2025-0020 | Journal eISSN: 1314-4081 | Journal ISSN: 1311-9702
Language: English
Page range: 22 - 38
Submitted on: Apr 14, 2025
Accepted on: Jun 30, 2025
Published on: Sep 25, 2025
Published by: Bulgarian Academy of Sciences, Institute of Information and Communication Technologies
In partnership with: Paradigm Publishing Services
Publication frequency: 4 issues per year

© 2025 Meeradevi,, Hrishikesh Haritas, Darshan Bankapure, Divyansh Mishra, published by Bulgarian Academy of Sciences, Institute of Information and Communication Technologies
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.