← Back to Portfolio
Neural Network Explainability
Preprint - Artificial Intelligence

A Lightweight Explainability Framework
for Neural Networks

Methods, Benchmarks, and Mobile Deployment

Yin LiMarch 2025DOI: 10.20944/preprints202503.1857.v1

Abstract

Explainability is increasingly crucial for real-world deployment of deep learning models, yet traditional explanation techniques can be prohibitively slow and memory-intensive on resource-constrained devices. This paper presents a novel lightweight explainability framework that significantly reduces the computational cost of generating explanations without compromising on quality. My approach focuses on an optimized Grad-CAM pipeline with sophisticated thresholding, advanced memory handling, and specialized evaluation metrics. I demonstrate speedups exceeding 300x over naive implementations while maintaining robust faithfulness and completeness scores. Through an extensive series of benchmarks, user studies, and statistical tests, I show that this framework is scalable, accurate, and deployable on edge devices such as Raspberry Pi, Android phones, and iPhones.

Key Contributions

  • Algorithmic Optimization: Novel thresholding technique retaining high-importance activations.
  • Unified Evaluation Suite: End-to-end benchmarking including speed, memory, and quality metrics.
  • Edge Deployment: Feasible execution on low-power devices (Raspberry Pi, Mobile).
  • Validation: User study (n=120) showing 67.6% faster interpretation time.

Performance Stats

300xSpeedup
<0.5MBMemory Overhead
60fpsOn Edge Devices

Methodology

The framework builds upon Grad-CAM but introduces Threshold-Based Simplification. We apply a dynamic percentile threshold $P$ to reduce the coarse localization map $L_c$ to a sparse mask. This drastically reduces the pixel processing load.

ALGORITHM SimplifyGradCAM(cam, threshold_percent)
    threshold = PERCENTILE(cam, 100 - threshold_percent)
    simplified_cam = ZEROS_LIKE(cam)
    FOR each pixel (i,j) in cam DO
        IF cam[i,j] > threshold THEN
            simplified_cam[i,j] = cam[i,j]
        END IF
    END FOR
    RETURN simplified_cam

We also support extensions for Vision Transformers (via attention map extraction) and RNNs (via temporal gradient analysis), ensuring the framework is model-agnostic.

Results & Discussion

Our evaluation confirms highly consistent performance across diverse domains including medical imaging and autonomous driving. The 10% thresholding strategy preserves 92% of the explanation faithfulness while reducing computation time by orders of magnitude compared to standard libraries like Captum.