Acta Electrotechnica et Informatica (Sep 2024)
Interpretability Using Reconstruction of Capsule Networks
Abstract
This paper evaluates the effectiveness of different decoder architectures in enhancing the reconstruction quality of Capsule Neural Networks (CapsNets), which impacts model interpretability. We compared linear, convolutional, and residual decoders to assess their performance in improving CapsNet reconstructions. Our experiments revealed that the Conditional Variational Autoencoder Capsule Network (CVAECapOSR) achieved the best reconstruction quality on the CIFAR-10 dataset, while the residual decoder outperformed others on the Brain Tumor MRI dataset. These findings highlight how improved decoder architectures can generate reconstructions of better quality, which can enhance changes by deforming output capsules, thereby making the feature extraction and classification processes within CapsNets more transparent and interpretable. Additionally, we evaluated the computational efficiency and scalability of each decoder, providing insights into their practical deployment in real-world applications such as medical diagnostics and autonomous driving.
Keywords