Have a personal or library account? Click to login
A Visually Explainable Dynamic Similarity Network for Few-Shot Classification Cover

A Visually Explainable Dynamic Similarity Network for Few-Shot Classification

Open Access
|Feb 2026

References

  1. F. Z. Mehrjardi, A. M. Latif, M. S. Zarchi, R. Sheikhpour, A survey on deep learning-based image forgery detection, Pattern Recognition, 144, 2023, 109778.
  2. X. Li, X. Yang, Z. Ma, J.-H. Xue, Deep metric learning for few-shot image classification: A review of recent developments, Pattern Recognition, 138, 2023, 109381.
  3. S. Lee, W. Moon, J.-P. Heo, Task discrepancy maximization for fine-grained few-shot classification, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, 5331–5340.
  4. D. Kang, H. Kwon, J. Min, M. Cho, Relational embedding for few-shot classification, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, 8822–8833.
  5. W. Chen, C. Si, Z. Zhang, L. Wang, Z. Wang, T. Tan, Semantic prompt for few-shot image recognition, arXiv preprint arXiv:2303.14123, 2023.
  6. B. Dong, P. Zhou, S. Yan, W. Zuo, Self-promoted supervision for few-shot transformer, In: European Conference on Computer Vision, Springer, 2022, 329–347.
  7. X. Zhang, D. Meng, H. Gouk, T. M. Hospedales, Shallow bayesian meta learning for real-world few-shot recognition, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, 651–660.
  8. M. Hiller, R. Ma, M. Harandi, T. Drummond, Re-thinking generalization in few-shot classification, In: Advances in Neural Information Processing Systems, 35, 2022, 3582–3595.
  9. F. Hao, F. He, L. Liu, F. Wu, D. Tao, J. Cheng, Class-aware patch embedding adaptation for few-shot image classification, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, 18905–18915.
  10. K. Wu, J. Zhang, H. Peng, M. Liu, B. Xiao, J. Fu, L. Yuan, TinyViT: Fast pretraining distillation for small vision transformers, In: European Conference on Computer Vision, Springer, 2022, 68–85.
  11. P. Bateni, R. Goyal, V. Masrani, F. Wood, L. Sigal, Improved few-shot visual classification, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, 14493–14502.
  12. A. Parnami, M. Lee, Learning from few examples: A summary of approaches to few-shot learning, arXiv preprint arXiv:2203.04291, 2022.
  13. F. Zhuang, Z. Qi, K. Duan, D. Xi, Y. Zhu, H. Zhu, H. Xiong, Q. He, A comprehensive survey on transfer learning, Proceedings of the IEEE, 109(1), 2020, 43–76.
  14. H.-J. Ye, H. Hu, D.-C. Zhan, F. Sha, Few-shot learning via embedding adaptation with set-to-set functions, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, 8808–8817.
  15. T. Ridnik, E. Ben-Baruch, A. Noy, L. Zelnik-Manor, ImageNet-21K Pretraining for the Masses, arXiv preprint arXiv:2104.10972, 2021.
  16. M. Dehghani, J. Djolonga, B. Mustafa, P. Padlewski, J. Heek, J. Gilmer, A. P. Steiner, M. Caron, R. Geirhos, I. Alabdulmohsin et al., Scaling vision transformers to 22 billion parameters, In: International Conference on Machine Learning, PMLR, 2023, 7480–7512.
  17. J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, L. Fei-Fei, ImageNet: A large-scale hierarchical image database, In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, IEEE, 2009, 248–255.
  18. M. Ren, E. Triantafillou, S. Ravi, J. Snell, K. Swersky, J. B. Tenenbaum, H. Larochelle, R. S. Zemel, Meta-learning for semi-supervised few-shot classification, arXiv preprint arXiv:1803.00676, 2018.
  19. L. Bertinetto, J. F. Henriques, P. H. Torr, A. Vedaldi, Meta-learning with differentiable closed-form solvers, arXiv preprint arXiv:1805.08136, 2018.
  20. W.-Y. Chen, Y.-C. Liu, Z. Kira, Y.-C. F. Wang, J.-B. Huang, A closer look at few-shot classification, arXiv preprint arXiv:1904.04232, 2019.
  21. E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, W. Chen et al., LoRA: Low-Rank Adaptation of Large Language Models, ICLR, 1(2), 2022, 3.
  22. N. Houlsby, A. Giurgiu, S. Jastrzebski, B. Morrone, Q. De Laroussilhe, A. Gesmundo, M. Attariyan, S. Gelly, Parameter-efficient transfer learning for NLP, In: International Conference on Machine Learning, PMLR, 2019, 2790–2799.
  23. D. Wertheimer, L. Tang, B. Hariharan, Few-shot classification with feature map reconstruction networks, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, 8012–8021.
  24. C. Zhang, Y. Cai, G. Lin, C. Shen, DeepEMD: Few-shot image classification with differentiable earth mover’s distance and structured classifiers, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, 12203–12213.
  25. C. Cao, Y. Zhang, Learning to compare relation: Semantic alignment for few-shot learning, IEEE Transactions on Image Processing, 31, 2022, 1462–1474.
  26. G. Koch, R. Zemel, R. Salakhutdinov et al., Siamese neural networks for one-shot image recognition, In: ICML Deep Learning Workshop, 2, Lille, 2015, 1–30.
  27. J. Snell, K. Swersky, R. Zemel, Prototypical networks for few-shot learning, In: Advances in Neural Information Processing Systems, 30, 2017.
  28. X. Li, J. Wu, Z. Sun, Z. Ma, J. Cao, J.-H. Xue, BSNet: Bi-Similarity Network for Few-shot Fine-grained Image Classification, IEEE Transactions on Image Processing, 30, 2020, 1318–1331.
  29. O. Vinyals, C. Blundell, T. Lillicrap, D. Wierstra et al., Matching networks for one shot learning, In: Advances in Neural Information Processing Systems, 29, 2016.
  30. J. Wu, T. Zhang, Y. Zhang, F. Wu, Task-aware part mining network for few-shot learning, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, 8433–8442.
  31. H. Chen, H. Li, Y. Li, C. Chen, Sparse spatial transformers for few-shot learning, Science China Information Sciences, 66(11), 2023, 210102.
  32. L. Li, B. Wang, M. Verma, Y. Nakashima, R. Kawasaki, H. Nagahara, SCOUTER: Slot Attention-based Classifier for Explainable Image Recognition, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, 1046–1055.
  33. G. Qi, H. Yu, Z. Lu, S. Li, Transductive few-shot classification on the oblique manifold, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, 8412–8422.
  34. J. Sun, J. Li, Few-shot classification with fork attention adapter, Pattern Recognition, 156, 2024, 110805.
  35. H. Cheng, S. Yang, J. T. Zhou, L. Guo, B. Wen, Frequency guidance matters in few-shot learning, In: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, 11814–11824.
  36. Z. Sun, W. Zheng, P. Guo, M. Wang, TST MFL: Two-stage training based metric fusion learning for few-shot image classification, Information Fusion, 113, 2025, 102611.
  37. Y. He, W. Liang, D. Zhao, H.-Y. Zhou, W. Ge, Y. Yu, W. Zhang, Attribute surrogates learning and spectral tokens pooling in transformers for few-shot learning, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, 9119–9129.
  38. T. Wu, T. Kobayashi, Geometric Mean Improves Loss For Few-Shot Learning, arXiv preprint arXiv:2501.14593, 2025.
  39. V. Petsiuk, A. Das, K. Saenko, Rise: Randomized input sampling for explanation of black-box models, arXiv preprint arXiv:1806.07421, 2018.
  40. Y. Rong, X. Lu, Z. Sun, Y. Chen, S. Xiong, ESPT: A self-supervised episodic spatial pretext task for improving few-shot learning, In: Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 2023, 9596–9605.
  41. X. Li, Z. Li, J. Xie, X. Yang, J.-H. Xue, Z. Ma, Self-reconstruction network for fine-grained few-shot classification, Pattern Recognition, 153, 2024, 110485.
  42. Z. Li, Z. Hu, W. Luo, X. Hu, SaberNet: Self-attention based effective relation network for few-shot learning, Pattern Recognition, 133, 2023, 109024.
  43. W. Samek, G. Montavon, S. Lapuschkin, C. J. Anders, K.-R. Müller, Explaining deep neural networks and beyond: A review of methods and applications, Proceedings of the IEEE, 109(3), 2021, 247–278.
  44. L. Grinsztajn, E. Oyallon, G. Varoquaux, Why do tree-based models still outperform deep learning on typical tabular data?, In: Advances in Neural Information Processing Systems, 35, 2022, 507–520.
  45. B. Wang, L. Li, Y. Nakashima, H. Nagahara, Learning bottleneck concepts in image classification, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, 10962–10971.
  46. F. Bodria, F. Giannotti, R. Guidotti, F. Naretto, D. Pedreschi, S. Rinzivillo, Benchmarking and survey of explanation methods for black box models, Data Mining and Knowledge Discovery, 37(5), 2023, 1719–1778.
  47. C. Wah, S. Branson, P. Welinder, P. Perona, S. Belongie, The Caltech-UCSD Birds-200-2011 Dataset, California Institute of Technology, Technical Report CNS-TR-2011-001, 2011.
  48. J. Zeng, Z. Xue, L. Zhang, Q. Lan, M. Zhang, Multistage relation network with dual-metric for few-shot hyperspectral image classification, IEEE Transactions on Geoscience and Remote Sensing, 61, 2023, 1–17.
  49. F. Sung, Y. Yang, L. Zhang, T. Xiang, P. H. Torr, T. M. Hospedales, Learning to compare: Relation network for few-shot learning, In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, 1199–1208.
  50. X. Li, Q. Song, J. Wu, R. Zhu, Z. Ma, J.-H. Xue, Locally-enriched cross-reconstruction for few-shot fine-grained image classification, IEEE Transactions on Circuits and Systems for Video Technology, 33(12), 2023, 7530–7540.
  51. Y. Liu, W. Zhang, C. Xiang, T. Zheng, D. Cai, X. He, Learning to affiliate: Mutual centralized learning for few-shot classification, In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, 14411–14420.
  52. J. Zhuang, T. Tang, Y. Ding, S. C. Tatikonda, N. Dvornek, X. Papademetris, J. Duncan, Adabelief optimizer: Adapting stepsizes by the belief in observed gradients, In: Advances in neural information processing systems, 33, 2020, 18795–18806.
  53. R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, D. Batra, Grad-cam: Visual explanations from deep networks via gradient-based localization, In: Proceedings of the IEEE international conference on computer vision, 2017, 618–626.
  54. H. Wang, Z. Wang, M. Du, F. Yang, Z. Zhang, S. Ding, P. Mardziel, X. Hu, Score-cam: Score-weighted visual explanations for convolutional neural networks, In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, 2020, 24–25.
  55. J. Qian, T. Wen, M. Ling, X. Du, H. Ding, Pixel-based clustering for local interpretable model-agnostic explanations, Journal of Artificial Intelligence and Soft Computing Research, 15, 2025.
  56. H. Liu, C. Wang, X. Jiang, M. Khishe, A few-shot learning approach for covid-19 diagnosis using quasi-configured topological spaces, Journal of Artificial Intelligence and Soft Computing Research, 14(1), 2023, 77–95.
  57. N. Shakhovska, A. Shebeko, Y. Prykarpatskyy, A novel explainable ai model for medical data analysis, Journal of Artificial Intelligence and Soft Computing Research, 14, 2024.
Language: English
Page range: 237 - 256
Accepted on: Dec 27, 2025
|
Published on: Feb 25, 2026
Published by: SAN University
In partnership with: Paradigm Publishing Services
Publication frequency: 4 issues per year

© 2026 Zirui Pei, Zuqiang Meng, Tingting Diao, Peng Miao, Yifan Meng, Chaohong Tan, published by SAN University
This work is licensed under the Creative Commons Attribution 4.0 License.