Have a personal or library account? Click to login
Three-Dimensional Line-of-Sight Estimation Based on RTACM-Net and Vision Transformer Cover

Three-Dimensional Line-of-Sight Estimation Based on RTACM-Net and Vision Transformer

By: Tingjuan Sang and  Wuqi Gao  
Open Access
|Dec 2025

References

  1. Philipe Ambrozio Dias, Damiano Malafronte, Henry Medeiros, et al. Gaze estimation for assisted living environments [J]. Computing Research Repository (CoRR), 2019.
  2. Zhang XC, Sugano Y, Fritz M, et al. MPIIGaze: Real world dataset and deep appearance-based gaze estimation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019, 41(1): 162–175. [doi: 10.1109/tpami.2017.2778103]
  3. Chen ZK, Shi BE. Appearance-based gaze estimation using dilated-convolutions. Proceedings of the 14th Asian Conference on Computer Vision. Perth: Springer, 2019. 309–324. [doi: 10.1007/978-3-030-20876-9_20]
  4. Cheng YH, Huang SY, Wang F, et al. A coarse-to-fine adaptive network for appearance-based gaze estimation. Proceedings of the 34th AAAI Conference on Artificial Intelligence. New York: AAAI Press, 2020. 1062310630. [doi:10.1609/aaai.v34i07.6636].
  5. Murthy LRD, Biswas P. Appearance-based gaze estimation using attention and difference mechanism. Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Nashville: IEEE, 2021. 3143–3152. [doi: 10.1109/cvprw53098.2021.00351]
  6. Cheng YH, Lu F. Gaze estimation using Transformer. Proceedings of the 26th International Conference on Pattern Recognition. Montreal: IEEE, 2022. 3341–3347. [doi: 10. 1109/icpr56361.2022.9956687]
  7. Abdelrahman A, Hempel T, Khalifa A, et al. L2CS-Net: Fine-grained gaze estimation in unconstrained environments. Proceedings of the 8th International Conference on Frontiers of Signal Processing. Corfu: IEEE, 2023. 98–102. [doi: 10.1109/icfsp59764.2023.10372944]
  8. Li YJ, Chen JH, Ma JX, et al. Gaze estimation based on convolutional structure and sliding window-based attention mechanism. Sensors, 2023, 23(13): 6226. [doi: 10.3390/s23136226]
  9. Tobias Fischer, Hyung Jin Chang, Yiannis Demiris. RT-GENE: Real-Time eye gaze estimation in natural environments[C]//Proceedings of the European Conference on Computer Vision (ECCV). 2018: 339–357.
  10. Zhang XC, Sugano Y, Fritz M, et al. It’s written all over your face: Full-face appearance-based gaze estimation. Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops. Honolulu: IEEE, 51 - 60. [doi: 10.1109/cvprw.2017.284]
  11. Fischer T, Chang HJ, Demiris Y. RT-GENE: Real-time eye gaze estimation in natural environments. Proceedings of the 15th European Conference on Computer Vision. Munich: Springer, 2018. 334–352. [doi: 10.1007/978-3-030-01249-6_21]
  12. Wang XH, Zhou J, Wang L, et al. BoT2L-Net: Appearancebased gaze estimation using bottleneck Transformer block and two identical losses in unconstrained environments. Electronics, 2023, 12(7): 1704. [doi: 10.3390/electronics 12071704]
  13. Petr Kellnhofer, Adria Recasens, Simon Stent, et al. Gaze360: Physically unconstrained gaze estimation in the wild [J]. Computing Research Repository (CoRR), 2019.
  14. Zhang XC, Sugano Y, Fritz M, et al. Appearance-based gaze estimation in the wild. Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition. Boston: IEEE, 2015. 4511 - 4520. [doi: 10.1109/cvpr.2015. 7299081]
  15. X. Pan et al., "On the Integration of Self-Attention and Convolution," 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 2022, pp. 805-815, doi: 10.1109/CVPR52688.2022.00089.
  16. Fernandez A. TeLU activation function for fast and stable deep learning [D]. University of South Florida, 2024.
Language: English
Page range: 12 - 21
Published on: Dec 31, 2025
In partnership with: Paradigm Publishing Services
Publication frequency: 4 issues per year

© 2025 Tingjuan Sang, Wuqi Gao, published by Xi’an Technological University
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.