References
- Wang, L., & Zhang, Y., “Economic and Environmental Costs of Urban Traffic Congestion,” Sustainable Cities and Society, vol. 89, p. 104328, 2023.
- M. C. Surgiu, I. M. Moise and E. A. Stanciu, “Studiu privind metode optime de detecție a vehiculelor în București,” AGIR, 2012.
- Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You Only Look Once: Unified, Real-Time Object Detection. În Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 779-788.
- Fang, W., & Chien, S., “Optimizing Commercial Vehicle Traffic in Urban Areas,” Transportation Research Part D: Transport and Environment, vol. 45, no. 3, pp. 112–125, 2022.
- Atoo, A. and Adeke, P., “Level of Service (LOS) of Freeway Segments within Makurdi Metropolis,” 2018.
- N. Dalal and B. Triggs, “Histograms of Oriented Gradients for Human Detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), San Diego, CA, USA, 2005, pp. 886–893.
- S. Ren, K. He, R. Girshick and J. Sun, “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 6, pp. 1137-1149, Jun. 2017.
- S. Li, X. Yang, X. Lin, Y. Zhang, and J. Wu, “Real-Time Vehicle Detection from UAV Aerial Images Based on Improved YOLOv5,” Sensors, vol. 23, no. 12, Jun. 2023, doi: 10.3390/s23125634.
- A. Geiger, P. Lenz and R. Urtasun, “Are we ready for autonomous driving? The KITTI vision benchmark suite,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2012, pp. 3354–3361.
- L. Wen et al., “UA-DETRAC: A new benchmark and protocol for multi-object detection and tracking,” Computer Vision and Image Understanding, vol. 193, Nov. 2015, doi: 10.1016/j.cviu.2020.102907.
- T. Y. Lin et al., “Microsoft COCO: Common objects in context,” in Proc. Eur. Conf. Comput. Vis., 2014, pp. 740–755.
- J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real time object detection,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 779–788.
- J. Redmon and A. Farhadi, “YOLOv3: An Incremental Improvement,” arXiv preprint arXiv:1804.02767, 2018.
- A. Bochkovskiy, C. Y. Wang and H. Y. M. Liao, “YOLOv4: Optimal Speed and Accuracy of Object Detection,” arXiv preprint arXiv:2004.10934, 2020.
- J. Jocher et al., “YOLOv8: A Cutting-edge Real-Time Object Detection Model,” Ultralytics, 2023.
- M. Cordts et al., “The Cityscapes Dataset for Semantic Urban Scene Understanding,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016.
- J. Lin et al., “Edge AI: On-demand deep learning model co-inference with device-edge synergy,” ACM SIGCOMM, 2019.
- https://raw.githubusercontent.com/ultralytics/assets/refs/heads/main/yolo/performance-comparison.png
- https://medium.com/analyticsvidhya/haar-cascades-explained-38210e57970d
- J. Terven and D. Cordova-Esparza, “A Comprehensive Review of YOLO Architectures in Computer Vision: From YOLOv1 to YOLOv8 and YOLO-NAS,” Machine Learning and Knowledge Extraction, vol. 5, no. 4, pp. 1680–1716, 2023.
- S. Pudaruth, I. M. Boodhun, and C. W. Onn, “Reducing Traffic Congestion Using Real-Time Traffic Monitoring with YOLOv8,” International Journal of Advanced Computer Science and Applications, vol. 15, no. 10, pp. 1068–1075, 2024.
- H. Lou, X. Duan, J. Guo, H. Liu, J. Gu, and L. Bi, “DC-YOLOv8: Small-Size Object Detection Algorithm Based on Camera Sensor,” Electronics, vol. 12, no. 10, p. 2323, 2023.