Have a personal or library account? Click to login
Predicting Vehicle Pose in Six Degrees of Freedom from Single Image in Real-World Traffic Environments Using Deep Pretrained Convolutional Networks and Modified Centernet Cover

Predicting Vehicle Pose in Six Degrees of Freedom from Single Image in Real-World Traffic Environments Using Deep Pretrained Convolutional Networks and Modified Centernet

Open Access
|Aug 2024

References

  1. Van Brummelen, Jessica, Marie O’Brien, Dominique Gruyer, and Homayoun Najjaran. “Autonomous vehicle perception: The technology of today and tomorrow.” Transportation research part C: emerging technologies 89 (2018): 384–406.
  2. Lefèvre, Stéphanie, Dizan Vasquez, and Christian Laugier. “A survey on motion prediction and risk assessment for intelligent vehicles.” ROBOMECH journal 1, no. 1 (2014): 1–14.
  3. Shladover, Steven E., Charles A. Desoer, J. Karl Hedrick, Masayoshi Tomizuka, Jean Walrand, W-B. Zhang, Donn H. McMahon, Huei Peng, Shahab Sheikholeslam, and Nick McKeown. “Automated vehicle control developments in the PATH program.” IEEE Transactions on vehicular technology 40, no. 1 (1991): 114–130.
  4. Kundu, Abhijit, Yin Li, and James M. Rehg. “3D-RCNN: Instance-level 3d object reconstruction via render-and-compare.” In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3559–3568. 2018.
  5. Chabot, Florian, Mohamed Chaouch, Jaonary Rabarisoa, Céline Teuliere, and Thierry Chateau. “Deep manta: A coarse-to-fine many-task network for joint 2d and 3d vehicle analysis from monocular image.” In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2040–2049. 2017.
  6. Wandinger U., “Introduction to LiDAR.” In LiDAR, (pp. 1–18). Springer, New York, NY, 2005.
  7. Arnold E., Al-Jarrah O.Y., Dianati M., Fallah S., Oxtoby D., Mouzakitis A. A., “survey on 3d object detection methods for autonomous driving applications.” IEEE Transactions on Intelligent Transportation Systems, 20, 3782–3795, 2019.
  8. Waldschmidt C., Hasch J., Menzel W., “Automotive radar—from first efforts to future systems.” IEEE Journal of Microwaves, 1(1), 135–148, 2021.
  9. He K., Zhang X., Ren S., Sun J., “Deep residual learning for image recognition.” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 770–778), 2016.
  10. Xie S., Girshick R., Dollár P., Tu Z., He K., “Aggregated residual transformations for deep neural networks.” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (pp. 1492–1500), 2017.
  11. Szegedy C., Ioffe S., Vanhoucke V., Alemi A. A., “Inception-v4, inception-ResNet, and the impact of residual connections on learning.” In the Thirty-first AAAI Conference on Artificial Intelligence, 2017.
  12. Huang G., Liu Z., Van Der Maaten L., Weinberger K. Q., “Densely connected convolutional networks.” In Proceedings of the IEEE conference on computer vision and pattern recognition, (pp. 4700–4708), 2017.
  13. Duan K., Bai S., Xie L., Qi H., Huang Q., Tian Q., “Centernet: Keypoint triplets for object detection.” In Proceedings of the IEEE/CVF International Conference on Computer Vision, (pp. 6569–6578), 2019.
  14. Kolekar S., Gite S., Pradhan B., Kotecha K., “Behavior Prediction of Traffic Actors for Intelligent Vehicle Using Artificial Intelligence Techniques: A Review.” IEEE Access, 9, 135034–135058, 2021.
  15. Everingham M., Van Gool L., Williams C. K. I., Winn J., Zisserman A., “The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results.”, http://www.pascalnetwork.org/challenges/VOC/voc2012/workshop/index.html.
  16. Kim J., Hong J., Park H., “Prospects of deep learning for medical imaging.” Precision and Future Medicine, 2(2), 37–52, 2018.
  17. He K., Gkioxari G., Dollar P., Girshick R., “Mask R-CNN.” In Proc. IEEE Int. Conf. Comp. Vis. pages 2980–2988. IEEE, 2017.
  18. Deng J., Dong W., Socher R., Li L.-J., Li K., Fei-Fei L., “Imagenet: A large-scale hierarchical image database.” In Proc. IEEE Conf. Comp. Vis. Patt. Recogn, pages 248–255. Ieee, 2009.
  19. Lin T.-Y., Maire M., Belongie S., Hays J., Perona P., Ramanan D., Dollar P., Zitnick C. L., “Microsoft coco: Common objects in context.” In Proc. Eur. Conf. Comp. Vis. pages 740–755. Springer, 2014.
  20. Geiger A., Lenz P., Urtasun R., “Are we ready for autonomous driving? The kitti vision benchmark suite.” In Proc. IEEE Conf. Comp. Vis. Patt. Recogn. Pages 3354–3361. IEEE, 2012.
  21. Lu F., Liu Z., Song X., Zhou D., Li W., Miao H., Manocha D., “Vehicle Perception from a Single Image for Autonomous Driving Using Deformable Model Representation and Deep Learning.” SSRN 4085267, 2022.
  22. Ke L., Li S., Sun Y., Tai Y. W., Tang C. K., “GSNet: Joint vehicle pose and shape reconstruction with geometrical and scene-aware supervision.” In European Conference on Computer Vision, Springer, Cham, (pp. 515–532), 2020.
  23. An K., Chen Y., Wang S., Xiao Z., “RCBi-CenterNet: an absolute pose policy for 3D object detection in autonomous driving.” Applied Sciences, 11(12), 5621, 2021.
  24. Pan S. J., Yang Q. A, “survey on transfer learning.” IEEE Transactions on Knowledge and Data Engineering, 22(10), 1345–1359, 2009.
  25. Khade S., Gite S., Pradhan B., “Iris Liveness Detection Using Multiple Deep Convolution Networks.” Big Data and Cognitive Computing, 6(2), 67, 2022.
  26. Samet N., Hicsonmez S., & Akbas E., “HoughNet: Integrating near and long-range evidence for visual detection.” IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(4), 4667–4681, 2022.
  27. Rawat W., & Wang Z., “Deep convolutional neural networks for image classification: A comprehensive review.” Neural Computation, 29(9), 2352–2449, 2017.
  28. Koonce, B., & Koonce, B., “ResNet50. Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization.” 63–72, 2021.
  29. Carranza-García M., Torres-Mateo J., Lara-Benítez P., & García-Gutiérrez J., “On the performance of one-stage and two-stage object detectors in autonomous vehicles using camera data.” Remote Sensing, 13(1), 89, 2020.
  30. Xie S., Girshick R., Dollár P., Tu Z., & He K., “Aggregated residual transformations for deep neural networks.” In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1492–1500), 2017.
  31. Kolekar Suresh, Shilpa Gite, Biswajeet Pradhan, and Abdullah Alamri, “Explainable AI in Scene Understanding for Autonomous Vehicles in Unstructured Traffic Environments on Indian Roads Using the Inception U-Net Model with Grad-CAM Visualization.” Sensors 22, no. 24: 9677, 2022. https://doi.org/10.3390/s22249677.
  32. Valev K., Schumann A., Sommer L., & Beyerer J., “A systematic evaluation of recent deep learning architectures for fine-grained vehicle classification.” In Pattern Recognition and Tracking XXIX (Vol. 10649, p. 1064902). SPIE. 2018.
  33. Wang S. H., & Zhang Y. D., DenseNet-201-based deep neural network with composite learning factor and precomputation for multiple sclerosis classification. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), 16(2s), 1–19, 2020.
  34. Song X., Wang P., Zhou D., Zhu R., Guan C., Dai Y., Yang R. “Apollocar3d: A large 3d car instance understanding benchmark for autonomous driving.” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, (pp. 5452–5462), 2019.
  35. Rouben, Y. F. D. “Pose estimation using overhead imagery and semantics.” (Doctoral dissertation, Massachusetts Institute of Technology), 2018.
  36. Llugsi R., El Yacoubi S., Fontaine A., Lupera P., “Comparison between Adam, AdaMax, and Adam W optimizers to implement a Weather Forecast based on Neural Networks for the Andean city of Quito.” In 2021 IEEE Fifth Ecu, 2021.
Language: English
Submitted on: Apr 18, 2024
Published on: Aug 6, 2024
Published by: Professor Subhas Chandra Mukhopadhyay
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2024 Suresh Kolekar, Shilpa Gite, Biswajeet Pradhan, Abdulla Alamri, published by Professor Subhas Chandra Mukhopadhyay
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.