Show/Hide Menu
Hide/Show Apps
Logout
Türkçe
Türkçe
Search
Search
Login
Login
OpenMETU
OpenMETU
About
About
Open Science Policy
Open Science Policy
Open Access Guideline
Open Access Guideline
Postgraduate Thesis Guideline
Postgraduate Thesis Guideline
Communities & Collections
Communities & Collections
Help
Help
Frequently Asked Questions
Frequently Asked Questions
Guides
Guides
Thesis submission
Thesis submission
MS without thesis term project submission
MS without thesis term project submission
Publication submission with DOI
Publication submission with DOI
Publication submission
Publication submission
Supporting Information
Supporting Information
General Information
General Information
Copyright, Embargo and License
Copyright, Embargo and License
Contact us
Contact us
A Unified Monocular Vision-Based Driving Model for Autonomous Vehicles With Multi-Task Capabilities
Date
2025-01-01
Author
Azak, Salim
Bozkaya, Frat
Tığlıoğlu, Şükrücan
Yusefi, Abdullah
Durdu, Akif
Metadata
Show full item record
This work is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License
.
Item Usage Stats
21
views
0
downloads
Cite This
The recent progress in autonomous driving primarily relies on sensor-rich systems, encompassing radars, LiDARs, and advanced cameras, in order to perceive the environment. However, human-operated vehicles showcase an impressive ability to drive based solely on visual perception. This study introduces an end-to-end method for predicting the steering angle and vehicle speed exclusively from a monocular camera image. Alongside the color image, which conveys scene texture and appearance details, a monocular depth image and a semantic segmentation image are internally derived and incorporated, offering insights into spatial and semantic environmental structures. This results in a total of three input images. Moreover, LSTM units are also employed to acquire temporal features. The proposed model demonstrates a significant enhancement in RMSE compared to the state-of-the-art, achieving a notable improvement of 44.96% for the steering angle and 4.39% for the speed on the Udacity dataset. Furthermore, tests on the CARLA and Sully Chen datasets yield results that outperform those reported in the literature. Extensive ablation studies are also conducted to showcase the effectiveness of each component. These findings highlight the potential of self-driving systems using visual input alone.
Subject Keywords
Autonomous driving
,
end-to-end learning
,
multi-task learning
,
self-driving car
,
steering estimation
URI
https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=105021248487&origin=inward
https://hdl.handle.net/11511/117254
Journal
IEEE Transactions on Intelligent Vehicles
DOI
https://doi.org/10.1109/tiv.2024.3483114
Collections
Department of Civil Engineering, Article
Citation Formats
IEEE
ACM
APA
CHICAGO
MLA
BibTeX
S. Azak, F. Bozkaya, Ş. Tığlıoğlu, A. Yusefi, and A. Durdu, “A Unified Monocular Vision-Based Driving Model for Autonomous Vehicles With Multi-Task Capabilities,”
IEEE Transactions on Intelligent Vehicles
, vol. 10, no. 9, pp. 4397–4408, 2025, Accessed: 00, 2025. [Online]. Available: https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=105021248487&origin=inward.