Show/Hide Menu
Hide/Show Apps
Logout
Türkçe
Türkçe
Search
Search
Login
Login
OpenMETU
OpenMETU
About
About
Open Science Policy
Open Science Policy
Open Access Guideline
Open Access Guideline
Postgraduate Thesis Guideline
Postgraduate Thesis Guideline
Communities & Collections
Communities & Collections
Help
Help
Frequently Asked Questions
Frequently Asked Questions
Guides
Guides
Thesis submission
Thesis submission
MS without thesis term project submission
MS without thesis term project submission
Publication submission with DOI
Publication submission with DOI
Publication submission
Publication submission
Supporting Information
Supporting Information
General Information
General Information
Copyright, Embargo and License
Copyright, Embargo and License
Contact us
Contact us
Object height estimation on monoscopic satellite images using deep learning
Download
Furkan_Gultekin_ms_thesis.pdf
Date
2024-1-22
Author
Gültekin, Furkan
Metadata
Show full item record
This work is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License
.
Item Usage Stats
371
views
192
downloads
Cite This
Conventional methodologies for deriving 3D information from satellite optical images often rely on intricate algorithms requiring substantial human supervision, numerous external parameters, and the acquisition of multiple images captured from different perspectives. This study aims to eliminate the need for manual parameter tuning and multiple image acquisition and completely automate the extraction of 3D object information from monoscopic satellite imagery by leveraging advanced deep learning algorithms. The Fused Height Estimation (Fused-HE) deep network model, which leverages the inherent characteristics of satellite imagery, is proposed within this scope. The model, featuring dual encoders, processes monoscopic satellite image inputs individually through the convolutional encoder for local feature extraction within neighboring pixels and the vision transformers encoder for global feature extraction using the relationships between objects in the image. The distinct feature outputs from the two encoders are concatenated in feature fusion blocks. Then, fused features are passed to the decoder blocks, and the height head produces the predicted heights. The proposed fused network is further improved by introducing an additional segmentation head to the model to assign height values to the correct pixel, and this model is named the Fused Segmentation Height Estimation (FusedSeg-HE). Comprehensive evaluations of individual models in the literature and the proposed fused networks demonstrate the proposed models provide both local and global feature extraction for height estimation, reduce root mean squared error between the estimated and ground truth height values by approximately 5%-13%, and increase accuracy by 4%-9% in terms of the delta threshold metric.
Subject Keywords
Deep learning
,
Monoscopic satellite image
,
Object height estimation
,
Local-global feature extraction
,
Convolutional neural networks
,
Vision transformers
URI
https://hdl.handle.net/11511/108758
Collections
Graduate School of Natural and Applied Sciences, Thesis
Citation Formats
IEEE
ACM
APA
CHICAGO
MLA
BibTeX
F. Gültekin, “Object height estimation on monoscopic satellite images using deep learning,” M.S. - Master of Science, Middle East Technical University, 2024.