Show/Hide Menu
Hide/Show Apps
Logout
Türkçe
Türkçe
Search
Search
Login
Login
OpenMETU
OpenMETU
About
About
Open Science Policy
Open Science Policy
Open Access Guideline
Open Access Guideline
Postgraduate Thesis Guideline
Postgraduate Thesis Guideline
Communities & Collections
Communities & Collections
Help
Help
Frequently Asked Questions
Frequently Asked Questions
Guides
Guides
Thesis submission
Thesis submission
MS without thesis term project submission
MS without thesis term project submission
Publication submission with DOI
Publication submission with DOI
Publication submission
Publication submission
Supporting Information
Supporting Information
General Information
General Information
Copyright, Embargo and License
Copyright, Embargo and License
Contact us
Contact us
XoFTR: Cross-modal Feature Matching Transformer
Date
2024-06-16
Author
Tuzcuoğlu, Önder
Köksal, Aybora
Sofu, Buğra
Kalkan, Sinan
Alatan, Abdullah Aydın
Metadata
Show full item record
This work is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License
.
Item Usage Stats
19
views
0
downloads
Cite This
We introduce XoFTR a cross-modal cross-view method for local feature matching between thermal infrared (TIR) and visible images. Unlike visible images TIR images are less susceptible to adverse lighting and weather conditions but present difficulties in matching due to significant texture and intensity differences. Current hand-crafted and learning-based methods for visible-TIR matching fall short in handling viewpoint scale and texture diversities. To address this XoFTR incorporates masked image modeling pre-training and fine-tuning with pseudo-thermal image augmentation to handle the modality differences. Additionally we introduce a refined matching pipeline that adjusts for scale discrepancies and enhances match reliability through sub-pixel level refinement. To validate our approach we collect a comprehensive visible-thermal dataset and show that our method outperforms existing methods on many benchmarks. Code and dataset at https://github.com/OnderT/XoFTR.
URI
https://openaccess.thecvf.com/content/CVPR2024W/IMW/html/Tuzcuoglu_XoFTR_Cross-modal_Feature_Matching_Transformer_CVPRW_2024_paper.html
https://hdl.handle.net/11511/111742
Conference Name
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition
Collections
Department of Electrical and Electronics Engineering, Conference / Seminar
Citation Formats
IEEE
ACM
APA
CHICAGO
MLA
BibTeX
Ö. Tuzcuoğlu, A. Köksal, B. Sofu, S. Kalkan, and A. A. Alatan, “XoFTR: Cross-modal Feature Matching Transformer,” presented at the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Washington, Amerika Birleşik Devletleri, 2024, Accessed: 00, 2024. [Online]. Available: https://openaccess.thecvf.com/content/CVPR2024W/IMW/html/Tuzcuoglu_XoFTR_Cross-modal_Feature_Matching_Transformer_CVPRW_2024_paper.html.