Show/Hide Menu
Hide/Show Apps
Logout
Türkçe
Türkçe
Search
Search
Login
Login
OpenMETU
OpenMETU
About
About
Open Science Policy
Open Science Policy
Open Access Guideline
Open Access Guideline
Postgraduate Thesis Guideline
Postgraduate Thesis Guideline
Communities & Collections
Communities & Collections
Help
Help
Frequently Asked Questions
Frequently Asked Questions
Guides
Guides
Thesis submission
Thesis submission
MS without thesis term project submission
MS without thesis term project submission
Publication submission with DOI
Publication submission with DOI
Publication submission
Publication submission
Supporting Information
Supporting Information
General Information
General Information
Copyright, Embargo and License
Copyright, Embargo and License
Contact us
Contact us
Multilingual Domain Adaptation for Speech Recognition Using LLMs
Date
2026-01-01
Author
Ulu, Elif Nehir
Derya, Ece
Tumer, Duygu
Demirel, Berkan
Karamanlıoğlu, Alper
Metadata
Show full item record
This work is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License
.
Item Usage Stats
530
views
0
downloads
Cite This
We present a practical pipeline for multilingual domain adaptation in automatic speech recognition (ASR) that combines the Whisper model with large language models (LLMs). Using Aya-23-8B, Common Voice transcripts in 22 languages are automatically classified into the Law and Healthcare domains, producing high-quality domain labels at a fraction of the manual cost. These labels drive parameterefficient (LoRA) fine-tuning of Whisper and deliver consistent relative Word Error Rate (WER) reductions of up to 14.3% for languages that contribute at least 800 in-domain utterances. A data-volume analysis reveals a clear breakpoint: gains become reliably large once that 800-utterance threshold is crossed, while monolingual tuning still rescues performance in truly low-resource settings. The workflow therefore shifts the key success factor from expensive hand labelling to scalable data acquisition, and can be replicated in new domains with minimal human intervention.
URI
https://hdl.handle.net/11511/117816
DOI
https://doi.org/10.1007/978-3-032-02548-7_32
Conference Name
28th International Conference on Text Speech and Dialogue-TSD-Annual
Collections
Department of Computer Engineering, Conference / Seminar
Citation Formats
IEEE
ACM
APA
CHICAGO
MLA
BibTeX
E. N. Ulu, E. Derya, D. Tumer, B. Demirel, and A. Karamanlıoğlu, “Multilingual Domain Adaptation for Speech Recognition Using LLMs,” Erlangen, Almanya, 2026, vol. 16029, Accessed: 00, 2025. [Online]. Available: https://hdl.handle.net/11511/117816.