Show/Hide Menu
Hide/Show Apps
Logout
Türkçe
Türkçe
Search
Search
Login
Login
OpenMETU
OpenMETU
About
About
Open Science Policy
Open Science Policy
Open Access Guideline
Open Access Guideline
Postgraduate Thesis Guideline
Postgraduate Thesis Guideline
Communities & Collections
Communities & Collections
Help
Help
Frequently Asked Questions
Frequently Asked Questions
Guides
Guides
Thesis submission
Thesis submission
MS without thesis term project submission
MS without thesis term project submission
Publication submission with DOI
Publication submission with DOI
Publication submission
Publication submission
Supporting Information
Supporting Information
General Information
General Information
Copyright, Embargo and License
Copyright, Embargo and License
Contact us
Contact us
Assessment of AI-generated front-end code quality: a comparative study
Download
thesis.pdf
Umut Güler_Yayımlama Fikri Mülkiyet Hakları ve Doğruluk Beyanı Jüri İmza Sayfası ve Öğrenci İmza Sayfası.pdf
Date
2026-1-20
Author
Güler, Umut
Metadata
Show full item record
This work is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License
.
Item Usage Stats
75
views
0
downloads
Cite This
The adoption of AI coding tools has significantly transformed software development practices, yet empirical evidence regarding the technical quality of AI-generated code and applications remains limited, particularly for modern front-end development. This thesis presents a comprehensive comparative evaluation of three commercially available AI coding tools (Cursor, Gemini CLI, and Windsurf), assessing their impact on the technical quality of front-end web applications developed using React, TypeScript, and Vite. A controlled experiment is designed for the implementations. Two application scenarios are used: a simple To-Do application, and a relatively complex Marketplace application. Each application for each AI tool is developed using two methodologies: single prompt generation and incremental build with iterative prompts. The technical qualities of these codebases and applications are evaluated with three industry-standard tools: Lighthouse for runtime performance, accessibility, and web standards; SonarQube for maintainability and code quality; and ESLint for code style compliance. Bundle sizes and interactive correction requirements were also measured as complementary assets. The results demonstrate that AI tools successfully generate functionally correct applications. However, quality outcomes vary significantly based on the AI tools, development methodology, and application complexity. It is observed that incremental development benefits the AI-generated applications consistently, regardless of the AI tool, in simple application scenarios. Conversely, complex applications exhibited significant quality degradations with incremental development, such as complexity increases and severe reliability drops for certain tool-methodology combinations. The findings reveal that each tool demonstrates a distinct performance profile in specific scenarios and metrics; no single tool dominates across all quality dimensions. This research provides evidence-based guidance for tool selection and development methodology based on application complexity requirements.
Subject Keywords
Generative AI
,
Front-end development
,
Code quality
,
AI coding tools
,
Software engineering
URI
https://hdl.handle.net/11511/118453
Collections
Graduate School of Informatics, Thesis
Citation Formats
IEEE
ACM
APA
CHICAGO
MLA
BibTeX
U. Güler, “Assessment of AI-generated front-end code quality: a comparative study,” M.S. - Master of Science, Middle East Technical University, 2026.