Abstract
The adoption of generative Artificial Intelligence (AI) tools in web development implementation tasks is increasing exponentially. This paper evaluates the performance of five leading Generative AI models: ChatGPT-4.0, DeepSeek-V3, Gemini-1.5, Copilot (March 2025 release), and Claude-3, in building HTML components. This study presents a structured evaluation of AI-generated HTML code produced by leading Generative AI models. We have designed a set of prompts for popular tasks to generate five standardized HTML components: a contact form, a navigation menu, a blog post layout, a product listing page, and a dashboard interface. The responses were evaluated across five dimensions: semantic structure, accessibility, efficiency, readability, and search engine optimization (SEO). Results show that while AI-generated HTML can achieve high validation scores, deficiencies remain in semantic structuring and accessibility, with measurable differences between models. The results show variation in the quality and structure of the generated HTML. These results provide practical insights into the limitations and strengths of the current use of AI tools in HTML development.
| Original language | English |
|---|---|
| Article number | 445 |
| Pages (from-to) | 445 |
| Number of pages | 1 |
| Journal | Technologies |
| Volume | 13 |
| Issue number | 10 |
| Early online date | 1 Oct 2025 |
| DOIs | |
| Publication status | Published - 1 Oct 2025 |
Keywords
- AI code generation
- frontend web components
- generative AI
- HTML development