Fine-tuning Large Language Models for Turkish Flutter Code Generation


Creative Commons License

Uluırmak B. A., Kurban R.

Sakarya University Journal of Computer and Information Sciences, vol.8, no.4, pp.637-650, 2025 (Scopus)

  • Publication Type: Article / Article
  • Volume: 8 Issue: 4
  • Publication Date: 2025
  • Doi Number: 10.35377/saucis...1722643
  • Journal Name: Sakarya University Journal of Computer and Information Sciences
  • Journal Indexes: Scopus, TR DİZİN (ULAKBİM)
  • Page Numbers: pp.637-650
  • Open Archive Collection: AVESIS Open Access Collection
  • Abdullah Gül University Affiliated: Yes

Abstract

The rapid advancement of large language models (LLMs) for code generation has largely centered on English programming queries. This paper targets a low-resource language scenario, Turkish, in Flutter mobile app development. Two representative LLMs (a 4B-parameter multilingual model and a 3B code-specialized model) on a new Turkish question-and-answer dataset for Flutter/Dart are fine-tuned in this study. Fine-tuning with parameter-efficient techniques yields dramatic improvements in code generation quality: Bilingual Evaluation Understudy (BLEU), Recall-Oriented Understudy for Gisting Evaluation (ROUGE-L), Metric for Evaluation of Translation with Explicit Ordering (METEOR), Bidirectional Encoder Representations from Transformers Score (BERTScore), and CodeBLEU scores show significant increases. The rate of correct solutions increased from ~30–70% (for base models) to 80–90% after fine-tuning. The performance trade-offs between models are analyzed, revealing that the multilingual model slightly outperforms the code-focused model in accuracy after fine-tuning. However, the code-focused model demonstrates faster inference speeds. These results demonstrate that even with very limited non-English training data, customizing LLMs can bridge the gap in code generation, enabling high-quality assistance for Turkish developers comparable to that for English. The dataset was released on GitHub to facilitate further research in multilingual code generation.