Translation with LLMs through Prompting with Long-Form Context

dc.contributor.authorAshqar, Huthaifa
dc.contributor.authorTami, Mohammad
dc.date.accessioned2025-10-16T15:27:08Z
dc.date.issued2025-08-02
dc.descriptionACL 2023, 61st Annual Meeting of the Association for Computational Linguistics, July 9th - 14th, 2023, Toronto, Canada
dc.description.abstractStable generation of text in low-resource languages is an unsolved issue in large language models. While Large Language Models (LLMs) can often produce good translations despite not being explicitly trained for this task, this does not hold for low-resource languages. LLMs are both more likely to generate off-target text (text in another language than intended) when prompted to translate to a low-resource language, and show increased instability in translation quality across prompt templates in low-resource languages. This study implemented a prepended monolingual text prompting method in the target language and used a context-and topic-aware with few-shot machine translation (MT). We quantified these methods for low-, mid-, and high-resource languages using OpenAI GPT-4o-mini and Google Gemini-1.5-flash. Gemini results showed that the use of contexttopic-aware with few-shot MT (CTAFSMT) significantly boosted the performance for the three language categories. However, this was not consistently observed in the case of ChatGPT. It was found that the significance of the results depended on the language itself rather than the level of its resources. This study is part of Stanford University's meta-study on whether LLMs can generate novel research ideas. The code, prompts, and results of the study can be found at https://github.com/HuthaifaAshqar/Translationwith-LLMs.
dc.description.sponsorshipWe would like to acknowledge that the original idea was submitted by Elizabeth Salesky from Google DeepMind for the purpose of this study, which is part of the funded Stanford University’s meta-study on whether LLMs can generate novel research ideas.
dc.description.urihttps://www.authorea.com/users/884736/articles/1318311-translation-with-llms-through-prompting-with-long-form-context
dc.format.extent7 pages
dc.genreconference papers and proceedings
dc.genrepreprints
dc.identifierdoi:10.13016/m2iz7n-bfyh
dc.identifier.urihttps://doi.org/10.36227/techrxiv.175416071.14800516/v1
dc.identifier.urihttp://hdl.handle.net/11603/40436
dc.language.isoen
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Data Science
dc.rightsAttribution 4.0 International
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.titleTranslation with LLMs through Prompting with Long-Form Context
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0002-6835-8338

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ACL_2023_Proceedings_LLMsTranslation_Preprint.pdf
Size:
176.47 KB
Format:
Adobe Portable Document Format