期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Building accurate translation-tailored large language models with language-aware instruction tuning
1
作者 Changtong ZAN Liang DING +3 位作者 Li SHEN Yibing ZHAN Xinghao YANG Weifeng LIU 《Frontiers of Information Technology & Electronic Engineering》 2025年第8期1341-1355,共15页
Large language models(LLMs)exhibit remarkable capabilities in various natural language processing tasks,such as machine translation.However,the large number of LLM parameters incurs significant costs during inference.... Large language models(LLMs)exhibit remarkable capabilities in various natural language processing tasks,such as machine translation.However,the large number of LLM parameters incurs significant costs during inference.Previous studies have attempted to train translation-tailored LLMs with moderately sized models by fine-tuning them on the translation data.Nevertheless,when performing translations in zero-shot directions that are absent from the fine-tuning data,the problem of ignoring instructions and thus producing translations in the wrong language(i.e.,the off-target translation issue)remains unresolved.In this work,we design a twostage fine-tuning algorithm to improve the instruction-following ability of translation-tailored LLMs,particularly for maintaining accurate translation directions.We first fine-tune LLMs on the translation data to elicit basic translation capabilities.At the second stage,we construct instruction-conficting samples by randomly replacing the instructions with the incorrect ones.Then,we introduce an extra unlikelihood loss to reduce the probability assigned to those samples.Experiments on two benchmarks using the LLaMA 2 and LLaMA 3 models,spanning 16 zero-shot directions,demonstrate that,compared to the competitive baseline translation-finetuned LLaMA,our method could effectively reduce the off-target translation ratio(up to-62.4 percentage points),thus improving translation quality(up to+9.7 bilingual evaluation understudy).Analysis shows that our method can preserve the model's performance on other tasks,such as supervised translation and general tasks.Code is released at https://github.com/alphadl/LanguageAware_Tuning. 展开更多
关键词 Zero-shot machine translation Off-target issue Large language model language-aware instruction tuning Instruction-conflicting sample
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部