A Rising Tide Lifts All Boats: MTQE Rewards for Idioms Improve General Translation Quality
Abstract
GRPO-style fine-tuning with MTQE models as rewards improves idiom translation by 14 points while enhancing general translation and cross-lingual capabilities.
Non-compositional expressions (e.g., idioms, proverbs, and metaphors) pose significant challenges for neural machine translation systems because their meanings cannot be derived from individual words alone. These expressions encode rich, cultural meaning, and have both figurative and literal meanings, making accurate translation difficult. Because models are fairly good at translating compositional text, we investigate GRPO-style fine-tuning using Machine Translation Quality Estimation (MTQE) models as reward functions to train models to better translate idioms. Using Chinese and Hindi idiom datasets, we find that idiom translation abilities improve by ~14 points, general, non-idiomatic translation implicitly improves by ~8 points, and cross-lingual translation abilities (trained on one language, evaluated on another) improves by ~6 points. Overall, our work quantifies the non-compositional translation gap and offers insights for developing LLMs with stronger cross-cultural and figurative language understanding.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- LLMs for Low-Resource Dialect Translation Using Context-Aware Prompting: A Case Study on Sylheti (2025)
- Structured Document Translation via Format Reinforcement Learning (2025)
- NLP Datasets for Idiom and Figurative Language Tasks (2025)
- From Scratch to Fine-Tuned: A Comparative Study of Transformer Training Strategies for Legal Machine Translation (2025)
- Remedy-R: Generative Reasoning for Machine Translation Evaluation without Error Annotations (2025)
- Comparing Approaches to Automatic Summarization in Less-Resourced Languages (2025)
- Can QE-informed (Re)Translation lead to Error Correction? (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 16
Browse 16 models citing this paperDatasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper