Abstract
Neural machine translation (NMT) has advanced with deep learning and large-scale multilingual models, yet translating low-resource languages often lacks sufficient training data and leads to hallucinations. This often results in translated content that diverges significantly from the source text. This research proposes a refined Contrastive Decoding (CD) algorithm that dynamically adjusts weights of log probabilities from strong expert and weak amateur models to mitigate hallucinations in low-resource NMT and improve translation quality. Advanced large language NMT models, including ChatGLM and LLaMA, are fine-tuned and implemented for their superior contextual understanding and cross-lingual capabilities. The refined CD algorithm evaluates multiple candidate translations using BLEU score, semantic similarity, and Named Entity Recognition accuracy. Extensive experimental results show substantial improvements in translation quality and a significant reduction in hallucination rates. Fine-tuned models achieve higher evaluation metrics compared to baseline models and state-of-the-art models. An ablation study confirms the contributions of each methodological component and highlights the effectiveness of the refined CD algorithm and advanced models in mitigating hallucinations. Notably, the refined methodology increased the BLEU score by approximately 30% compared to baseline models.
| Original language | English |
|---|---|
| Pages (from-to) | 1104-1117 |
| Number of pages | 14 |
| Journal | CAAI Transactions on Intelligence Technology |
| Volume | 10 |
| Issue number | 4 |
| DOIs | |
| Publication status | Published - Aug 2025 |
| MoE publication type | A1 Journal article-refereed |
Keywords
- artificial intelligence
- artificial neural network
- computer vision
- deep learning
- deep neural networks
- large language model
Fingerprint
Dive into the research topics of 'Large Language Models With Contrastive Decoding Algorithm for Hallucination Mitigation in Low-Resource Languages'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver