A comprehensive overview of LLM-based approaches for machine translation
Abstract
Statistical machine translation (SMT) used parallel corpora and statistical models, to identify translation patterns and probabilities. Although this method had advantages, it had trouble with idiomatic expressions, context-specific subtleties, and intricate linguistic structures. The subsequent introduction of deep neural networks such as recurrent neural networks (RNNs), long short-term memory (LSTMs), transformers with attention mechanisms, and the emergence of large language model (LLM) frameworks has marked a paradigm shift in machine translation in recent years and has entirely replaced the traditional statistical approaches. The LLMs are able to capture complex language patterns, semantics, and context because they have been trained on enormous volumes of text data. Our study summarizes the most significant contributions in the literature related to LLM prompting, fine-tuning, retrieval augmented generation, improved transformer variants for faster translation, multilingual LLMs, and quality estimation with LLMs. This new research direction guides the development of more efficient and innovative solutions to address the current challenges of LLMs, including hallucinations, translation bias, information leakage, and inaccuracy due to language inconsistencies.
Keywords
Large language models; Low-resource languages; Machine translation; Prompting; Quality estimation; Tuning
Full Text:
PDFDOI: http://doi.org/10.11591/ijeecs.v38.i1.pp344-356
Refbacks
- There are currently no refbacks.
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Indonesian Journal of Electrical Engineering and Computer Science (IJEECS)
p-ISSN: 2502-4752, e-ISSN: 2502-4760
This journal is published by the Institute of Advanced Engineering and Science (IAES) in collaboration with Intelektual Pustaka Media Utama (IPMU).