TY - GEN
T1 - A Novel Translation-Driven Approach to Enhance LLM Performance on Low-Resource Languages
AU - Ofer, Moshe
AU - Zamler, Orel
AU - Azaria, Amos
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Large Language Models (LLMs) excel in highresource languages but struggle with low-resource languages due to limited training data and insufficient representation during pre-training. This disparity creates significant barriers for deploying advanced NLP technologies across diverse linguistic communities. This paper presents TALL (Trainable Architecture for Enhancing LLM Performance in Low-Resource Languages), a novel framework that strategically integrates an LLM with two bilingual translation models to bridge the performance gap between high and low-resource languages. TALL transforms lowresource inputs into high-resource representations through a multi-stage pipeline, leveraging the LLM's robust capabilities while preserving essential linguistic features through carefully designed dimension alignment layers and custom transformer components. The architecture addresses the challenge of integrating models with different hidden dimensions and representation spaces, enabling seamless knowledge transfer across languages. Our comprehensive experiments on Hebrew demonstrate significant improvements over several competitive baselines, including direct LLM use, naive translation approaches, finetuning strategies, and soft prompting techniques. Notably, TALL achieves up to 5. 5 9% accuracy compared to 2. 9 3% for the next best approach, representing a substantial performance gain. The architecture employs a parameter-efficient strategy, freezing large pre-trained components while training only lightweight adapter modules, effectively balancing computational efficiency with performance gains. This approach makes TALL particularly suitable for resource-constrained environments while maintaining strong cross-lingual transfer capabilities. Code is available in https://github.com/MosheOfer1/TALL
AB - Large Language Models (LLMs) excel in highresource languages but struggle with low-resource languages due to limited training data and insufficient representation during pre-training. This disparity creates significant barriers for deploying advanced NLP technologies across diverse linguistic communities. This paper presents TALL (Trainable Architecture for Enhancing LLM Performance in Low-Resource Languages), a novel framework that strategically integrates an LLM with two bilingual translation models to bridge the performance gap between high and low-resource languages. TALL transforms lowresource inputs into high-resource representations through a multi-stage pipeline, leveraging the LLM's robust capabilities while preserving essential linguistic features through carefully designed dimension alignment layers and custom transformer components. The architecture addresses the challenge of integrating models with different hidden dimensions and representation spaces, enabling seamless knowledge transfer across languages. Our comprehensive experiments on Hebrew demonstrate significant improvements over several competitive baselines, including direct LLM use, naive translation approaches, finetuning strategies, and soft prompting techniques. Notably, TALL achieves up to 5. 5 9% accuracy compared to 2. 9 3% for the next best approach, representing a substantial performance gain. The architecture employs a parameter-efficient strategy, freezing large pre-trained components while training only lightweight adapter modules, effectively balancing computational efficiency with performance gains. This approach makes TALL particularly suitable for resource-constrained environments while maintaining strong cross-lingual transfer capabilities. Code is available in https://github.com/MosheOfer1/TALL
KW - cross-lingual transfer
KW - Hebrew NLP
KW - large language models
KW - low-resource languages
KW - parameter-efficient adaptation
UR - https://www.scopus.com/pages/publications/105031913129
U2 - 10.1109/ICTAI66417.2025.00052
DO - 10.1109/ICTAI66417.2025.00052
M3 - ???researchoutput.researchoutputtypes.contributiontobookanthology.conference???
AN - SCOPUS:105031913129
T3 - Proceedings - International Conference on Tools with Artificial Intelligence, ICTAI
SP - 347
EP - 354
BT - Proceedings - 2025 IEEE 37th International Conference on Tools with Artificial Intelligence, ICTAI 2025
PB - IEEE Computer Society
T2 - 37th IEEE International Conference on Tools with Artificial Intelligence, ICTAI 2025
Y2 - 3 November 2025 through 5 November 2025
ER -