Contextual Learning Support for Low-Resource Language Large Language Models: Efficient Training Strategies for Zero-Shot and Few-Shot Learning

Authors

  • Tanya Xiang School of Physics and Engineering, University of Birmingham, Birmingham, United Kingdom Author

DOI:

https://doi.org/10.71222/gkj3aw70

Keywords:

low-resource languages, large language models, contextual learning, transfer learning, zero-shot learning, few-shot learning

Abstract

In natural language processing tasks for low-resource languages, building high-quality large language models faces numerous challenges, including data scarcity, domain mismatch, and limited annotation resources. This paper proposes an efficient training strategy based on contextual learning, aimed at addressing the difficulties encountered by low-resource languages in zero-shot and few-shot learning scenarios. The approach leverages cross-lingual knowledge transfer from high-resource languages and systematically enhances contextual information in prompts and representations to maximize the model’s generalization ability. We explore the combination of multi-task learning and self-supervised learning to exploit heterogeneous corpora, using existing multilingual and monolingual resources for pretraining. A lightweight fine-tuning stage with a small amount of labeled data is then employed for targeted adaptation to specific downstream tasks and languages. The proposed framework is designed to be computationally efficient, reducing training cost while maintaining or improving performance. Experimental results on a range of language understanding and generation benchmarks demonstrate significant improvements in task performance across various low-resource languages under both zero-shot and few-shot conditions. Ablation studies further highlight the contribution of contextual learning components and cross-lingual transfer mechanisms. These findings provide practical guidance for developing scalable large language models for underrepresented languages and offer new ideas and methods for future research on inclusive and resource-efficient language technologies.

References

1. Z. X. Yong, C. Menghini, and S. Bach, "Lexc-gen: Generating data for extremely low-resource languages with large language models and bilingual lexicons," in Findings of the Association for Computational Linguistics: EMNLP 2024, Nov. 2024, pp. 13990–14009.

2. M. A. Hasan, P. Tarannum, K. Dey, I. Razzak, and U. Naseem, "Do large language models speak all languages equally? A comparative study in low-resource settings," arXiv preprint arXiv:2408.02237, 2024.

3. M. Adeyemi, A. Oladipo, R. Pradeep, and J. Lin, "Zero-shot cross-lingual reranking with large language models for low-resource languages," in Proc. 62nd Annu. Meeting Assoc. Comput. Linguistics (Vol. 2: Short Papers), Aug. 2024, pp. 650–656.

4. P. Guo, Y. Ren, Y. Hu, Y. Li, J. Zhang, X. Zhang, and H. Y. Huang, "Teaching large language models to translate on low-resource languages with textbook prompting," in Proc. 2024 Joint Int. Conf. Comput. Linguistics, Lang. Resources and Evaluation (LREC-COLING 2024), May 2024, pp. 15685–15697.

5. N. Kholodna, S. Julka, M. Khodadadi, M. N. Gumus, and M. Granitzer, "LLMs in the loop: Leveraging large language model annotations for active learning in low-resource languages," in Joint Eur. Conf. Mach. Learn. Knowl. Discovery Databases, Cham: Springer Nature Switzerland, Aug. 2024, pp. 397–412.

6. T. Zhong, Z. Yang, Z. Liu, R. Zhang, W. You, Y. Liu, ... and T. Liu, "Opportunities and challenges of large language models for low-resource languages in humanities research," arXiv preprint arXiv:2412.04497, 2024.

7. G. Mani and G. B. Namomsa, "Large language models (LLMs): Representation matters, low-resource languages and multi-modal architecture," in 2023 IEEE AFRICON, Sep. 2023, pp. 1–6.

8. V. Hangya, H. S. Saadi, and A. Fraser, "Improving low-resource languages in pre-trained multilingual language models," in Proc. 2022 Conf. Empirical Methods Natural Lang. Process., Dec. 2022, pp. 11993–12006.

9. N. Kadyrbek, Z. Tuimebayev, M. Mansurova, and V. Viegas, "The development of small-scale language models for low-resource languages, with a focus on Kazakh and direct preference optimization," Big Data Cogn. Comput., vol. 9, no. 5, p. 137, 2025.

10. J. McGiff and N. S. Nikolov, "Overcoming data scarcity in generative language modelling for low-resource languages: A systematic review," arXiv preprint arXiv:2505.04531, 2025.

11. N. Bui, G. Nguyen, N. Nguyen, B. Vo, L. Vo, T. Huynh, ... and M. Dinh, "Fine-tuning large language models for improved health communication in low-resource languages," Comput. Methods Programs Biomed., vol. 263, p. 108655, 2025.

12. H. Avetisyan and D. Broneske, "Large language models and low-resource languages: An examination of Armenian NLP," Findings Assoc. Comput. Linguistics: IJCNLP-AACL 2023 (Findings), pp. 199–210, 2023.

13. T. A. Chang, C. Arnett, Z. Tu, and B. Bergen, "When is multilinguality a curse? Language modeling for 250 high-and low-resource languages," in Proc. 2024 Conf. Empirical Methods Natural Lang. Process., Nov. 2024, pp. 4074–4096.

14. P. W. Khoboko, V. Marivate, and J. Sefara, "Optimizing translation for low-resource languages: Efficient fine-tuning with custom prompt engineering in large language models," Mach. Learn. Appl., vol. 20, p. 100649, 2025.

15. T. Harrington, C. Navarro, E. Davidson, C. Salazar, and C. Langston, "Recursive instruction tuning of large language models for low-resource languages," Authorea Preprints, 2024.

16. X. Liang, Y. M. J. Khaw, S. Y. Liew, T. P. Tan, and D. Qin, "Towards low-resource languages machine translation: A language-specific fine-tuning with LoRA for specialized large language models," IEEE Access, 2025.

17. K. Ogueji, Y. Zhu, and J. Lin, "Small data? No problem! Exploring the viability of pretrained multilingual language models for low-resourced languages," in Proc. 1st Workshop Multilingual Representation Learn., Nov. 2021, pp. 116–126.

18. S. Joshi, M. S. Khan, A. Dafe, K. Singh, V. Zope, and T. Jhamtani, "Fine tuning LLMs for low resource languages," in 2024 5th Int. Conf. Image Process. Capsule Netw. (ICIPCN), Jul. 2024, pp. 511–519.

Downloads

Published

11 April 2026

How to Cite

Xiang, T. (2026). Contextual Learning Support for Low-Resource Language Large Language Models: Efficient Training Strategies for Zero-Shot and Few-Shot Learning. Pinnacle Academic Press Proceedings Series, 10, 344-352. https://doi.org/10.71222/gkj3aw70