Comparing Human and LLM Annotations in Low-Resource Language NLP Tasks

Authors

  • A.Hima Bindu Assistant Professor, Department Of Eee,, Bhoj Reddy Engineering College For Women, India. Author
  • Gundla Harshitha, Sangannagare Chandana B. Tech Students, Department Of Cse, Bhoj Reddy Engineering College For Women, India. Author

Abstract

In Natural Language Processing (NLP), annotated
datasets play a crucial role in training and
evaluating machine learning models. However, in
low-resource languages, the availability of highquality
annotated data is extremely limited due to
linguistic complexity, lack of standardization, and
scarcity of expert annotators. With the rise of Large
Language Models (LLMs), such as GPT and similar
models, there is growing interest in using these
models to generate annotations automatically. This
study compares human-generated annotations with
those generated by LLMs for NLP tasks such as partof-
speech tagging, named entity recognition, and
sentiment analysis in low-resource languages. The
comparison is based on precision, recall, and F1-
score, along with qualitative analysis. Our findings
show that while LLMs can provide reasonable
annotations in many cases, human annotations still
outperform them in linguistic nuance, context
understanding, and domain specificity. However,
LLMs show potential in speeding up the annotation
process and supporting human annotators through
pre-annotation. This research highlights the
complementary strengths of humans and LLMs and
proposes a hybrid annotation workflow for building
better NLP resources in low-resource settings.

Downloads

Published

2025-06-10

Issue

Section

Articles

How to Cite

Comparing Human and LLM Annotations in Low-Resource Language NLP Tasks. (2025). International Journal of Engineering and Science Research, 15(3s), 288-297. https://www.ijesr.org/index.php/ijesr/article/view/156

Most read articles by the same author(s)