🤖 AI Summary
This work addresses named entity recognition (NER) for the low-resource language Nepali. We conduct the first systematic evaluation of mainstream large language models (LLMs) on generative zero-shot and few-shot NER without fine-tuning. To enable rigorous assessment, we construct a manually annotated Nepali NER test set and empirically evaluate both open- and closed-source LLMs using zero-shot, few-shot, and chain-of-thought prompting strategies. Results demonstrate that LLMs possess foundational NER capability, achieving up to 72.4% accuracy; moreover, our lightweight prompt engineering approach significantly enhances adaptability to low-resource settings. This study fills a critical gap by establishing the first publicly available benchmark for Nepali NER, providing a reproducible methodological framework and the inaugural performance benchmark for generative NER in low-resource languages.
📝 Abstract
Generative Artificial Intelligence (GenAI), particularly Large Language Models (LLMs), has significantly advanced Natural Language Processing (NLP) tasks, such as Named Entity Recognition (NER), which involves identifying entities like person, location, and organization names in text. LLMs are especially promising for low-resource languages due to their ability to learn from limited data. However, the performance of GenAI models for Nepali, a low-resource language, has not been thoroughly evaluated. This paper investigates the application of state-of-the-art LLMs for Nepali NER, conducting experiments with various prompting techniques to assess their effectiveness. Our results provide insights into the challenges and opportunities of using LLMs for NER in low-resource settings and offer valuable contributions to the advancement of NLP research in languages like Nepali.