sunnyhuma171 / GLiNER

Generalist model for NER

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

GLiNER: Generalist Model for Named Entity Recognition using Bidirectional Transformer

GLiNER Logo

Overview

Named Entity Recognition (NER) is essential in various Natural Language Processing (NLP) applications. Traditional NER models are effective but limited to a set of predefined entity types. In contrast, Large Language Models (LLMs) can extract arbitrary entities through natural language instructions, offering greater flexibility. However, their size and cost, particularly for those accessed via APIs like ChatGPT, make them impractical in resource-limited scenarios. In this paper, we introduce a compact NER model trained to identify any type of entity. Leveraging a bidirectional transformer encoder, our model, GLiNER, facilitates parallel entity extraction, an advantage over the slow sequential token generation of LLMs. Through comprehensive testing, GLiNER demonstrate strong performance, outperforming both ChatGPT and fine-tuned LLMs in zero-shot evaluations on various NER benchmarks.

Pretrained Weight

  • Download the pre-trained weights for GLiNER using the following link: Pretrained Weight

Training Data

  • Find the training data used for training GLiNER here: Training Data

Evaluation Data

  • Evaluate the performance of GLiNER using the provided evaluation data (provided by instruction_IE): Evaluation Data

Paper

  • Explore the details of the GLiNER model and its performance in the accompanying paper: GLiNER Paper

Contact

If you have any questions or need further assistance, please feel free to email us at urchade.zaratiana@gmail.com.

We hope GLiNER proves to be a valuable resource for your Named Entity Recognition tasks. Thank you for your interest in our project!

@misc{zaratiana2023gliner,
      title={GLiNER: Generalist Model for Named Entity Recognition using Bidirectional Transformer}, 
      author={Urchade Zaratiana and Nadi Tomeh and Pierre Holat and Thierry Charnois},
      year={2023},
      eprint={2311.08526},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

About

Generalist model for NER


Languages

Language:Python 92.8%Language:Jupyter Notebook 7.2%