There are 5 repositories under pretrained-language-model topic.
YAYI 2 是中科闻歌研发的新一代开源大语言模型,采用了超过 2 万亿 Tokens 的高质量、多语言语料进行预训练。(Repo for YaYi 2 Chinese LLMs)
Foundation Architecture for (M)LLMs
A curated list of pretrained sentence and word embedding models
An optimized deep prompt tuning strategy comparable to fine-tuning across scales and tasks
Summarization Papers
中文法律LLaMA (LLaMA for Chinese legel domain)
Code associated with the Don't Stop Pretraining ACL 2020 paper
word2vec, sentence2vec, machine reading comprehension, dialog system, text classification, pretrained language model (i.e., XLNet, BERT, ELMo, GPT), sequence labeling, information retrieval, information extraction (i.e., entity, relation and event extraction), knowledge graph, text generation, network embedding
Papers and Datasets on Instruction Tuning and Following. ✨✨✨
ACL'2023: DiffusionBERT: Improving Generative Masked Language Models with Diffusion Models
MWPToolkit is an open-source framework for math word problem(MWP) solvers.
Worth-reading papers and related resources on attention mechanism, Transformer and pretrained language model (PLM) such as BERT. 值得一读的注意力机制、Transformer和预训练语言模型论文与相关资源集合
[NeurIPS 2023] This is the code for the paper `Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias`.
[NeurIPS 2021] COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining
EMNLP'23 survey: a curation of awesome papers and resources on refreshing large language models (LLMs) without expensive retraining.
On Transferability of Prompt Tuning for Natural Language Processing
Bamboo-7B Large Language Model
[WWW 2022] Topic Discovery via Latent Space Clustering of Pretrained Language Model Representations
[KDD22] Official PyTorch implementation for "Towards Unified Conversational Recommender Systems via Knowledge-Enhanced Prompt Learning".
CODER: Knowledge infused cross-lingual medical term embedding for term normalization. [JBI, ACL-BioNLP 2022]
[NeurIPS 2022] Generating Training Data with Language Models: Towards Zero-Shot Language Understanding
Implementation of ICLR 21 paper: Probing BERT in Hyperbolic Spaces
ELECTRA기반 한국어 대화체 언어모델
Official implementation of the ACL 2024: Scientific Inspiration Machines Optimized for Novelty
BioBART: Pretraining and Evaluation of A Biomedical Generative Language Model [ACL-BioNLP 2022]
Implementation of "TransPolymer: a Transformer-based language model for polymer property predictions" in PyTorch
[EMNLP 2022] This is the code repo for our EMNLP‘22 paper "COCO-DR: Combating Distribution Shifts in Zero-Shot Dense Retrieval with Contrastive and Distributionally Robust Learning".
Translate Natural Language Processing to SPARQL Query and vice versa
Awesome LLM Self-Consistency: a curated list of Self-consistency in Large Language Models
[Paper][AAAI 2023] DUET: Cross-modal Semantic Grounding for Contrastive Zero-shot Learning
EMNLP'2022: BERTScore is Unfair: On Social Bias in Language Model-Based Metrics for Text Generation
The source code used for paper "Empower Entity Set Expansion via Language Model Probing", published in ACL 2020.
This repository contains the code for the paper in Findings of EMNLP 2021: "EfficientBERT: Progressively Searching Multilayer Perceptron via Warm-up Knowledge Distillation".