There are 54 repositories under bert topic.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Natural Language Processing Tutorial for Deep Learning Researchers
Embed images and sentences into fixed-length vectors with CLIP
Pre-Training with Whole Word Masking for Chinese BERT（中文BERT-wwm系列模型）
大规模中文自然语言处理语料 Large Scale Chinese Corpus for NLP
💥 Fast State-of-the-Art Tokenizers optimized for Research and Production
Official implementations for various pre-training models of ERNIE-family, covering topics of Language Understanding & Generation, Multimodal Understanding & Generation, and beyond.
Google AI 2018 BERT pytorch implementation
:mag: Haystack is an open source NLP framework that leverages pre-trained Transformer models. It enables developers to quickly implement production-ready semantic search, question answering, summarization and document ranking for a wide range of NLP applications.
Easy-to-use and powerful NLP library with Awesome model zoo, supporting wide-range of NLP tasks from research to industrial applications, including Neural Search, Question Answering, Information Extraction and Sentiment Analysis end-to-end system.
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
Tensorflow solution of NER task Using BiLSTM-CRF model with Google BERT Fine-tuning And private Server services
A LITE BERT FOR SELF-SUPERVISED LEARNING OF LANGUAGE REPRESENTATIONS, 海量中文预训练ALBERT模型
Tutorials on getting started with PyTorch and TorchText for sentiment analysis.
pycorrector is a toolkit for text error correction. 文本纠错，Kenlm，ConvSeq2Seq，BERT，MacBERT，ELECTRA，ERNIE，Transformer，T5等模型实现，开箱即用。
A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
State of the Art Natural Language Processing
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Leveraging BERT and c-TF-IDF to create easily interpretable topics.
该仓库主要记录 NLP 算法工程师相关的顶会论文研读笔记
Implementation of BERT that could load official pre-trained models for feature extraction and prediction
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project: http://casl-project.ai/
LightSeq: A High Performance Library for Sequence Processing and Generation
RoBERTa中文预训练模型: RoBERTa for Chinese
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
Top2Vec learns jointly embedded topic, document and word vectors.
A curated list of pretrained sentence and word embedding models
Multi-Task Deep Neural Networks for Natural Language Understanding
Must-read papers on prompt-based tuning for pre-trained language models.
bert nlp papers, applications and github resources, including the newst xlnet ， BERT、XLNet 相关论文和 github 项目
This repository contains demos I made with the Transformers library by HuggingFace.
Datasets, tools, and benchmarks for representation learning of code.
Language Understanding Evaluation benchmark for Chinese: datasets, baselines, pre-trained models,corpus and leaderboard
Jupyter Notebook tutorials on solving real-world problems with Machine Learning & Deep Learning using PyTorch. Topics: Face detection with Detectron 2, Time Series anomaly detection with LSTM Autoencoders, Object Detection with YOLO v5, Build your first Neural Network, Time Series forecasting for Coronavirus daily cases, Sentiment Analysis with BERT.