There are 12 repositories under language-modeling topic.
Differentiable architecture search for convolutional and recurrent networks
Plug and Play Language Model implementation. Allows to steer topic and attributes of GPT-2 models.
Keras implementation of BERT with pre-trained weights
👄 The most accurate natural language detection library for Go, suitable for long and short text alike
A Modern C++ Data Sciences Toolkit
End-to-end ASR/LM implementation with PyTorch
Tasks Assessing Protein Embeddings (TAPE), a set of five biologically relevant semi-supervised learning tasks spread across different domains of protein biology.
Official PyTorch Repo for "ReZero is All You Need: Fast Convergence at Large Depth"
An implementation of DeepMind's Relational Recurrent Neural Networks (NeurIPS 2018) in PyTorch.
A Contrastive Framework for Neural Text Generation
XLNet for generating language.
Lyrics Generator aka Character-level Language Modeling with Multi-layer LSTM Recurrent Neural Network
Curso práctico: NLP de cero a cien 🤗
Use tensorflow's tf.scan to build vanilla, GRU and LSTM RNNs
Tasks Assessing Protein Embeddings (TAPE), a set of five biologically relevant semi-supervised learning tasks spread across different domains of protein biology. (DEPRECATED)
A fast LSTM Language Model for large vocabulary language like Japanese and Chinese
Independently Recurrent Neural Networks (IndRNN) implemented in pytorch.
Comparatively fine-tuning pretrained BERT models on downstream, text classification tasks with different architectural configurations in PyTorch.
仿 Scikit-Learn 设计的深度学习自然语言处理框架, 支持约 40 种模型类, 涵盖语言模型、文本分类、NER、MRC、机器翻译等各个领域
Training an n-gram based Language Model using KenLM toolkit for Deep Speech 2
Character-Level language models
Pre-training of Language Models for Language Understanding
Protein sequence classification with self-supervised pretraining
Rank-One Model Editing for Locating and Editing Factual Knowledge in GPT
Code accompanying our ECCV-2020 paper on 3D Neural Listeners.
Attentive Federated Learning for Private NLM
Web Native language Workbench with Projectional Web Editor
Recurrent Neural Networks (RNN, GRU, LSTM) and their Bidirectional versions (BiRNN, BiGRU, BiLSTM) for word & character level language modelling in Theano
[JMLR 2020] NeurIPS 2019 MicroNet Challenge Efficient Language Modeling, Champion
A toolkit for pre-processing large source code corpora
Official repository for the paper "Going Beyond Linear Transformers with Recurrent Fast Weight Programmers" (NeurIPS 2021)
MozoLM: A language model (LM) serving library
Implement modern LSTM cell by tensorflow and test them by language modeling task for PTB. Highway State Gating, Hypernets, Recurrent Highway, Attention, Layer norm, Recurrent dropout, Variational dropout.
Neural Networks for Protein Sequence Alignment