Nln's repositories
UnsupervisedMT
Phrase-Based & Neural Unsupervised Machine Translation
babyai
BabyAI platform. A testbed for training agents to understand and execute language commands.
biobert
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
biobert-pretrained
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
BioRelEx
BioRelEx: Biological Relation Extraction Benchmark
BoundaryNet
BoundaryNet - A Semi-Automatic Layout Annotation Tool
Clinical-Trial-Parser
Library for converting clinical trial eligibility criteria to a machine-readable format.
clinicalBERT
repository for Publicly Available Clinical BERT Embeddings
crfpar
Code for ACL'20 paper "Efficient Second-Order TreeCRF for Neural Dependency Parsing" and IJCAI'20 paper "Fast and Accurate Neural CRF Constituency Parsing".
ct_warc_to_doc
Source code to extract content from commoncrawl news corpus and upload to S3
DeepFaceLab
DeepFaceLab is a tool that utilizes deep learning to recognize and swap faces in pictures and videos. Includes prebuilt ready to work standalone Windows 7,8,10 binary (look readme.md).
doccano
Open source text annotation tool for machine learning practitioner.
docker-elk
The Elastic stack (ELK) powered by Docker and Compose.
faiss
A library for efficient similarity search and clustering of dense vectors.
few-shot
Repository for few-shot learning machine learning projects
gpt-2-simple
Python package to easily retrain OpenAI's GPT-2 text-generating model on new texts
leo
Implementation of Meta-Learning with Latent Embedding Optimization
lit
The Language Interpretability Tool: Interactively analyze NLP models for model understanding in an extensible and framework agnostic interface.
MAML-Pytorch
Elegant PyTorch implementation of paper Model-Agnostic Meta-Learning (MAML)
parser
A collection of state-of-the-art syntactic parsing models based on Biaffine Parser.
PreSumm
code for EMNLP 2019 paper Text Summarization with Pretrained Encoders
ProphetNet
ProphetNet: Predicting Future N-gram for Sequence-to-Sequence Pre-training https://arxiv.org/pdf/2001.04063.pdf
SpanBERT
Code for using and evaluating SpanBERT.
unilm
UniLM AI - Large-scale Self-supervised Pre-training across Tasks, Languages, and Modalities
wikiextractor
A tool for extracting plain text from Wikipedia dumps