There are 6 repositories under huggingface-transformers topic.
🧠💬 Articles I wrote about machine learning, archived from MachineCurve.com.
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
Trained models & code to predict toxic comments on all 3 Jigsaw Toxic Comment Challenges. Built using ⚡ Pytorch Lightning and 🤗 Transformers. For access to our API, please email us at contact@unitary.ai.
Multimodal model for text and tabular data with HuggingFace transformers as building block for text data
[EMNLP 2022] A Unified Framework and Analysis for Structured Knowledge Grounding with Text-to-Text Language Models
Sentiment analysis neural network trained by fine-tuning BERT, ALBERT, or DistilBERT on the Stanford Sentiment Treebank.
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
Package to compute Mauve, a similarity score between neural text and human text. Install with `pip install mauve-text`.
Build and train state-of-the-art natural language processing models using BERT
Amazon SageMaker Local Mode Examples
QuickAI is a Python library that makes it extremely easy to experiment with state-of-the-art Machine Learning models.
extending stable diffusion prompts with suitable style cues using text generation
Fast Inference Solutions for BLOOM
Dreambooth implementation based on Stable Diffusion with minimal code.
Indonesian Language Models and its Usage
MLOps for Vision Models (TensorFlow) from 🤗 Transformers with TensorFlow Extended (TFX)
[2021 훈민정음 한국어 음성•자연어 인공지능 경진대회] 대화요약 부문 알라꿍달라꿍 팀의 대화요약 학습 및 추론 코드를 공유하기 위한 레포입니다.
KLUE 데이터를 활용한 HuggingFace Transformers 튜토리얼
Serve scikit-learn, XGBoost, TensorFlow, and PyTorch models with AWS Lambda container images support.
make differentially private training of transformers easy
Adapt Transformer-based language models to new text domains
This repo is the home of Romanian Transformers.
Phoneme Recognition using pre-trained models Wav2vec2, HuBERT and WavLM. Throughout this project, we compared specifically three different self-supervised models, Wav2vec (2019, 2020), HuBERT (2021) and WavLM (2022) pretrained on a corpus of English speech that we will use in various ways to perform phoneme recognition for different languages with a network trained with Connectionist Temporal Classification (CTC) algorithm.
Multi-Class Text Classification for products based on their description with Machine Learning algorithms and Neural Networks (MLP, CNN, Distilbert).
The source code for the Cutoff data augmentation approach proposed in this paper: "A Simple but Tough-to-Beat Data Augmentation Approach for Natural Language Understanding and Generation".
Post-process Amazon Textract results with Hugging Face transformer models for document understanding
Tensorflow, Pytorch, Huggingface Transformer, Fastai, etc. tutorial Colab Notebooks.
BioELECTRA
This repository contains Natural Language Generation (NLG) models aimed on generating text of fairy tales (Markov Chain, LSTM neural network, GPT-2 Transformers).
[Pytorch] A GPT2-ChatBot implemented by huggingface/transformers
This repository contains the implementation of the paper -- Bi-Bimodal Modality Fusion for Correlation-Controlled Multimodal Sentiment Analysis
spaCy-wrap is a wrapper library for spaCy for including fine-tuned transformers from Huggingface in your spaCy pipeline allowing you to include existing fine-tuned models within your SpaCy workflow.