There are 23 repositories under language-models topic.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
Open-source offline translation library written in Python
:house_with_garden: Fast & easy transfer learning for NLP. Harvesting language models for the industry. Focus on Question Answering.
Reasoning in Large Language Models: Papers and Resources, including Chain-of-Thought, Instruction-Tuning and Multimodality.
[Paper List] Papers integrating knowledge graphs (KGs) and large language models (LLMs)
LangChain & Prompt Engineering tutorials on Large Language Models (LLMs) such as ChatGPT with custom data. Jupyter notebooks on loading and indexing data, creating prompt templates, CSV agents, and using retrieval QA chains to query the custom data. Projects for using a private LLM (Llama 2) for chat with PDF files, tweets sentiment analysis.
A family of diffusion models for text-to-audio generation.
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
💁 Awesome Treasure of Transformers Models for Natural Language processing contains papers, videos, blogs, official repo along with colab Notebooks. 🛫☑️
日本語LLMまとめ - Overview of Japanese LLMs
[ACL 2023] Reasoning with Language Model Prompting: A Survey
[ACL 2021] LM-BFF: Better Few-shot Fine-tuning of Language Models https://arxiv.org/abs/2012.15723
PhoBERT: Pre-trained language models for Vietnamese (EMNLP-2020 Findings)
LLM-based ontological extraction tools, including SPIRES
Official implementation for HyenaDNA, a long-range genomic foundation model built with Hyena
Happy Transformer makes it easy to fine-tune and perform inference with NLP Transformer models.
This repository contains a collection of papers and resources on Reasoning in Large Language Models.
Crosslingual Generalization through Multitask Finetuning
ChatGPT Universe is fleeting notes on ChatGPT, GPT, and large language models (LLMs)
Pre-trained models and language resources for Natural Language Processing in Polish
Latency and Memory Analysis of Transformer Models for Training and Inference
💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client
Scaling Data-Constrained Language Models
ExtremeBERT is a toolkit that accelerates the pretraining of customized language models on customized datasets, described in the paper “ExtremeBERT: A Toolkit for Accelerating Pretraining of Customized BERT”.
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including an implementation of kNN-LM and kNN-MT
Unofficial API Wrapper for Perplexity.ai + Account Generator
This repository will guide you to create your own Smart Virtual Assistant like Google Assistant using Open AI's ChatGPT, Whisper. The entire solution is created using Python & Gradio.
This repository contains landmark research papers in Natural Language Processing that came out in this century.