David Herel (DavidHerel)

DavidHerel

Geek Repo

Company:CIIRC

Location:Czech Republic

Home Page:davidherel.com

Github PK Tool:Github PK Tool

David Herel's starred repositories

collapse-lm-iclr

"Collapse of Self-trained Language Models" codebase for ICLR 2024

Language:PythonLicense:MITStargazers:2Issues:0Issues:0

fundus

A very simple news crawler with a funny name

Language:PythonLicense:MITStargazers:210Issues:0Issues:0

minbpe

Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.

Language:PythonLicense:MITStargazers:8772Issues:0Issues:0

API

Documentation and Samples for the Official HN API

License:MITStargazers:11176Issues:0Issues:0

GNews

A Happy and lightweight Python Package that Provides an API to search for articles on Google News and returns a JSON response.

Language:PythonLicense:MITStargazers:593Issues:0Issues:0

litgpt

20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.

Language:PythonLicense:Apache-2.0Stargazers:8890Issues:0Issues:0

llama.cpp

LLM inference in C/C++

Language:C++License:MITStargazers:61907Issues:0Issues:0

llama2.c

Inference Llama 2 in one file of pure C

Language:CLicense:MITStargazers:16861Issues:0Issues:0

gigaGPT

a small code base for training large models

Language:PythonLicense:Apache-2.0Stargazers:244Issues:0Issues:0
Language:PythonStargazers:1Issues:0Issues:0

arctic_shift

Making Reddit data accessible to researchers, moderators and everyone else. Interact with the data through large dumps, an API or web interface.

Language:TypeScriptStargazers:203Issues:0Issues:0

nanoRWKV

RWKV in nanoGPT style

Language:PythonLicense:MITStargazers:159Issues:0Issues:0

sota_lm

Repository containing state-of-the-art ensemble for several language modelling benchmarks

Language:HTMLLicense:MITStargazers:7Issues:0Issues:0

lit-llama

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Language:PythonLicense:Apache-2.0Stargazers:5899Issues:0Issues:0

TinyLlama

The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.

Language:PythonLicense:Apache-2.0Stargazers:7363Issues:0Issues:0

TAADpapers

Must-read Papers on Textual Adversarial Attack and Defense

Language:PythonLicense:MITStargazers:1472Issues:0Issues:0

EvolvingModularRobots_Unity

Software for Evolving Modular Robots in Unity

Language:PythonLicense:GPL-3.0Stargazers:11Issues:0Issues:0

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs.

Language:PythonLicense:MITStargazers:34930Issues:0Issues:0

llama

Inference code for Llama models

Language:PythonLicense:NOASSERTIONStargazers:54291Issues:0Issues:0

cc_net

Tools to download and cleanup Common Crawl data

Language:PythonLicense:MITStargazers:938Issues:0Issues:0

awd-lstm-lm

LSTM and QRNN Language Model Toolkit for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:1959Issues:0Issues:0

why-I-hate-wow-private-servers

Reasons why most of WoW private servers sucks

License:AGPL-3.0Stargazers:93Issues:0Issues:0

davidherel.github.io

Source code for my website: https://davidherel.com

Language:HTMLStargazers:2Issues:0Issues:0

semantics-preserving-encoder

Python library providing a simple, fully supervised sentence embedding technique for textual adversarial attacks.

Language:PythonLicense:MITStargazers:12Issues:0Issues:0

TabPFN

Official implementation of the TabPFN paper (https://arxiv.org/abs/2207.01848) and the tabpfn package.

Language:PythonLicense:Apache-2.0Stargazers:1144Issues:0Issues:0
Language:Jupyter NotebookStargazers:5Issues:0Issues:0
Language:PythonStargazers:436Issues:0Issues:0

LaMDA-rlhf-pytorch

Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.

Language:PythonLicense:MITStargazers:458Issues:0Issues:0

deep_learning_curriculum

Language model alignment-focused deep learning curriculum

Stargazers:1168Issues:0Issues:0

RWKV-LM

RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

Language:PythonLicense:Apache-2.0Stargazers:12015Issues:0Issues:0