Renato Negrinho (negrinho)

negrinho

Geek Repo

Company:Carnegie Mellon University

Location:Pittsburgh

Home Page:http://www.cs.cmu.edu/~negrinho/

Github PK Tool:Github PK Tool

Renato Negrinho's starred repositories

devops-exercises

Linux, Jenkins, AWS, SRE, Prometheus, Docker, Python, Ansible, Git, Kubernetes, Terraform, OpenStack, SQL, NoSQL, Azure, GCP, DNS, Elastic, Network, Virtualization. DevOps Interview Questions

Language:PythonLicense:NOASSERTIONStargazers:62866Issues:1179Issues:82

ollama

Get up and running with Llama 2, Mistral, Gemma, and other large language models.

gpt-engineer

Specify what you want it to build, the AI asks for clarification, and then builds it.

Language:PythonLicense:MITStargazers:50004Issues:498Issues:438

tinygrad

You like pytorch? You like micrograd? You love tinygrad! ❤️

Language:PythonLicense:MITStargazers:23370Issues:264Issues:555

alpaca-lora

Instruct-tune LLaMA on consumer hardware

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:18081Issues:157Issues:466

qdrant

Qdrant - High-performance, massive-scale Vector Database for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/

Language:RustLicense:Apache-2.0Stargazers:17447Issues:114Issues:1003

llama2.c

Inference Llama 2 in one file of pure C

SuperAGI

<⚡️> SuperAGI - A dev-first open source autonomous AI agent framework. Enabling developers to build, manage & run useful autonomous agents quickly and reliably.

Language:PythonLicense:MITStargazers:14200Issues:164Issues:394

gorilla

Gorilla: An API store for LLMs

Language:PythonLicense:Apache-2.0Stargazers:9629Issues:99Issues:108

minbpe

Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.

Language:PythonLicense:MITStargazers:7751Issues:77Issues:21

ZLUDA

CUDA on AMD GPUs

Language:RustLicense:Apache-2.0Stargazers:7345Issues:96Issues:118

gemma.cpp

lightweight, standalone C++ inference engine for Google's Gemma models.

Language:C++License:Apache-2.0Stargazers:5105Issues:34Issues:56

gemma_pytorch

The official PyTorch implementation of Google's Gemma models

Language:PythonLicense:Apache-2.0Stargazers:4889Issues:37Issues:29

AITemplate

AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (NVIDIA GPU) and MatrixCore (AMD GPU) inference.

Language:PythonLicense:Apache-2.0Stargazers:4430Issues:83Issues:237

OLMo

Modeling, training, eval, and inference code for OLMo

Language:PythonLicense:Apache-2.0Stargazers:3645Issues:41Issues:128

neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime

Language:PythonLicense:Apache-2.0Stargazers:1900Issues:32Issues:180

intel-extension-for-transformers

⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡

Language:PythonLicense:Apache-2.0Stargazers:1883Issues:26Issues:123

reflexion

[NeurIPS 2023] Reflexion: Language Agents with Verbal Reinforcement Learning

Language:PythonLicense:MITStargazers:1860Issues:29Issues:28

DeepSpeed-MII

MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.

Language:PythonLicense:Apache-2.0Stargazers:1580Issues:36Issues:246

prm800k

800,000 step-level correctness labels on LLM solutions to MATH problems

Language:PythonLicense:MITStargazers:1255Issues:112Issues:14

oneAPI-samples

Samples for Intel® oneAPI Toolkits

Language:C++License:MITStargazers:820Issues:30Issues:127

intel-extension-for-tensorflow

Intel® Extension for TensorFlow*

Language:C++License:NOASSERTIONStargazers:303Issues:17Issues:59

optimum-intel

🤗 Optimum Intel: Accelerate inference with Intel optimization tools

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:301Issues:36Issues:59

OLMo-Eval

Evaluation suite for LLMs

Language:PythonLicense:Apache-2.0Stargazers:245Issues:6Issues:2

awesome-llm-human-preference-datasets

A curated list of Human Preference Datasets for LLM fine-tuning, RLHF, and eval.

License:MITStargazers:217Issues:7Issues:0

zeno

AI Data Management & Evaluation Platform

Language:SvelteLicense:MITStargazers:206Issues:8Issues:231

neural-speed

An innovative library for efficient LLM inference via low-bit quantization

Language:C++License:Apache-2.0Stargazers:139Issues:6Issues:20

SpeculativeDecodingPapers

📰 Must-read papers and blogs on Speculative Decoding ⚡️

License:Apache-2.0Stargazers:121Issues:6Issues:0

llm-on-ray

Pretrain, finetune and serve LLMs on Intel platforms with Ray

Language:PythonLicense:Apache-2.0Stargazers:33Issues:8Issues:54