James (jamesliu)

jamesliu

Geek Repo

Company:Bay Jarvis

Location:Palo Alto, CA

Home Page:jamesliuai.com

Github PK Tool:Github PK Tool

James's repositories

nanoPPO

An efficient implementation of the Proximal Policy Optimization (PPO) algorithm with linear and attention policy for reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:6Issues:3Issues:0

nanoDPO

A nimble and innovative implementation of the Direct Preference Optimization (DPO) algorithm with Causal Transformer and LSTM model, inspired by the paper of DPO in fine-tuning unsupervised Language Models

Language:PythonLicense:Apache-2.0Stargazers:5Issues:3Issues:0

nanoTransformer

A PyTorch-based featuring an efficiently implemented Transformer model. The core of our attention mechanisms is powered by torch.einsum, ensuring clean, readable, and highly optimized tensor operations.

Language:PythonStargazers:2Issues:2Issues:0

nChain

a flexible and efficient implementation to create LLM bots over extensible dataset.

Language:PythonStargazers:2Issues:0Issues:0

amago

a simple and scalable agent for training adaptive policies with sequence-based RL

License:MITStargazers:0Issues:0Issues:0

anthropic-cookbook

A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.

License:MITStargazers:0Issues:0Issues:0

autogen

Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ

Language:Jupyter NotebookLicense:CC-BY-4.0Stargazers:0Issues:0Issues:0
Stargazers:0Issues:0Issues:0

devika

Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.

License:MITStargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

litgpt

Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

License:Apache-2.0Stargazers:0Issues:0Issues:0

LLaMA-Factory

Unify Efficient Fine-tuning of 100+ LLMs

License:Apache-2.0Stargazers:0Issues:0Issues:0

llm-foundry

LLM training code for MosaicML foundation models

License:Apache-2.0Stargazers:0Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Stargazers:0Issues:0Issues:0

LLMLingua

To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.

License:MITStargazers:0Issues:0Issues:0
License:MITStargazers:0Issues:0Issues:0

long-range-arena

Long Range Arena for Benchmarking Efficient Transformers

License:Apache-2.0Stargazers:0Issues:0Issues:0

magicoder

Magicoder: Source Code Is All You Need

Language:PythonLicense:MITStargazers:0Issues:1Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0
Stargazers:0Issues:0Issues:0

OpenMoE

A family of open-sourced Mixture-of-Experts (MoE) Large Language Models

Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0
Language:ShellStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0

routerbench

The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System

Stargazers:0Issues:0Issues:0

ScaleLLM

A high-performance inference system for large language models, designed for production environments.

License:Apache-2.0Stargazers:0Issues:0Issues:0

SPIN

The official implementation of Self-Play Fine-Tuning (SPIN)

License:Apache-2.0Stargazers:0Issues:0Issues:0

Time-LLM

[ICLR 2024] Official implementation of "Time-LLM: Time Series Forecasting by Reprogramming Large Language Models"

License:Apache-2.0Stargazers:0Issues:0Issues:0

TinyLlama

The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.

License:Apache-2.0Stargazers:0Issues:0Issues:0

Transformers_Are_What_You_Dont_Need

The best repository showing why transformers don’t work in time series forecasting and showcasing the best SOTA non transformer models.

Stargazers:0Issues:0Issues:0