ldwang's repositories

License:MITStargazers:0Issues:0Issues:0

ComfyUI

The most powerful and modular stable diffusion GUI, api and backend with a graph/nodes interface.

License:GPL-3.0Stargazers:0Issues:0Issues:0

deita

Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DiT

Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"

License:NOASSERTIONStargazers:0Issues:0Issues:0

Firefly

Firefly(流萤): 中文对话式大语言模型

Language:PythonStargazers:0Issues:0Issues:0

FlagScale

FlagScale is a Large Language Model (LLM) toolkit based on open-sourced projects.

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

how-to-optim-algorithm-in-cuda

how to optimize some algorithm in cuda.

Stargazers:0Issues:0Issues:0

Latte

Latte: Latent Diffusion Transformer for Video Generation.

License:Apache-2.0Stargazers:0Issues:0Issues:0

LESS

Preprint: Less: Selecting Influential Data for Targeted Instruction Tuning

License:MITStargazers:0Issues:0Issues:0

LLaMA-Factory

Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

LLaMA-Pro

Progressive LLaMA with Block Expansion.

License:Apache-2.0Stargazers:0Issues:0Issues:0

llm-foundry

LLM training code for MosaicML foundation models

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
License:Apache-2.0Stargazers:0Issues:0Issues:0

LLMTest_NeedleInAHaystack

Doing simple retrieval from LLM models at various context lengths to measure accuracy

License:NOASSERTIONStargazers:0Issues:0Issues:0

megalodon

Reference implementation of Megalodon 7B model

License:MITStargazers:0Issues:0Issues:0

MiniCPM

MiniCPM-2.4B: An end-side LLM outperforms Llama2-13B.

License:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

MoE-LLaVA

Mixture-of-Experts for Large Vision-Language Models

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

OLMo

Modeling, training, eval, and inference code for OLMo

License:Apache-2.0Stargazers:0Issues:0Issues:0

Open-Sora-Plan

This project aim to reproducing Sora (Open AI T2V model), but we only have limited resource. We deeply wish the all open source community can contribute to this project.

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:0Issues:0Issues:0

open_clip

An open source implementation of CLIP.

License:NOASSERTIONStargazers:0Issues:0Issues:0

QAnything

Question and Answer based on Anything.

License:Apache-2.0Stargazers:0Issues:0Issues:0

QuRating

Select LM Training Data Based on Qualitative Aspects of Text

Language:PythonStargazers:0Issues:0Issues:0

qwen-vllm

通义千问VLLM推理部署DEMO

Stargazers:0Issues:0Issues:0

Qwen1.5

Qwen1.5 is the improved version of Qwen, the large language model series developed by Qwen team, Alibaba Cloud.

Stargazers:0Issues:0Issues:0

SiT

Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"

License:MITStargazers:0Issues:0Issues:0

stable-diffusion

A latent text-to-image diffusion model

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:0Issues:0Issues:0

stable-weight-decay-regularization

[NeurIPS 2023] The PyTorch Implementation of Scheduled (Stable) Weight Decay.

License:MITStargazers:0Issues:0Issues:0
License:MITStargazers:0Issues:0Issues:0

Video-LLaVA

Video-LLaVA: Learning United Visual Representation by Alignment Before Projection

License:Apache-2.0Stargazers:0Issues:0Issues:0