Xiaoyu Zhang (BBuf)

BBuf

Geek Repo

Company:SkyWork

Location:ChengDu

Home Page:www.giantpandacv.com

Github PK Tool:Github PK Tool

Xiaoyu Zhang's repositories

tvm_mlir_learn

compiler learning resources collect.

how-to-optim-algorithm-in-cuda

how to optimize some algorithm in cuda.

how-to-learn-deep-learning-framework

how to learn PyTorch and OneFlow

License:Apache-2.0Stargazers:232Issues:6Issues:0

giantpandacv.com

www.giantpandacv.com

Language:PythonLicense:NOASSERTIONStargazers:137Issues:9Issues:2
Language:PythonStargazers:13Issues:3Issues:0
Language:PythonStargazers:10Issues:0Issues:0

mlc-llm-code-analysis

Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.

Language:PythonLicense:Apache-2.0Stargazers:8Issues:1Issues:0

tvm

Open deep learning compiler stack for cpu, gpu and specialized accelerators

Language:PythonLicense:Apache-2.0Stargazers:7Issues:1Issues:0

cpp_related_tips

📚 C/C++ 技术面试基础知识总结,包括语言、程序库、数据结构、算法、系统、网络、链接装载库等知识及面试经验、招聘、内推等信息。This repository is a summary of the basic knowledge of recruiting job seekers and beginners in the direction of C/C++ technology, including language, program library, data structure, algorithm, system, network, link loading library, interview experience, recruitment, recommendation, etc.

Language:C++License:NOASSERTIONStargazers:5Issues:1Issues:0
Language:CudaStargazers:4Issues:1Issues:0

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:4Issues:1Issues:0
Language:PythonLicense:MITStargazers:3Issues:2Issues:0

opencompass

OpenCompass is an LLM evaluation platform, supporting a wide range of models (LLaMA, LLaMa2, ChatGLM2, ChatGPT, Claude, etc) over 50+ datasets.

Language:PythonLicense:Apache-2.0Stargazers:3Issues:0Issues:0

tvm-cn

TVM Documentation in Chinese Simplified / TVM 中文文档

Language:JavaScriptStargazers:3Issues:1Issues:0

How_to_optimize_in_GPU

This is a series of GPU optimization topics. Here we will introduce how to optimize the CUDA kernel in detail. I will introduce several basic kernel optimizations, including: elementwise, reduce, sgemv, sgemm, etc. The performance of these kernels is basically at or near the theoretical limit.

Language:CudaLicense:Apache-2.0Stargazers:2Issues:1Issues:0

tokenizers-cpp

Universal cross-platform tokenizers binding to HF and sentencepiece

Language:C++License:Apache-2.0Stargazers:2Issues:1Issues:0

FasterTransformer

Transformer related optimization, including BERT, GPT

Language:C++License:Apache-2.0Stargazers:1Issues:1Issues:0
Language:PythonLicense:NOASSERTIONStargazers:1Issues:1Issues:0

LLaMA-Factory

Easy-to-use LLM fine-tuning framework (LLaMA, BLOOM, Mistral, Baichuan, Qwen, ChatGLM)

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

stb_image_example

std based image encoder decoder example (c++)

Language:CStargazers:1Issues:2Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:1Issues:1Issues:0

ChatRWKV

ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

fastllm

纯c++的全平台llm加速库,支持python调用,chatglm-6B级模型单卡可达10000+token / s,支持glm, llama, moss基座,手机端流畅运行

Language:C++Stargazers:0Issues:1Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of autoregressive language models.

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

RWKV-CUDA

The CUDA version of the RWKV language model ( https://github.com/BlinkDL/RWKV-LM )

Language:CudaStargazers:0Issues:1Issues:0

tvm_gpu_gemm

play gemm with tvm

Language:CudaStargazers:0Issues:1Issues:0