Ruonan Wang (rnwang04)

rnwang04

Geek Repo

Company:Intel

Location:Shanghai, China

Github PK Tool:Github PK Tool

Ruonan Wang's repositories

Language:HTMLStargazers:0Issues:0Issues:0

bloomz.cpp

C++ implementation for BLOOM

Language:CLicense:MITStargazers:0Issues:0Issues:0
Language:HTMLLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Folder-Structure-Conventions

Folder / directory structure options and naming conventions for software projects

License:MITStargazers:0Issues:0Issues:0

HelloGitHub

:octocat: 分享 GitHub 上有趣、入门级的开源项目。Share interesting, entry-level open source projects on GitHub.

Language:PythonStargazers:0Issues:0Issues:0

ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

lightning

The most intuitive, flexible, way to build PyTorch models and lightning apps that glue together everything around the models, without the pain of infrastructure, cost management, scaling and everything else.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llama.cpp

LLM inference in C/C++

Language:C++License:MITStargazers:0Issues:0Issues:0

LongLoRA

Code and documents of LongLoRA and LongAlpaca

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

neural-compressor

Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0