Vui Seng Chua (vuiseng9)

vuiseng9

Geek Repo

Company:@Intel

Github PK Tool:Github PK Tool

Vui Seng Chua's repositories

nncf

PyTorch*-based Neural Network Compression Framework for enhanced OpenVINO™ inference

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

AMX-TMUL-Code-Samples

Code samples related to Intel(R) AMX

Language:CStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0

Diff-Pruning

Structural Pruning for Diffusion Models

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

diffusers

🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DLAI-LangChain-LLM-App

In LangChain for LLM Application Development, you will gain essential skills in expanding the use cases and capabilities of language models in application development using the LangChain framework.

Language:Jupyter NotebookStargazers:0Issues:0Issues:0

EAGLE

[ICML'24] EAGLE: Speculative Sampling Requires Rethinking Feature Uncertainty

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

fairseq

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

hf-peft

🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

ipex

A Python package for extending the official PyTorch that can easily obtain performance on Intel platform

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, ModelScope, etc

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

llm-awq

AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

meta-sam

The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:0Issues:0Issues:0

mlperf-inference

Reference implementations of MLPerf™ inference benchmarks

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

mlperf-v3.0-intel

This repository contains the results and code for the MLPerf™ Inference v3.0 benchmark.

License:Apache-2.0Stargazers:0Issues:0Issues:0

mlperf-v3.1-intel

This repository contains the results and code for the MLPerf™ Inference v3.1 benchmark.

License:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

optimum

🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools

License:Apache-2.0Stargazers:0Issues:0Issues:0

optimum-intel

Accelerate inference of 🤗 Transformers with Intel optimization tools

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

PowerInfer

High-speed Large Language Model Serving on PCs with Consumer-grade GPUs

Language:CLicense:MITStargazers:0Issues:0Issues:0

sd-perf

quick script to profile stable diffusion performance

Language:PythonStargazers:0Issues:2Issues:0

smoothquant

SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

SparseFinetuning

Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

speculative-sampling

Simple implementation of Speculative Sampling in NumPy for GPT-2.

Stargazers:0Issues:0Issues:0

SqueezeLLM

SqueezeLLM: Dense-and-Sparse Quantization

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

Teaching-Intel-Intrinsics-for-SIMD-Parallelism

Teaching Vectorization and SIMD using Intel Intrinsics in a Computer Organization and Architecture class

Language:C++License:MITStargazers:0Issues:0Issues:0

torchinfo

View model summaries in PyTorch!

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

transformers

🤗Transformers: State-of-the-art Natural Language Processing for Pytorch and TensorFlow 2.0.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:MITStargazers:0Issues:0Issues:0