AinL (gmlwns2000)

gmlwns2000

User data from Github https://github.com/gmlwns2000

Company:School of Complete Solitude : Undergraduate Student @ KAIST

Location:Anyang, Korea

Home Page:neuralaction.github.io

GitHub:@gmlwns2000


Organizations
Kawaian
NeuralAction

AinL's repositories

sharkshark-4k

Upscale Twitch stream and restream into Twitch or RTMP or File.

Language:PythonLicense:NOASSERTIONStargazers:15Issues:2Issues:3

sea-attention

Official Implementation of SEA: Sparse Linear Attention with Estimated Attention Mask (ICLR 2024)

Language:PythonStargazers:10Issues:3Issues:0

streaming-llm-triton

OpenAI Triton Implementation of Streaming LLM

Language:PythonStargazers:8Issues:2Issues:0

sttabt

[ICLR2023] Official code of Sparse Token Transformer with Attention Back-Tracking

Language:Jupyter NotebookStargazers:6Issues:3Issues:0
Language:TeXLicense:GPL-3.0Stargazers:4Issues:1Issues:0
Language:PythonStargazers:4Issues:1Issues:0
Language:PythonLicense:NOASSERTIONStargazers:3Issues:6Issues:0

cs454-project

CS454 2023 F Team 4

Language:Jupyter NotebookStargazers:1Issues:1Issues:0

InfiniGen

InfiniGen: Efficient Generative Inference of Large Language Models with Dynamic KV Cache Management (OSDI'24)

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0

pypareto-native

Numba optimized version of `pypareto`. Sorting chains for pareto frontier extraction

Language:PythonLicense:MITStargazers:1Issues:0Issues:0

RULER-hip

This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?

Language:PythonLicense:Apache-2.0Stargazers:1Issues:0Issues:0
Language:Jupyter NotebookStargazers:0Issues:1Issues:0
Stargazers:0Issues:0Issues:0

EXAONE-3.5

Official repository for EXAONE 3.5 built by LG AI Research

License:NOASSERTIONStargazers:0Issues:0Issues:0
Stargazers:0Issues:1Issues:0

gmlwns2000.github.io

AcadHomepage: A Modern and Responsive Academic Personal Homepage

Language:SCSSLicense:MITStargazers:0Issues:0Issues:0

hip-attention

Training-free Post-training Efficient Sub-quadratic Complexity Attention. Implemented with OpenAI Triton.

Language:PythonStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0
Language:PythonStargazers:0Issues:1Issues:0
Language:Jupyter NotebookStargazers:0Issues:1Issues:0

InfiniteBench-hip

Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718

Language:PythonLicense:MITStargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of language models.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

loft-hip

LOFT: A 1 Million+ Token Long-Context Benchmark

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

LongBench-hip

LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

LongLM

LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

sglang-hip12

SGLang is a fast serving framework for large language models and vision language models. See hip12-offload-add-offload-cache

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

triton-fix-autotune

Development repository for the Triton language and compiler

Language:C++License:MITStargazers:0Issues:0Issues:0

vllm-timber

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0