Mark (M-a-r-k)

M-a-r-k

Geek Repo

0

followers

0

following

Github PK Tool:Github PK Tool

Mark's starred repositories

LLMLingua

[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.

Language:PythonLicense:MITStargazers:4605Issues:0Issues:0