AttentionVCGitHubVC
Trending ReposTrending BuildersPortfolio
AttentionVC

Built by JustinZ and Jennie

AttentionVCA product by AttentionVC
Back to Trending
microsoft
microsoft/

LLMLingua

[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.

Python
Stars

5.9k

+3 today+10 /wk+31 /mo
Forks

352

Issues

108

Watchers

5.9k

Star History

Repository Info

LicenseMIT
CreatedJul 7, 2023
Last push10/28/2025
Homepagellmlingua.com/
Open on GitHub