Search Portage & Overlays:
Newest
News
Repository news
GLSAs
Browse
USE Flags
Overlays
More...
dev-python
/llmlingua
To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
Screenshots
https://github.com/microsoft/LLMLingua
llmlingua-0.2.1
~amd64 ~x86
dev quality python_targets_python3_11 python_targets_python3_12 python_targets_python3_13 python_targets_python3_14
View
Download
Browse
License: MIT
Overlay:
pypi
ChangeLog
USE Flags
Dependencies
Reverse Deps
Related Bugs