dev-python/llmlingua
To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
Bugs
| Bug # | Severity | Platform | Status | Description |
|---|
These bugs were grabbed from http://bugs.gentoo.org and have only passed a preliminary search using the package title,
to do a more through search please visit: http://bugs.gentoo.org.


View
Download
Browse