Overview
Compresr reduces LLM token costs through intelligent context compression.
Compression Models
espresso_v1General-purpose compression — no query needed. Removes redundant tokens while preserving meaning. Ideal for pre-compressing documents, system prompts, or any context you want to reuse across multiple queries.
latte_v1Query-specific compression that preserves tokens relevant to a given query. Ideal for RAG pipelines and Q&A systems where you want to keep answer-relevant information while compressing the rest.
Quick Start
- Get your API key from the Dashboard
- Install the SDK:
pip install compresr - Start compressing — see the Quick Start guide