Open Research

Proof, not promises.

We publish our benchmarks, architectural decisions, and performance data. Every claim is backed by reproducible results.

21.2%
Cache Hit Rate
Requests served without touching the LLM
3.3%
Token Savings
Net reduction in input tokens at 100 conversations
<1ms
Cache Latency
Response time for exact cache hits via KV
0
Errors
100% success rate across all benchmark runs

See the savings for yourself

Integrate Decyra in 3 lines of code. Watch your AI agent costs drop as the cache warms — backed by the same methodology in our published benchmarks.