Point your base_url at cache.kaissa.ai,
keep your own API key, and we cache your LLM calls semantically —
zero code changes.
Sign in and paste your OpenAI or Anthropic key. We store it encrypted and issue you a Kaissa virtual key.
Update your base_url. Use your Kaissa virtual key as the API key. Everything else stays the same.
Every semantically similar request hits our HNSW cache (0.3ms). Only true misses reach the provider and cost tokens.
Sign in with Google, paste your API key, change one env var. You're caching.
Sign in with GoogleAdd your OpenAI or Anthropic key. We store it encrypted and route your requests through it.