Building AI Agents with Volary
Volary is a system of persistent memory that can integrate with any AI agent. Route your LLM requests through Volary and it automatically learns from past conversations, injecting relevant knowledge into future sessions.
How it works
- 1Proxy requests through Volary – point your agent at a Volary agent endpoint instead of directly at OpenAI or Anthropic. Volary forwards the request to your configured backend.
- 2Reflections are extracted – after successful conversations, Volary analyses the exchange and extracts factual knowledge (where to find things) and procedural knowledge (how to avoid mistakes).
- 3Knowledge is injected – on future requests, Volary retrieves relevant reflections via vector similarity search and injects them as context before forwarding to your LLM provider.
Setup
To start using Volary with your agent, sign up in the Volary dashboard and follow the setup steps to create an “API Provider” agent.
Once you're done, you can configure your agent with its URL and access token to communicate through Volary as you would with your normal LLM backend.
Key endpoints
Each agent exposes API-compatible endpoints that you can use as drop-in replacements:
| Endpoint | Description |
|---|---|
POST /v0/orgs/{org}/agents/{agent}/v1/chat/completions | OpenAI-compatible chat completions with reflection injection |
POST /v0/orgs/{org}/agents/{agent}/v1/messages | Anthropic-compatible messages with reflection injection |
POST /v0/orgs/{org}/agents/{agent}/v1/responses | OpenAI Responses API with stateful conversation chains |
GET /v0/orgs/{org}/agents/{agent}/v0/mcp | Act as a Model Context Protocol server for your agent. |
See the API Reference for complete endpoint documentation.
