Skip to main content

Building AI Agents with Volary

Volary is a system of persistent memory that can integrate with any AI agent. Route your LLM requests through Volary and it automatically learns from past conversations, injecting relevant knowledge into future sessions.

How it works

  1. 1Proxy requests through Volary – point your agent at a Volary agent endpoint instead of directly at OpenAI or Anthropic. Volary forwards the request to your configured backend.
  2. 2Reflections are extracted – after successful conversations, Volary analyses the exchange and extracts factual knowledge (where to find things) and procedural knowledge (how to avoid mistakes).
  3. 3Knowledge is injected – on future requests, Volary retrieves relevant reflections via vector similarity search and injects them as context before forwarding to your LLM provider.

Setup

To start using Volary with your agent, sign up in the Volary dashboard and follow the setup steps to create an “API Provider” agent.

Once you're done, you can configure your agent with its URL and access token to communicate through Volary as you would with your normal LLM backend.

Key endpoints

Each agent exposes API-compatible endpoints that you can use as drop-in replacements:

EndpointDescription
POST /v0/orgs/{org}/agents/{agent}/v1/chat/completionsOpenAI-compatible chat completions with reflection injection
POST /v0/orgs/{org}/agents/{agent}/v1/messagesAnthropic-compatible messages with reflection injection
POST /v0/orgs/{org}/agents/{agent}/v1/responsesOpenAI Responses API with stateful conversation chains
GET /v0/orgs/{org}/agents/{agent}/v0/mcpAct as a Model Context Protocol server for your agent.

See the API Reference for complete endpoint documentation.