PersistQ combines powerful semantic search, local embeddings, and transparent pricing to give your AI applications long-term memory at scale.
Save hundreds per month with local Transformers.js embeddings. No per-token charges, no OpenAI API costs.
Vector-based search finds relevant memories even with fuzzy queries. Powered by pgvector and automatic embeddings.
Get started in seconds with our intuitive REST API. Works with any language or framework.
Your data never leaves your infrastructure for embeddings. No third-party AI service dependencies.
Built on battle-tested PostgreSQL with pgvector extension. No vendor lock-in, familiar technology.
One prompt and Claude Code sets up PersistQ via MCP. Zero manual configuration needed.
Everything you need to build AI applications with long-term memory
Store memories with automatic embedding generation. Retrieve by semantic search or exact match.
Organize memories with groups and tags. Filter searches by category for precise results.
Attach custom metadata to memories. Search and filter by any field.
Create, read, update, and delete memories via simple API endpoints.
Semantic search powered by 384-dimensional embeddings. Find relevant memories by meaning.
Traditional text-based search for exact matches. Fast and reliable.
Combines semantic and keyword search for best results. Automatic query optimization.
Filter by groups, tags, metadata, and date ranges. Narrow down results efficiently.
Clear documentation with examples in multiple languages. Quick start guides included.
Web dashboard to view memories, manage API keys, and monitor usage.
Built-in rate limiting to protect your application. Configurable per tier.
Track API usage, storage, and costs. Basic analytics on free tier, advanced on paid.