Common questions about Seizn Memory API, from getting started to advanced operations.
Q1What is Seizn Memory and what problem does it solve?
Seizn Memory is an AI memory infrastructure that provides persistent, searchable memory for AI applications. Unlike vector databases that only store/search vectors, Seizn includes the full product layer: memory extraction, policy management, key management, deletion, audit logs, and SDKs. It solves the problem of maintaining context across AI sessions and enabling personalized AI experiences.
Getting Started
Q2How is Seizn different from vector databases like Pinecone or Weaviate?
Vector databases are storage/search infrastructure for vectors. Seizn is a complete memory system built on top of that, providing: automatic memory extraction from conversations, memory type classification, namespace/scope management, API key rotation, audit logging, SDKs, and governance features. Think of it as 'memory infrastructure' vs 'search infrastructure'.
Getting Started
Q3Do I need RAG to use Seizn?
No. The most common starting point is: store memory -> search -> inject into prompt. RAG (automatic context composition + response generation) is the next step. You can start simple and add complexity as needed.
Getting Started
Q4What's the fastest way to build a PoC?
1) POST /api/memories to store a user preference. 2) GET /api/memories to search. 3) Inject the results into your LLM prompt. 4) Later, add /api/extract for auto-extraction and /api/query for memory-augmented responses.
Getting Started
Q5What should I store in /api/memories?
Store information that remains valid across conversations: preferences (tone, language, format), facts (job, tools, project structure), instructions ("always summarize in tables"), relationships ("Alice is the team lead"). Avoid storing temporary or session-specific data unless using session scope.
Core Concepts
Q6What data should I NOT store?
Never store: passwords, API keys, tokens, session cookies (auth info), SSN, passport numbers, bank accounts (PII), credit card info (payment data). For temporary data, use session scope with TTL if needed.
Core Concepts
Q7Why is namespace important?
Namespace separates memories by project/tenant/environment. Without it, data gets mixed, search quality drops, and deletion/export becomes painful. Recommended: 'org:acme/app:chat/env:prod' or 'project:myapp/env:staging'. Never use just 'default' in production.
Core Concepts
Q8When should I use scope (user/session/agent)?
user: Preferences that apply to a user globally. session: Goals or context valid only for this conversation. agent: Rules specific to one agent in a multi-agent system. Using scope properly reduces prompt length and improves response consistency.
Core Concepts
Q9What are memory_types and why do they matter?
memory_type classifies memories: fact (unchanging info), preference (user choices), instruction (rules to follow), relationship (people/org connections), experience (past events). This is the most powerful axis for filtering, deletion, and policy application.
Core Concepts
Q10How do threshold and limit work?
limit: Number of candidate memories to retrieve (too low = miss relevant ones, too high = noisy context). threshold: 0-1 similarity cutoff (higher = stricter). Start with limit=10, threshold=0.7. If missing memories, lower threshold to 0.6 and raise limit to 20. If getting irrelevant results, raise threshold to 0.75-0.8.
Search & Retrieval
Q11Why are my search results irrelevant?
Usually one of: 1) namespaces are mixed, 2) too many memories stored (noise), 3) threshold too low, 4) content is too abstract ('likes things' vs specific facts). Fix: separate namespaces, make content specific, raise threshold.
Search & Retrieval
Q12Search quality dropped as memories grew. What do I do?
Add importance scoring and keep only high-importance memories. Use TTL to auto-expire old memories. Periodically merge similar memories into summaries. Separate namespaces to reduce search scope.
Search & Retrieval
Q13How should I use /api/extract?
Recommended flow: 1) Call with auto_store=false to preview extracted memories. 2) Show results to user for confirmation. 3) Store only approved memories. 4) Once extraction quality is proven, switch to auto_store=true for automation.
Extraction
Q14What's the difference between model=haiku and model=sonnet?
haiku: Faster, cheaper, good for most cases. sonnet: More accurate, better for important extractions (onboarding, contracts, policies). Use haiku for bulk/initial extraction, sonnet for high-stakes scenarios.
Extraction
Q15Can I modify a memory after storing it?
Yes, but the recommended pattern is: create new memory + delete/archive old one. This approach is better for audit trails and prevents regression issues.
Operations
Q16How do I delete memories?
Two approaches: 1) Delete by ID (precise), 2) Delete by namespace (bulk cleanup). For enterprise/compliance, ensure 'complete deletion + audit log'. Document your deletion policy clearly.
Operations
Q17I'm getting 429 Too Many Requests. What should I do?
Biggest cost drivers: 1) Extract frequency - reduce calls. 2) Search scope - use namespace to narrow. 3) Model choice - use haiku for routine, sonnet for important. 4) Caching - cache repeated queries. 5) Batch operations when possible.
Operations
Q19Can I use the API key in browser (frontend)?
Not recommended - high risk of key exposure. Call Seizn from your server (Next.js Route Handler, Cloudflare Worker, serverless function) and have the browser call your server. Never expose API keys to client-side code.
Security & Compliance
Q20What documentation does my security/legal team need?
They'll want: 1) Data scope (what's stored/not stored), 2) Encryption (at rest: AES-256, in transit: TLS), 3) Tenant isolation method, 4) Deletion/retention policy, 5) Audit log access, 6) Key rotation/expiration policy. Keep a Security & Governance page in your docs.
Security & Compliance
Q21What SDKs are available?
We provide official SDKs for TypeScript/JavaScript (npm install seizn), Python (pip install seizn), and REST API for any language. All SDKs include TypeScript types, automatic retries, and streaming support.
Getting Started
Q22What's the typical integration timeline?
Basic integration (store/search memories) takes 1-2 hours. Adding extraction takes another 2-4 hours. Full production deployment with governance policies typically takes 1-2 weeks including testing and security review.
Getting Started
Q23How do I handle multi-language content?
Seizn's embeddings support 100+ languages. Store content in any language and search will work cross-lingually. For best results, keep a 'language' metadata field and filter by it when you want language-specific results.
Search & Retrieval
Q24Can I combine semantic search with keyword filtering?
Yes. Use the 'filters' parameter to narrow by metadata (namespace, scope, memory_type, custom fields) before semantic ranking. This hybrid approach gives you both precision and recall.
Search & Retrieval
Q25How do I handle extraction failures?
Check the 'status' field in the response. If extraction fails: 1) Verify input format is correct, 2) Check if content is too short (<50 chars often fails), 3) Try a different model. Set up retry logic with exponential backoff for transient errors.
Extraction
Q26Can I customize what gets extracted?
Yes. Use the 'extraction_prompt' parameter to guide the model. You can specify: which memory types to extract, minimum importance threshold, specific fields to look for, and custom output schema.
Extraction
Q27How do I migrate from another memory/RAG system?
Export your existing data to JSON with fields: content, metadata, created_at. Use our bulk import API (POST /api/memories/import) with batches of 100-500 items. Run a parallel validation query set to ensure quality matches.
Operations
Q28What's the recommended backup strategy?
Use the export API (GET /api/memories/export) to create periodic backups. Schedule daily exports for critical namespaces. Store exports in your own object storage (S3, GCS) with versioning enabled.
Operations
Q29Is Seizn SOC 2 compliant?
Seizn is built on SOC 2 Type II compliant infrastructure (Supabase). Our own SOC 2 certification is in progress (ETA Q3 2026). EU region deployment is available for Enterprise customers by request. Contact our sales team for security questionnaires or compliance documentation.
Security & Compliance
Q30How do I implement user data deletion for GDPR/CCPA?
Use DELETE /api/memories with user_id filter to remove all memories for a user. Enable audit logging to prove deletion. Set retention policies to auto-delete after a period. Export user data with GET /api/memories/export?user_id=xxx for data portability requests.
Security & Compliance
Q31What is the Graph-RAG memory system?
Seizn uses a Graph-RAG architecture that combines vector embeddings with knowledge graph relationships. Memories are stored as nodes with typed edges (relates_to, supports, contradicts, etc.). This enables: contextual recall through graph traversal, automatic community detection for topic clustering, and temporal knowledge tracking with validity periods.
Core Concepts
Q32What is the memory tier system (Hot/Warm/Cold)?
Seizn automatically organizes memories into tiers based on usage: Hot tier (frequently accessed, kept in fast cache), Warm tier (occasionally accessed, standard retrieval), Cold tier (rarely used, archived but searchable). The tier manager automatically promotes/demotes memories based on access patterns, optimizing both performance and cost.
Core Concepts
Q33How does cross-lingual search work?
Seizn stores both original content and a canonical English representation for each memory. Search queries are matched against both embeddings, enabling scenarios like searching in English and finding Hindi memories. We use multilingual embedding models (BGE-M3, LaBSE) that support 100+ languages with high cross-lingual alignment.
Search & Retrieval
Q34What languages are supported for memory extraction?
Seizn supports 35+ languages for memory extraction and search, including: English, Chinese (Simplified/Traditional), Hindi, Spanish, French, German, Japanese, Korean, Russian, Ukrainian, Arabic, and all major Indic languages (Tamil, Telugu, Bengali, etc.). Language is auto-detected with 95%+ accuracy.
Search & Retrieval
Q35What is MCP integration and how do I use it?
MCP (Model Context Protocol) allows AI assistants like Claude Desktop to directly access Seizn memories. Install our MCP server (npx @seizn/mcp-server), add it to your Claude Desktop config, and the assistant can automatically save and retrieve memories during conversations. This enables persistent personalization across sessions.
Getting Started
Q36What technology powers Seizn's memory system?
Seizn is built on: Next.js 16 with React Server Components, Supabase (PostgreSQL + pgvector for vector search), Voyage AI voyage-3 model (1024-dim embeddings), Claude for memory extraction and analysis. We use HNSW indexes for sub-100ms vector search, WebSocket realtime subscriptions, and Redis for caching.
Getting Started
Q37How do I integrate Seizn with my LLM workflow?
Three main approaches: 1) REST API - direct HTTP calls for full control, 2) SDKs (TypeScript/Python) - type-safe wrappers with retry logic, 3) MCP Server - automatic integration with Claude Desktop and other MCP-compatible assistants. For production, we recommend the SDK with server-side calls to protect your API key.
Operations
Q38What is the Context API and how do I use it?
The Context API (GET /api/context) returns a pre-formatted string ready to inject into your LLM prompt. It combines: user profile, recent memories, relevant facts, and graph context. Use format=brief (~500 tokens), detailed (~1500 tokens), or extended (~3000 tokens) based on your context budget.
Operations
Q39What is the Seizn MCP Server and which editors does it support?
The Seizn MCP Server (seizn-mcp) is a Model Context Protocol server that lets AI coding assistants directly access your Seizn memories. It supports Claude Code, Cursor, Windsurf, Cline, and any MCP-compatible client natively. For editors without MCP support (GitHub Copilot, Aider, OpenAI Codex), we provide config file sync that generates CLAUDE.md, AGENTS.md, .cursorrules, .windsurfrules, .github/copilot-instructions.md, .clinerules, and CONVENTIONS.md from your memories.
MCP & Dev Tools
Q40How does multi-editor config sync work?
Seizn can export your memories as editor-specific configuration files. Use the sync_config_files MCP tool or POST /api/config-sync API. Supported formats: CLAUDE.md (Claude Code), AGENTS.md (OpenAI Codex), .cursor/rules (Cursor), .cursorrules (Cursor legacy), .windsurfrules (Windsurf), .github/copilot-instructions.md (GitHub Copilot), .clinerules (Cline), CONVENTIONS.md (Aider). Push syncs memories to files, pull imports files back to Seizn. This means your AI preferences follow you across every editor.
MCP & Dev Tools
Q41How do I authenticate the MCP server without copying API keys?
Seizn supports OAuth Device Flow (RFC 8628) for browser-based authentication. Run the auth_login MCP tool β it generates a human-readable code like 'ABCD-1234', opens your browser to seizn.com/auth/device, and you approve access. The token is saved to ~/.seizn/credentials.json automatically. No more copying API keys between terminals.
MCP & Dev Tools
Q42What MCP Resources are available?
Seizn MCP Server exposes read-only resources that AI assistants can access: seizn://memories/recent (last 10 memories), seizn://profile (user profile), seizn://graph/summary (knowledge graph stats), seizn://memories/project/{name} (project-filtered memories), seizn://context/{format} (pre-formatted context in brief/detailed/extended), and seizn://docs/setup/{editor} (editor-specific setup guides).
MCP & Dev Tools
Q43Can I get notified when memories change?
Yes. Use the Webhook tools (create_webhook, list_webhooks, delete_webhook, webhook_deliveries) in MCP or the REST API at /api/webhooks. Subscribe to events like memory.created, memory.updated, memory.deleted and receive HTTPS POST notifications. This enables real-time sync between Seizn and your custom integrations.
MCP & Dev Tools
Q44How does auto context loading work with projects?
When you call session_init with a cwd (current working directory), Seizn auto-detects your project from package.json, pyproject.toml, or Cargo.toml. It then loads project-specific memories and instructions automatically. This means your AI assistant always has the right context for the project you're working on β no manual configuration needed.
MCP & Dev Tools
Q45How do I set up Seizn MCP in my editor?
For Claude Code: add to ~/.claude/settings.json under mcpServers with command 'npx' and args ['-y', 'seizn-mcp@latest']. For Cursor: go to Settings > MCP Servers and add the same config. For Windsurf: Settings > MCP. For Cline: MCP settings. Set SEIZN_API_KEY in the env block, or use auth_login for browser-based OAuth. Full setup guides are available at seizn.com/docs/integrations.
MCP & Dev Tools
Still have questions?
Check out our documentation or get in touch with our support team.