The universal memory API for the AI era

Stop building retrieval from scratch. Personalise LLMs for your users. Built for developers who ship. Start building DOCS

Context

is everything

Without it, even the smartest AI is just an expensive chatbot

$ init vector_database
Way too expensive. Time to switch.
Painfully slow. Let's try another.
Won't scale. Back to square one.
Maintenance nightmare. Need alternatives.
$ choose embedding_model
Which model fits your use case?
Confusing performance tradeoffs
Can't keep up with new releases
$ handle format_parsing
Markdown: Tables break everything
HTML: Scripts and styles interfere
PDF: Layout ruins extraction
Word docs: Unpredictable formatting
$ calculate scaling_costs
Costs explode at production scale
Performance degrades as data grows
Engineering hours pile up fast
$ setup connection_sync
Sync failures between data sources
API rate limits during large syncs
Images: Need vision models now?
Audio/Video: Transcription costs soar
$ init multimodal_support
Websites: JS & rate limits are messy
PDFs: OCR fails, extraction inconsistent
Authentication tokens expire constantly
$ init vector_database
Way too expensive. Time to switch.
Painfully slow. Let's try another.
Won't scale. Back to square one.
Maintenance nightmare. Need alternatives.
$ choose embedding_model
Which model fits your use case?
Confusing performance tradeoffs
Can't keep up with new releases
$ handle format_parsing
Markdown: Tables break everything
HTML: Scripts and styles interfere
PDF: Layout ruins extraction
Word docs: Unpredictable formatting
$ calculate scaling_costs
Costs explode at production scale
Performance degrades as data grows
Engineering hours pile up fast
$ setup connection_sync
Sync failures between data sources
API rate limits during large syncs
Images: Need vision models now?
Audio/Video: Transcription costs soar
$ init multimodal_support
Websites: JS & rate limits are messy
PDFs: OCR fails, extraction inconsistent
Authentication tokens expire constantly
FEATURES  •  FEATURES  •  FEATURES

Unlock the Full Potential of Your Data

const response = await fetch('https://api.supermemory.ai/v3/memories', {
  method: 'POST',
  headers: {
    'Authorization': 'Bearer sm_ywdhjSbiDLkLIjjVotSegR_rsq3ZZKNRJmVr12p4ItTcf'
  },
  body: JSON.stringify({
    content: 'My name is Shreyans.',
      // or https://example.com 
      // or https://example.com/page.pdf
    metadata: {
      user_id: '123'
    }
  }),
})

const data = await response.json()
const response = await fetch('https://api.supermemory.ai/v3/memories', {
  method: 'GET',
  headers: {
    'Authorization': 'Bearer sm_ywdhjSbiDLkLIjjVotSegR_rsq3ZZKNRJmVr12p4ItTcf',
  },
  body: JSON.stringify({
    q: "What's my name?"
  })
})

const data = await response.json()
const response = await fetch('https://api.supermemory.ai/v3/connections/onedrive', {
  method: 'POST',
  headers: {
    'Authorization': 'Bearer sm_ywdhjSbiDLkLIjjVotSegR_rsq3ZZKNRJmVr12p4ItTcf',
  }
});

const data = await response.json();
solution  •  solution  •  solution

We’ve seen what it’s like to build memory infrastructure the hard way — so we built supermemory to make it disappear.

product  •  product  •  product

Unlimited context API. One simple 
switch.

Just add api.supermemory.ai/v3 to your OpenAI base URL — and get automatic long-term context across conversations.

import OpenAI from "openai"

const
client = new OpenAI({  
  baseUrl: "https://api.supermemory.ai/v3/https://api.openai.com/v1/"
})
Start building DOCS
FEATURES  •  FEATURES  •  FEATURES

Build the memory layer your product deserves

01.
Your data grows. Supermemory keeps up

Enterprise-Grade Performance at Any Scale

Supermemory is built to handle billions of data points with low-latency retrieval — whether you're indexing documents, video, or structured product data.

02.
No heavy lifting. Just smart, connected infrastructure

Seamless Integration 
Across Teams & Tools

Connect directly to your existing stack — from Notion to Google Drive to custom CRMs — with flexible APIs and SDKs that let every team tap into memory instantly.

03.
Own your data. Maintain compliance. Stay in control

Secure by Design. 

Fully Controllable.

Deploy Supermemory in the cloud, on-prem, or directly on-device — with full control over where and how your data is stored.

FEATURES  •  FEATURES  •  FEATURES

It just clicks with your stack

Interoperability

Model-agnostic APIs

Supermemory works with any LLM provider. So you can keep switching, without lock-in.

Performance

Sub-400ms latency 
at scale

Supermemory is built for speed and scale. We re-imagined RAG to be faster and more efficient.

EFFICIENCY

Best in class
performance

Supermemory delivers stronger precision and recall at every benchmark. And it’s ridiculously easy to start.

comparison of evaluation metrics: supermemory vs major memory provider
Tooling

Works with AI SDK, Langchain, and more

Supermemory works with any LLM provider. So you can keep switching, without lock-in.

Language Agnostic

Deploy in a day, 
not months

SDKs available for Python and Javascript.

Learn more

product  •  product  •  product

Add context to your agentic apps with few lines of code

Supermemory provides SDKs to make integration as simple as possible

npm install 'supermemory'
pip install 'supermemory'
Start building DOCS

Trusted by Open Source,
enterprise, and more than

35,000

of you

#1 Product of the day at Product hunt

Starred by over 9,000 users on Github

Flow useS supermemory to build the cursor for writing

Medtech Vendors uses supermemory to search through 500k vendors

Mixus uses Supermemory to power co-intelligence Agentic platform