hero-grid
hero-gradient

The universal memory API for the AI era

Get supermemory TM

Context engine for your app. Personalise LLMs for your users.

Built for developers who ship. Start buildingDOCS

hero-grid gradient-icon

Context
is everything

Without it, even the smartest AI is just an expensive chatbot

memory-engineering
  • $ init multimodal support
    • error
      Websites: JS & rate limits are messy
    • error
      PDFs: OCR fails, extraction inconsistent
    • error
      Authentication tokens expire constantly
  • $ init vector_database
    • error
      Way too expensive. Time to switch.
    • error
      Painfully slow. Let's try another.
    • error
      Won't scale. Back to square one.
    • error
      Maintenance nightmare. Need alternatives.
  • $ choose embedding_model
    • warning
      Which model fits your use case?
    • warning
      Confusing performance tradeoffs
    • warning
      Can't keep up with new releases
  • $ handle format_parsing
    • warning
      Markdown: Tables break everything
    • warning
      HTML: Scripts and styles interfere
    • warning
      PDF: Layout ruins extraction
  • $ calculate scaling_costs
    • error
      Costs explode at production scale
    • error
      Performance degrades as data grows
    • error
      Engineering hours pile up fast
  • $ setup connection_sync
    • error
      Sync failures between data sources
    • error
      API rate limits during large syncs
    • error
      Images: Need vision models now?
    • error
      Audio/Video: Transcription costs soar
  • $ init multimodal support
    • error
      Websites: JS & rate limits are messy
    • error
      PDFs: OCR fails, extraction inconsistent
    • error
      Authentication tokens expire constantly
  • $ init vector_database
    • error
      Way too expensive. Time to switch.
    • error
      Painfully slow. Let's try another.
    • error
      Won't scale. Back to square one.
    • error
      Maintenance nightmare. Need alternatives.
  • $ choose embedding_model
    • warning
      Which model fits your use case?
    • warning
      Confusing performance tradeoffs
    • warning
      Can't keep up with new releases
  • $ handle format_parsing
    • warning
      Markdown: Tables break everything
    • warning
      HTML: Scripts and styles interfere
    • warning
      PDF: Layout ruins extraction
  • $ calculate scaling_costs
    • error
      Costs explode at production scale
    • error
      Performance degrades as data grows
    • error
      Engineering hours pile up fast
  • $ setup connection_sync
    • error
      Sync failures between data sources
    • error
      API rate limits during large syncs
    • error
      Images: Need vision models now?
    • error
      Audio/Video: Transcription costs soar
solution • solution • solution

We've seen what it's like to build memory infrastructure the hard way — so we built supermemory to make it disappear.

product • product • product

Edit one line.
Get longer threads,
cost savings
memory.

Just add api.supermemory.ai/v3 to your OpenAI base URL — and get automatic long-term context across conversations.

import OpenAI from "openai"
const client = new OpenAI({
   baseUrl: "https://api.supermemory.ai/v3/https://api.openai.com/v1/"
})
Start building DOCS
features • features • features

Unlock the Full Potential of Your Data

const response = await fetch('https://api.supermemory.ai/v3/memories', {
  method: 'POST',
  headers: {
    'Authorization': 'Bearer sm_ywdhjSbiDLkLIjjVotSegR_rsq3ZZKNRJmVr12p4ItTcf'
  },
  body: JSON.stringify({
    content: 'You can add text',
      // or a url https://example.com 
      // or pdfs, videos, images. https://example.com/page.pdf
    metadata: {
      user_id: '123'
    }
  }),
})

const data = await response.json();
features • features • features

Build the memory layer your product deserves

01.
Your data grows. Supermemory keeps up

Enterprise-Grade Performance at Any Scale

Supermemory is built to handle billions of data points with low-latency retrieval — whether you're indexing documents, video, or structured product data.

02.
No heavy lifting. Just smart, connected infrastructure

Seamless Integration Across Teams & Tools

Connect directly to your existing stack — from Notion to Google Drive to custom CRMs — with flexible APIs and SDKs that let every team tap into memory instantly.

03.
Own your data. Maintain compliance. Stay in control

Secure by Design.
Fully Controllable.

Deploy Supermemory in the cloud, on-prem, or directly on-device — with full control over where and how your data is stored.

features • features • features

It just clicks with your stack

Model-interoperable APIs
Interoperability

Model-interoperable APIs

Supermemory works with any LLM provider. So you can keep switching, without lock-in. Switch between models. keep your memory.

Sub-400ms latency at scale
Performance

Sub-400ms latency at scale

Supermemory is built for speed and scale. We re-imagined RAG to be faster and more efficient.

Best in class performance
Efficiency

Best in class performance

Supermemory delivers stronger precision and recall at every benchmark. And it's ridiculously easy to start.

comparison of evaluation metrics:
supermemory vs major memory provider
comparison of evaluation metrics:<br> <span class="text-gradient">supermemory</span> vs major memory provider
Works with AI SDK, Langchain, and more
Tooling

Works with AI SDK, Langchain, and more

Supermemory works with any LLM provider. So you can keep switching, without lock-in.

Deploy in a day, not months Python, Typescript, Next.js, and React logos
Language Agnostic

Deploy in a day, not months

SDKs available for Python and Javascript.

Learn more

product • product • product

Add context to your agentic apps with few lines of code

Supermemory provides SDKs to make integration as simple as possible

Start building DOCS
hero-grid

Trusted by Open Source, enterprise, and more than

testimonial

of you

Product Hunt
#1 Product of the day at Product hunt
Github
Starred by over 9,000 users on Github
Flow
Flow uses supermemory to build the cursor for writing
Medtech Vendors
Medtech Vendors uses supermemory to search through 500k vendors
Mixus
Mixus uses Supermemory to power co-intelligence Agentic platform