Beyond Context Windows: Building Cognitive Memory for AI Agents

Abstract

AI agents are rapidly changing how users interact with software, yet most agentic systems today operate with little to no intelligent memory, relying instead on brittle context-window heuristics or short-term state. This limitation fundamentally constraints personalization, reasoning, and long-term adaptation.

To build agents that function as long-term personalized assistants, the memory problem must be addressed head-on.

In this talk, we introduce LinkedIn’s Cognitive Memory Agent (CMA), a horizontal memory platform designed to power stateful, context-aware, and personalized AI agents at scale. We present the architecture and design principles behind CMA, and how it enables agents to accumulate, reason over, and act upon long-term user knowledge. CMA tackles the memory challenge through three core components.

  • An ingestion layer that determines how to interpret unstructured inputs, what information to extract, and when and how to store it
  • A layered memory system comprising semantic memory (structured knowledge), episodic memory (time-indexed events), working memory (in-session context), and procedural memory (reasoning traces and plans).
  • A retrieval orchestration layer that disambiguates user intent, dynamically retrieves relevant memories across layers, and synthesizes responses.

Reinforcement learning continuously optimizes both ingestion and retrieval for quality, efficiency, and adaptability.

Together, these components enable agents to move beyond simple recall of prior interactions. CMA continuously ingests signals from user behavior and environment, infers latent preferences, and applies reasoning to deliver proactive, personalized assistance.

We conclude with insights from deploying CMA in large-scale products such as the LinkedIn Hiring Assistant, and explore how these experiences point toward a new generation of intelligent AI agents.


Speaker

Karthik Ramgopal

Distinguished Engineer & Tech Lead of the Product Engineering Team @LinkedIn, 15+ Years of Experience in Full-Stack Software Development

Distinguished Engineer and the Uber technical lead for the Product Engineering team at LinkedIn, leading ~5000 engineers responsible for developing and operating all of LinkedIn’s member and customer-facing products. 15+ years of experience in full-stack software development, design, and architecture, across product and infrastructure engineering teams. Role entails strategically shaping LinkedIn's technical vision across a wide range of domains with hands-on contributions to code and design on specific projects, with a primary focus on Generative AI applications (products and internal productivity use cases) in the recent past.

 

Read more
Find Karthik Ramgopal at:

From the same track

Session AI/LLM

Reliable Retrieval for Production AI Systems

Tuesday Mar 17 / 10:35AM GMT

Search is central to many AI systems. Everyone is building RAG and agents right now, but few are building reliable retrieval systems.

Speaker image - Lan Chu

Lan Chu

AI Tech Lead and Senior Data Scientist

Session AI/ML

Refreshing Stale Code Intelligence

Tuesday Mar 17 / 01:35PM GMT

Coding models are helping software developers move faster than ever, but weirdly, the models themselves are not keeping up. They are trained on months-old snapshots of open source code. They have never seen your internal codebase, let alone the code you wrote yesterday.

Speaker image - Jeff Smith

Jeff Smith

CEO & Co-Founder @ Neoteny AI, AI Engineer, Researcher, Author, Ex-Meta/FAIR

Session AI

Rewriting All of Spotify's Code Base, All the Time

Tuesday Mar 17 / 11:45AM GMT

We don't need LLMs to write new code. We need them to clean up the mess we already made.In mature organizations, we have to maintain and migrate the existing codebase. Engineers are constantly balancing new feature development with endless software upkeep.

Speaker image - Jo  Kelly-Fenton

Jo Kelly-Fenton

Engineer @Spotify

Speaker image - Aleksandar Mitic

Aleksandar Mitic

Software Engineer @Spotify

Session applied ai

Building an AI Gateway Without Frameworks: One Platform, Many Agents

Tuesday Mar 17 / 03:55PM GMT

Early AI integrations often start small: wrap an inference API, add a prompt, ship a feature. At Zoox, that approach grew into Cortex, a production AI gateway supporting multiple model providers, multiple modalities, and agentic workflows with dozens of tools, serving over 100 internal clients.

Speaker image - Amit Navindgi

Amit Navindgi

Staff Software Engineer @Zoox

Session

Sync Agents in Production: Failure Modes and Fixes

Tuesday Mar 17 / 05:05PM GMT

As models improve, we are starting to build long-running, asynchronous agents such as deep research agents and browser agents that can execute multi-step workflows autonomously. These systems unlock new use cases, but they fail in ways that short-lived agents do not.

Speaker image - Meryem Arik

Meryem Arik

Co-founder and CEO @Doubleword (previously TitanML)