In-context memory loads everything into the agent’s active context window. It is fast and simple, but token costs scale linearly and accuracy degrades as context grows. External memory (Mem0, Zep, vector stores) persists facts across sessions with 90% fewer tokens and 26% better accuracy. But neither solves a third problem: enterprise agents that need governed business definitions, lineage, and policies that were never in a conversation. This guide covers all three architectures, their real costs, and an honest decision framework.
Quick comparison table
Permalink to “Quick comparison table”| Dimension | In-Context Memory | External Memory (Mem0/Zep) | Structured Context Layer |
|---|---|---|---|
| What it is | Everything loaded into the active context window | Facts extracted from conversations, stored in vector DB | Pre-governed enterprise knowledge: definitions, lineage, policies |
| Cost model | ~$5,000/day at 10K interactions (GPT-4o, 100K tokens) | ~$333/day, a 90% token savings vs full-context | Query-based; not proportional to context size |
| Latency | Baseline | 91% lower p95 latency than full-context (Mem0) | Typically sub-millisecond for structured queries |
| Accuracy | Degrades mid-context (Liu et al., “lost in the middle”) | 26% better than full-context on LOCOMO benchmark | Not retrieval-dependent; query returns exact match |
| Freshness | Always current (session-scoped) | Depends on re-ingestion cadence; stale risk | Depends on governance pipeline refresh cycle |
| Enterprise fit | Prototypes, short tasks | User personalization, session continuity | Multi-agent, regulated industries, complex data estates |
| Cold-start | Immediate, no prior context needed | Requires prior conversations to extract knowledge | Immediate, knowledge predates any agent |
| Governance | None | None, no access controls or certification | Native: lineage, access controls, certified definitions |
What is in-context memory?
Permalink to “What is in-context memory?”In-context memory is everything the AI agent sees during a single inference call: the system prompt, conversation history, retrieved chunks, tool outputs, and few-shot examples loaded into the context window. It requires no retrieval step, the model has direct attention over all of it, but it is finite, session-scoped, and priced per token.
How it works
Permalink to “How it works”In-context memory is the agent’s active working RAM. Every token the model attends to during one inference call (system prompt, conversation history, retrieved document chunks, tool outputs) is in-context memory. There is no retrieval latency; it is fully accessible in zero hops. After the call ends, it is gone.
Current context window limits: GPT-4o (128K tokens), Claude Sonnet 4 (200K tokens, 1M in beta), GPT-5.4 (1.05M tokens with 2x pricing above 272K). Those numbers are large. They are not unlimited.
Token pricing is the pressure point. GPT-4o costs $5 per million input tokens. At 100K tokens per request, each call costs $0.50. Scale that to 10,000 agent interactions per day: $5,000 per day in context costs alone. Mem0’s published research (arXiv:2504.19413) puts the problem in concrete terms: full-context approaches consume approximately 26,031 tokens per conversation. At that rate, the cost is not hypothetical; it is the primary reason production teams move off naive context-stuffing strategies.
In-context memory started as the only option, before RAG and before memory frameworks. For short, self-contained tasks, it remains optimal. It breaks down as knowledge bases grow: thousands of policy documents, metric definitions, and entity relationships do not fit in any practical context window. Long-running multi-agent pipelines compound the problem; context accumulates across tool calls, and earlier instructions are effectively crowded out.
Practitioners in production have a name for this: context rot. As sessions grow longer, the model stops attending to instructions placed at the top of the system prompt. This is not a prompt engineering problem that better formatting fixes. It is a structural property of how transformer attention works, and it matters deeply for enterprise agents.
For a deeper treatment of where context windows break down, see Atlan’s guide to LLM context window limitations.
When in-context memory is the right choice
Permalink to “When in-context memory is the right choice”- Short, self-contained tasks: Single-turn requests where all relevant context fits within the prompt without competing for token budget.
- Small, static knowledge bases: Knowledge that fits comfortably in the context window without triggering mid-context degradation.
- Prototypes and demos: Speed of setup matters; cost optimization is not yet a requirement.
- No session persistence needed: The agent does not need to recall prior interactions, and each call is independent.
What is external memory?
Permalink to “What is external memory?”External memory extracts facts from conversations, stores them in a vector database or knowledge graph, and retrieves relevant chunks at inference time. Systems like Mem0 deliver 26% better accuracy, 91% lower latency, and 90% token savings versus full-context approaches, making them the practical choice for conversational agents that need session continuity without proportional token costs.
How it works
Permalink to “How it works”The pipeline has four steps: embed, store, retrieve, inject. Documents and conversation turns are chunked, embedded as vectors, stored in a vector database (Pinecone, Weaviate, Chroma), then retrieved via similarity search and injected into the context window at query time.
External memory covers three types of knowledge. Episodic memory captures conversation history, user preferences, and past interactions; this is the Mem0 and Zep pattern. Semantic memory holds factual knowledge extracted from documents. Procedural memory encodes tool-use patterns and workflow routines.
The performance data from Mem0 (arXiv:2504.19413) is concrete: 26% accuracy improvement over full-context approaches on the LOCOMO benchmark, 91% reduction in p95 latency, and 90% token savings. That translates to approximately 1,764 tokens per conversation versus 26,031 for full-context. At 10,000 daily interactions with GPT-4o pricing, the cost drops to roughly $333 per day versus $5,000 for full-context approaches. Zep takes a different approach, using a temporal knowledge graph. In practice, graph construction latency runs several hours; memory footprints exceed 600,000 tokens per conversation; and immediate post-ingestion retrieval often fails. The simpler extraction pipeline in Mem0 wins on production viability.
External memory has five failure modes worth understanding:
- Chunking mismatch — fixed 512-token chunks split semantic units, and a 0.87 cosine similarity score can still return wrong answers 40% of the time due to semantic-but-not-contextual relevance.
- Stale data — vector indexes not refreshed surface superseded policies with no staleness signal to the model.
- Semantic mismatch — the distribution gap between query phrasing and document phrasing degrades retrieval for exact-match requirements like product codes or contract clauses.
- Governance blindness — retrieval systems have no awareness of access policies, metric certification status, or the difference between a canonical definition and a draft.
- The cold-start problem — Mem0-style systems cannot answer “what does ‘net revenue’ mean at this company” because that knowledge was never in any conversation.
That fifth failure mode is where the agent context layer becomes relevant, and where the comparison between memory and knowledge infrastructure matters most. For more on how context engineering addresses these gaps at the system level, see Atlan’s guide to context engineering.
When external memory is the right choice
Permalink to “When external memory is the right choice”- Conversational agents with returning users: User preferences and prior interactions are the primary knowledge source; session continuity across days or weeks is required.
- Large unstructured knowledge bases: Documents, emails, and transcripts too large for the context window but primarily unstructured.
- Cost-sensitive deployments: 90% token reduction makes external memory compelling when conversation history is the core knowledge type.
- Personalization use cases: Customer support agents and personal assistants where user-specific memory is the differentiator.
The “lost in the middle” problem: why bigger context windows don’t solve it
Permalink to “The “lost in the middle” problem: why bigger context windows don’t solve it”Liu et al. (2024, TACL) found that language model performance peaks when relevant information appears at the beginning or end of the input context, and degrades significantly when it sits in the middle. This is not a prompt engineering problem. It is a structural property of transformer attention that persists even in explicitly long-context models.
What the research found
Permalink to “What the research found”The paper (Nelson F. Liu, Kevin Lin, John Hewitt et al., “Lost in the Middle: How Language Models Use Long Contexts,” TACL 12:157-173, 2024, arXiv:2307.03172) tested this across multiple model families. The finding held consistently: models attend most strongly to content at the beginning and end of context, and accuracy degrades for content positioned in the middle, even when models were explicitly fine-tuned for long-context use. This is structural, not incidental.
For enterprise agents, this has direct consequences. Knowledge bases injected as large context blocks are disproportionately affected. The most critical definitions, such as those that determine whether a revenue figure is GAAP or adjusted, may land mid-context. Practitioners observe this as context rot: as sessions grow, the agent appears to forget instructions that are nominally still present.
The intuitive response is to use a larger context window. GPT-5.4’s 1.05 million token window seems to make the problem go away. It does not. Two factors undermine the large-context solution. First, GPT-5.4 charges 2x pricing above 272K tokens, meaning large context windows are materially more expensive at scale. Second, research from 2024 and 2025 (arXiv:2407.16833; arXiv:2501.01880) confirms the nuanced tradeoff: long context consistently outperforms RAG on accuracy benchmarks when sufficiently resourced, but the cost advantage of retrieval and the “lost in the middle” degradation make RAG the practical choice for most production workloads.
The architectural implication is precise. If accuracy at enterprise scale matters, context must be managed, either via selective retrieval (external memory) or via structured query (the context layer). “Stuff it all in” breaks at the scale and budget constraints of production enterprise systems.
For more on the limits of context windows as a solution, see Atlan’s guide to LLM context window limitations.
The structured context layer: the third option for enterprises
Permalink to “The structured context layer: the third option for enterprises”A structured context layer is pre-governed enterprise knowledge (metric definitions, data lineage, ontology, and business policies) that agents query directly at inference time via MCP or API. Unlike in-context stuffing or conversation-memory extraction, it does not require prior sessions to exist. The knowledge was never conversational; it was always infrastructure.
What it is and how it works
Permalink to “What it is and how it works”The structured context layer is not assembled from conversation history. It is maintained as governed enterprise knowledge that agents query when they need it.
Five components make up the five-component agent context layer:
- Semantic layer: Governed metric definitions, such as what “revenue” means at this company, certified and versioned.
- Ontology and identity: Cross-system entity resolution, mapping the same customer across CRM, ERP, and data warehouse.
- Operational playbooks: Routing rules, escalation logic, and decision procedures agents follow.
- Data lineage: Provenance, including where a number came from and whether the pipeline was certified.
- Active metadata: Decision memory, certification status, and ownership tracking.
Agents query this layer via MCP, SQL, or API, not via embedding similarity. The question “what is net revenue?” does not find the closest vector to that phrase; it queries a governed glossary entry. That distinction eliminates the retrieval accuracy problem entirely for structured knowledge.
This is architecturally distinct from both other approaches. Memory asks: what did the user say before? In-context injection asks: how much can I fit in the prompt? The structured context layer asks: what does this organization know about this topic, in a form an agent can reason over?
The difference between memory and knowledge matters here. Memory captures what happened: episodic, session-scoped. Knowledge encodes what is true: governed, persistent, predating any agent. Enterprises need both. But they are architecturally distinct, and conflating them produces agents that are either expensive, inaccurate, or both.
Evidence for the impact of structured context is concrete. Snowflake’s internal ontology experiment showed that adding a knowledge layer improved agent answer accuracy by 20% and reduced tool calls by 39%, not from better retrieval, but from structured knowledge infrastructure. Workday VP Joe DosSantos described the failure mode precisely: “We built a revenue analysis agent and it couldn’t answer one question. We started to realize we were missing this translation layer.” That is not a memory problem. It is a governed knowledge infrastructure problem. Hallucination rates across LLMs without proper context grounding run 50 to 82% (PMC research); Workday achieved a 5x accuracy improvement with structured context layers. Gartner forecasts that 40% of agentic AI projects will be canceled by 2027 due to missing structured context, and predicts context engineering will appear in 80% of AI tools by 2028, improving agent accuracy by 30% or more.
To understand how this layer connects to a context graph as its underlying data structure, see the full treatment of agent context layer architecture.
When to consider a structured context layer
Permalink to “When to consider a structured context layer”- Multiple agents need consistent definitions: If Agent A and Agent B each have their own memory stores, they can “remember” different definitions of the same metric. A structured context layer eliminates that inconsistency at the source.
- Knowledge predates the agent: Business glossaries, lineage maps, and policies exist in data catalogs long before any AI agent is deployed. Extraction is unnecessary.
- Regulated industries: Compliance, auditability, and access control are requirements that finance and healthcare teams cannot waive.
- Multi-system entity resolution: “Customer” means different things in CRM, ERP, and the data warehouse. An ontology layer resolves this before the agent sees the query.
In-context vs external memory vs structured context: detailed comparison
Permalink to “In-context vs external memory vs structured context: detailed comparison”The sharpest differences appear in three dimensions: cost model (token proportionality vs query-based), knowledge source (session-scoped vs conversation-extracted vs pre-governed), and governance posture (none vs none vs native). For short tasks, in-context wins on simplicity. For conversational agents, external memory wins on cost. For enterprise-scale multi-agent systems, structured context wins on accuracy and auditability.
Detailed comparison across 10 dimensions
Permalink to “Detailed comparison across 10 dimensions”| Dimension | In-Context Memory | External Memory (Mem0/Zep) | Structured Context Layer |
|---|---|---|---|
| Primary mechanism | Token injection: everything in the active context window | Embed, store, retrieve, inject | Direct query against governed knowledge graph |
| Knowledge source | Session content, system prompt, retrieved chunks | Prior conversation history, extracted facts | Pre-governed definitions, lineage, policies |
| Token cost at scale | ~26,031 tokens per conversation; ~$5,000/day at 10K interactions | ~1,764 tokens per conversation; ~$333/day at 10K interactions | Query-based; not proportional to context size |
| Accuracy failure mode | “Lost in the middle”: degrades at long contexts (Liu et al.) | Chunking mismatch, semantic gap, stale index | Governance pipeline lag; outdated definitions if uncertified |
| Cold-start behavior | Immediate, no prior context required | Fails without prior conversations | Immediate, knowledge predates any agent |
| Session persistence | None, cleared after every call | Persistent across sessions | Persistent, continuously maintained |
| Governance and auditability | None | None | Native: lineage, access controls, certification status |
| Multi-agent consistency | Per-agent; no shared source of truth | Per-agent memory store; inconsistency risk | Shared; all agents query the same governed layer |
| Setup complexity | Low, prompt engineering only | Medium, embedding pipeline and vector DB operations | High, requires governance investment upfront |
| Best fit | Short tasks, prototypes, demos | Conversational agents, personalization | Enterprise multi-agent, regulated industries |
Real-world example: revenue attribution at a financial services firm
Permalink to “Real-world example: revenue attribution at a financial services firm”Three agents operate at a financial services firm: a reporting agent that queries Snowflake, a compliance agent that checks regulatory thresholds, and a customer-facing agent that summarizes account performance.
With in-context memory, each agent’s system prompt encodes a different version of “net revenue,” injected by a different team at setup time. With external memory, each agent “remembers” the last conversation it had about revenue. Three agents, three conversations, three answers. With a structured context layer, all three agents query the same certified glossary entry: net_revenue, the GAAP definition certified by the CFO office, last updated 2026-01-15. One answer, auditable, consistent.
Decision framework: which architecture fits your scenario
Permalink to “Decision framework: which architecture fits your scenario”No single architecture wins across all scenarios. The decision turns on three variables: knowledge type (session-generated vs pre-existing), scale (single agent vs enterprise fleet), and governance requirements (none vs regulated). Most production enterprise systems combine all three: in-context for immediate task instructions, external memory for user history, and a structured context layer as the substrate that makes both reliable.
Routing matrix
Permalink to “Routing matrix”Choose in-context memory when:
- The task is single-turn and self-contained, with all relevant information in the user’s request.
- The knowledge base is small, static, and fits within the context window without mid-context degradation.
- You are prototyping and speed of setup matters more than cost or accuracy at scale.
- The agent runs infrequently and daily cost impact is negligible.
Choose external memory (Mem0/Zep/vector store) when:
- The agent operates in conversational settings where user preferences and prior interactions define the value.
- Session continuity across days or weeks is a product requirement, such as for personal assistants or customer support.
- The knowledge base is large but primarily unstructured, consisting of documents, emails, and transcripts.
- User-specific personalization is the core use case, not shared enterprise knowledge.
Choose a structured context layer when:
- Multiple agents need access to the same definitions, policies, and relationships, and inconsistency is a business risk.
- Enterprise data spans multiple systems with different entity identifiers, as in the “customer” problem at scale.
- Compliance, auditability, or access control are non-negotiable, as in financial services and healthcare.
- The knowledge agents need was never in a conversation; it lives in data catalogs, business glossaries, and governance systems.
- Gartner’s 40% cancellation forecast is a real risk: agents failing because they lack structured, governed context, not just session memory.
The hybrid reality:
Production enterprise systems typically use all three. The question is not which to use in isolation; it is where the governing logic lives. The structured context layer is the substrate that makes in-context and external memory reliable: it certifies what facts mean before they are retrieved or injected.
For more on why 40% of agentic AI projects fail without structured context, and on whether enterprises need a context layer between data and AI, the Atlan context layer page covers both in detail.
How Atlan approaches the in-context vs external memory question
Permalink to “How Atlan approaches the in-context vs external memory question”Atlan’s position: the in-context vs external memory debate is framed around the wrong problem. Both architectures treat context as something assembled or extracted at runtime. The real gap is governed enterprise knowledge (definitions, lineage, and policies) that exists as organizational infrastructure and that agents need access to without any conversational history.
The challenge
Permalink to “The challenge”When enterprise teams build agents with in-context stuffing or vector retrieval, they hit a specific failure: the agent cannot correctly interpret the organization’s own data. Joe DosSantos, VP Enterprise Data and Analytics at Workday, described it precisely: “We built a revenue analysis agent and it couldn’t answer one question. We started to realize we were missing this translation layer.”
Neither in-context memory nor external memory solves that problem. In-context memory cannot hold the full organizational knowledge graph in a single call. External memory cannot extract “what does ‘net revenue’ mean at this company” from conversations if no one has ever discussed it with the agent. The agent does not know what “revenue” means at this company, what “customer” resolves to across five source systems, or which lineage path is certified. That is not a memory problem; it is a governed knowledge infrastructure problem.
Atlan’s approach
Permalink to “Atlan’s approach”Atlan maintains a continuously governed context layer: 18 million-plus cataloged assets, 1,300-plus business glossary terms (at CME Group scale), certified metric definitions, cross-system lineage, and access-controlled policies. Agents query this layer via MCP, the protocol that reached 97 million monthly downloads by March 2026 with adoption from Anthropic, OpenAI, Microsoft, and AWS.
The enterprise context layer that Atlan provides is not an alternative to conversation memory. It is the substrate that governs what memory retrieves and what in-context injection means. Active metadata, which updates as data changes rather than only when someone edits a prompt, means the context layer stays current without requiring manual prompt maintenance.
Gartner predicts context engineering will appear in 80% of AI tools by 2028, improving agent accuracy by 30% or more. Atlan’s view: the teams that win are the ones treating context as infrastructure, not as a runtime assembly problem.
Customer evidence
Permalink to “Customer evidence”CME Group cataloged 18 million assets and 1,300-plus business glossary terms in year one. That volume of structured knowledge pre-exists any conversation; no extraction is required. Workday achieved a 5x accuracy improvement with structured context grounding versus a hallucination-prone baseline (PMC research shows 50 to 82% hallucination rates across LLMs without proper grounding). Snowflake’s internal ontology experiment (20% accuracy improvement, 39% fewer tool calls from adding a knowledge layer) demonstrates the same architectural shift that Atlan’s context layer enables via its governed metadata graph.
See how CME Group achieved context at speed with Atlan’s enterprise context layer.
To understand how to build a memory layer for AI agents that includes the structured knowledge substrate, and how the types of AI agent memory map to these three architectural options, see Atlan’s guide on how to implement an enterprise context layer for AI and the agent context layer guide for implementation-level detail.
Wrapping up
Permalink to “Wrapping up”The in-context vs external memory debate assumes the only question is how to deliver context, by injection or retrieval. It misses a prior question: where does authoritative enterprise knowledge live?
In-context memory and external memory both treat knowledge as something you assemble at runtime. Neither can answer “what does ‘revenue’ mean at this company” if that answer was never in a conversation and never fits cleanly in a prompt. External memory solves the token cost and session continuity problem. In-context memory solves the simplicity and latency problem. The structured context layer solves the governed knowledge problem.
Most enterprise teams will need all three. The context layer is the substrate that determines whether the other two produce answers you can trust. For the full architecture of how these pieces connect, see Atlan’s agent context layer guide and the Atlan context layer product page.
FAQs about in-context memory vs external memory for AI agents
Permalink to “FAQs about in-context memory vs external memory for AI agents”1. What is the difference between in-context memory and external memory for AI agents?
Permalink to “1. What is the difference between in-context memory and external memory for AI agents?”In-context memory is everything loaded into the agent’s active context window during a single inference call, including the system prompt, conversation history, and retrieved chunks. External memory stores facts extracted from prior conversations in a vector database and retrieves relevant chunks at query time. In-context memory is session-scoped and clears after each call; external memory persists across sessions. The tradeoff is simplicity vs scalability: in-context requires no infrastructure, while external memory requires embedding pipelines and retrieval management.
2. Can a large context window replace external memory for AI agents?
Permalink to “2. Can a large context window replace external memory for AI agents?”Not reliably. Liu et al. (TACL, 2024) found that model accuracy degrades significantly when relevant information sits in the middle of long contexts, even in models explicitly designed for long-context use. This “lost in the middle” effect means stuffing a 200K or 1M context window does not guarantee the model attends to all of it. External memory’s selective retrieval addresses this by injecting only relevant context, not the full history.
3. What is the “lost in the middle” problem in LLMs?
Permalink to “3. What is the “lost in the middle” problem in LLMs?”“Lost in the middle” is a finding from Liu et al. (arXiv:2307.03172, TACL 2024): language models perform best when relevant information appears at the beginning or end of the input context, and accuracy degrades when that information sits in the middle. The cause is structural, as transformer attention weights information at context boundaries more heavily than mid-context positions. It affects even models with explicit long-context training, making large context windows an unreliable solution for knowledge-dense enterprise workloads.
4. What are the cost tradeoffs between in-context and external memory approaches?
Permalink to “4. What are the cost tradeoffs between in-context and external memory approaches?”Mem0’s published research (arXiv:2504.19413) quantifies the gap: full-context approaches use approximately 26,031 tokens per conversation; memory-managed approaches use approximately 1,764 tokens, a 15x difference. At GPT-4o pricing ($5 per million input tokens) and 10,000 daily interactions, full-context costs roughly $5,000 per day versus $333 per day for memory-managed approaches. External memory delivers a 90% token reduction and 91% lower p95 latency. The tradeoff: external memory requires embedding infrastructure and introduces retrieval accuracy risk.
5. What is a structured context layer and how does it differ from a memory layer?
Permalink to “5. What is a structured context layer and how does it differ from a memory layer?”A structured context layer is pre-governed enterprise knowledge (metric definitions, data lineage, ontology, and policies) that agents query directly at inference time. A memory layer (Mem0, Zep) extracts and stores facts from prior conversations. The distinction is architectural: memory captures what happened; structured context encodes what is true. A structured context layer can answer “what does ‘net revenue’ mean at this company” even if no agent has ever discussed it, because that knowledge was always organizational infrastructure, not conversational history.
6. How do enterprise AI agents handle memory across multiple agents?
Permalink to “6. How do enterprise AI agents handle memory across multiple agents?”Per-agent memory stores (Mem0, Zep, custom vector databases) create a consistency risk: two agents can “remember” different definitions of the same metric if their conversation histories diverged. Enterprise-scale multi-agent architectures address this with a shared context substrate, a governed knowledge layer all agents query. Without it, multi-agent systems produce inconsistent answers to identical questions depending on which agent handles the request. Gartner forecasts that 40% of agentic AI projects will be canceled by 2027 because agents lack this structured context.
7. When should I use RAG instead of stuffing context into the prompt?
Permalink to “7. When should I use RAG instead of stuffing context into the prompt?”RAG is preferable when your knowledge base is too large for the context window, when cost matters at scale, or when the “lost in the middle” effect would degrade accuracy with full-context injection. Prompt stuffing remains valid for small, static knowledge bases in short single-turn tasks. Research (arXiv:2407.16833) confirms that long-context approaches outperform RAG on accuracy benchmarks when sufficiently resourced, but RAG’s cost advantage and selective retrieval make it the practical choice for most production workloads.
8. What is the best memory architecture for multi-agent enterprise systems?
Permalink to “8. What is the best memory architecture for multi-agent enterprise systems?”Most production enterprise systems use all three memory types in combination: in-context memory for immediate task instructions and retrieved chunks, external memory for user session history and personalization, and a structured context layer for shared enterprise knowledge including definitions, lineage, and policies. The shared context layer is the most important component for multi-agent consistency: without it, agents operating from separate memory stores will produce different answers to identical questions about the same data. The structured layer governs what the other two retrieve.
External citations
Permalink to “External citations”-
Liu, N.F., Lin, K., Hewitt, J. et al. “Lost in the Middle: How Language Models Use Long Contexts.” Transactions of the Association for Computational Linguistics, 12:157-173, 2024. https://arxiv.org/abs/2307.03172
-
“Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory.” arXiv:2504.19413, 2026. https://arxiv.org/abs/2504.19413
-
“Retrieval Augmented Generation or Long-Context LLMs? A Comprehensive Study and Hybrid Approach.” arXiv:2407.16833, 2024. https://arxiv.org/abs/2407.16833
-
“Long Context vs. RAG for LLMs: An Evaluation and Revisits.” arXiv:2501.01880, January 2025. https://arxiv.org/abs/2501.01880
-
Snowflake Blog: “Agent Context Layer: Trustworthy Data Agents.” https://www.snowflake.com/en/blog/agent-context-layer-trustworthy-data-agents/
-
Gartner: “Context Engineering.” https://www.gartner.com/en/articles/context-engineering
-
Anthropic: “Effective Context Engineering for AI Agents.” https://www.anthropic.com/engineering/effective-context-engineering-for-ai-agents
Share this article
