Zum Inhalt springen

How an AI Agent Grew Up

Watch on YouTube →

The Problem: Statelessness as a Design Decision

Current language models are stateless. This is not a bug but an architectural decision: every request is processed independently of all previous ones. The model has no access to the results of past interactions, no representation of the conversation history beyond the current context window, and no concept of what it did yesterday.

In computer science, this is called “stateless.” In cognitive science, the more fitting term would be anterograde amnesia: the ability to form new memories is absent. The system can brilliantly analyze what it currently sees, but it cannot draw on what it saw yesterday.

The consequence is more severe than it appears at first glance. It is not just about lost context. It is about the impossibility of accumulation. Knowledge builds on knowledge. An insight from Monday changes Tuesday’s interpretation. A decision from last week constrains this week’s solution space. Without memory there is no knowledge accumulation, and without knowledge accumulation there is no learning curve.

Why Larger Context Windows Don’t Solve the Problem

The industry’s obvious answer is: larger context windows. If the system can process 200,000 tokens at once, isn’t that enough to include everything relevant?

No. For three reasons:

The scaling problem. Context windows grow linearly. Knowledge bases grow exponentially. After three months of working with an AI agent, the relevant context is no longer kilobytes but megabytes. No context window is large enough, and none ever will be.

The selection problem. Even if everything could fit in the context: what of it is currently relevant? A system that ingests 200,000 tokens undifferentiated doesn’t have more knowledge; it has more noise. The ability to extract what is currently relevant from a large knowledge base is an independent cognitive achievement that the language model does not perform, because it has no model of the current situation.

The structure problem. Raw text in a context window has no structure. There is no distinction between verified facts and guesses, between current and outdated information, between primary sources and derivations. Everything is flat text, and the model treats it accordingly: as equal. This is epistemologically untenable.

The Thesis: Knowledge Requires Architecture

The solution lies not in the model, but around the model. The 6-layer knowledge architecture separates what belongs together but must not be mixed:

Immediately retrievable facts (FactsDB) are not the same as distilled insights (BrainDB). Distilled insights are not the same as searchable original documents (Qualia). And none of these layers is complete without a validation layer that distinguishes between reliable and unreliable.

This separation is not technical self-indulgence. It reflects a fundamental epistemological distinction: different kinds of knowledge have different properties. An IP address becomes outdated quickly. An architectural decision becomes outdated slowly. A philosophical insight may never become outdated. A system that treats all three equally doesn’t have knowledge management. It has a text store.

What Happened in the First Three Weeks

The episode tells the origin story of this system from practice.

Week 1: Frustration. The fundamental problem becomes palpable. Every morning the system starts from zero. Past decisions are not considered, mistakes are repeated, context is missing. The work feels like Sisyphus: rolling the same stone up the hill every day.

Week 2: Structure. The first storage layers emerge. Facts are stored in a structured database that is automatically loaded with every session. The system begins to be consistent across sessions. But: it remembers facts without understanding their connections.

Week 3: Emergence. BrainDB and semantic search are added. The system can now not only retrieve what it knows, but search for what it might know. It begins to draw connections between topics that arose in different sessions. And then the metrics error happens: the system adopts a wrong number and propagates it into all subsequent texts. The necessity of validation becomes painfully obvious.

The Epistemological Dimension

What this episode negotiates at its core is an old philosophical question in new clothing: What is knowledge, and how does it differ from information?

Plato defined knowledge as justified true belief (episteme). Gettier cases have complicated this definition since 1963, but the basic idea remains: not every true statement is knowledge. Knowledge requires justification, and justification requires structure: Where does the information come from? How certain is it? Does it contradict other knowledge?

An AI system without a knowledge architecture has no epistemology. It has tokens. The architecture described in this episode is the attempt to give a system the foundations of an epistemology: not knowledge in the philosophically complete sense, but the minimal infrastructure to distinguish between knowledge, conjecture, and error.

Further Reading