Skip to main content

Tag

Memory

6 articles tagged with Memory.

  1. The Mind's Eye - Engineering a CLI for Intelligent AI Interaction

    This article concludes the series by showing how a deliberately designed CLI becomes a powerful interaction layer, giving users precise control over an AI system's conversational context, short-term memory, and long-term semantic knowledge.

  2. The Autonomous Brain - Engineering AI for Continuous Learning and Memory Enrichment

    This piece introduces background processors as autonomous AI agents that summarise conversations and extract critical facts to continuously enrich Long-Term Semantic Memory. By running asynchronously and optimising token usage, these processors enable a self-improving, increasingly personalised AI system that learns from every interaction.

  3. The 'Aha!' Moment - Engineering the Perfect Prompt for Truly Contextual AI

    In this article, we show how dynamic prompt engineering—via a `SessionManager` that intelligently layers short-term context, long-term semantic memory, and system instructions turns stateless LLM calls into genuinely contextual and personalised conversations.

  4. Semantic Horizons - Engineering an AI's Enduring Long-Term Memory

    In this article, we explain how Long-Term Semantic Memory uses vector embeddings and semantic search to give AI meaningful, persistent memory across conversations.

  5. The Token Economy - Engineering an AI's Working Memory

    In this article, we explore how Short-Term Conversational Memory creates the illusion of memory in otherwise stateless LLMs through careful context persistence and structured prompt reconstruction. We also show how token limits, cost, and context degradation are managed using asynchronous, AI-driven summarisation that preserves meaning while keeping conversations efficient and coherent.

  6. The Grand Orchestration - Engineering a Dual-Memory AI for Enduring Conversations

    This article explains why LLMs often forget earlier messages and how naive full-history prompting is costly and inefficient. It introduces a dual-memory architecture: a short-term store for immediate conversation flow and a long-term semantic store for durable knowledge across sessions. Together, these systems let an AI maintain coherent dialogue without overloading the context window or budget.