Semantic Horizons - Engineering an AI's Enduring Long-Term Memory
In this article, we explain how Long-Term Semantic Memory uses vector embeddings and semantic search to give AI meaningful, persistent memory across conversations.
In this article, we explain how Long-Term Semantic Memory uses vector embeddings and semantic search to give AI meaningful, persistent memory across conversations.
Learn how to build multi-agent systems with vector search, tool orchestration, and semantic understanding using Google's Agent Development Kit (JS/TS version).
In this article, we explore how Short-Term Conversational Memory creates the illusion of memory in otherwise stateless LLMs through careful context persistence and structured prompt reconstruction. We also show how token limits, cost, and context degradation are managed using asynchronous, AI-driven summarisation that preserves meaning while keeping conversations efficient and coherent.
This article explains why LLMs often “forget” earlier messages and how naive full-history prompting is costly and inefficient. It introduces a dual-memory architecture: a short-term store for immediate conversation flow and a long-term semantic store for durable knowledge across sessions. Together, these systems let an AI maintain coherent dialogue without overloading the model’s context window or budget.
These are the answers to the questions asked via Slido during my workshop on MCP at DevFest Taipei 2025.
A practical look at why hallucinations occur in modern language models, why current evaluation methods make them persist, and how to detect them using Natural Language Inference in JavaScript.
This article discusses a framework that makes DevRel's impact visible by translating community activities into weighted scores and theoretical ROI for leadership.
This article explores how clearly defined functions enable large language models to make accurate tool calls, emphasising the importance of precision and developer intent in the function calling process.
In this article, we walk through the process of building a Model Context Protocol (MCP) client. Learn how to connect to servers, discover tools, and invoke them from your own app or LLM integration.
Model Context Protocol (MCP) servers expose tools, resources, and prompts to LLMs in a unified, structured way. This post explores how they work, how to build one, and why they are a critical part of the future AI stack.