LLM Foundations

Core mechanics of large language models — training, inference, tokenization, and the mathematics of next-token prediction.

Conceptual visualization of agent guardrails enforcing permission boundaries on autonomous AI tool calls and outputs
MONA explainer 11 min

What Are Agent Guardrails? How Permission Systems Constrain AI

Agent guardrails enforce permission boundaries on autonomous AI. Learn how Claude SDK, NeMo, and Llama Guard constrain …

Concentric runtime checkpoints around an LLM agent showing input, output, and tool-call boundaries with permeable filters
MONA explainer 11 min

Prerequisites for Agent Guardrails: Tool Use and Runtime Limits

Agent guardrails are runtime classifiers wrapped around tool-use loops — useful, partial, and demonstrably evadable. …

Autonomous agent paused at an interrupt checkpoint awaiting human approval before resuming a workflow
MONA explainer 12 min

Prerequisites and Technical Limits of HITL for AI Agents

HITL for agents is easy to start and hard to scale. Learn the prerequisites — durable state, idempotency, escalation — …

Geometric visualization of an approval gate paused between an autonomous agent and a tool call
MONA explainer 11 min

Human-in-the-Loop for AI Agents: How Approval Gates Work

Human-in-the-loop for AI agents pauses autonomous workflows at risky steps and routes them to a human gate. Here's how …

Diagram of an LLM agent loading checkpoint snapshots from a thread before each reasoning step
MONA explainer 10 min

Agent State Management: Threads, Checkpointers, Hard Limits

Agent state is not memory — it is plumbing that replays snapshots between steps. Mona explains threads, checkpointers, …

Graph of state snapshots linked by a checkpoint thread across reasoning turns inside an agent runtime
MONA explainer 10 min

Agent State Management: How Checkpointing Persists Memory Across Turns

Agent state management decides whether your agent remembers. See how LangGraph checkpointers, threads, and reducers …

Sequence of tool calls forming an agent trajectory graded against a reference path
MONA explainer 10 min

Agent Evaluation: How Trajectory Analysis Measures AI Agents

Agent evaluation grades the path, not just the final answer. Learn how trajectory analysis exposes silent reasoning …

Layered diagram of agent evaluation showing outcome judgment, trajectory analysis, and cost-per-task observability stacked over a benchmark surface.
MONA explainer 11 min

Agent Evaluation Prerequisites: LLM-as-Judge to Cost-Per-Task

Agent evaluation needs three signals: outcome, trajectory, cost. Learn why LLM-as-judge has known biases and where major …

Layered diagram of an agent loop showing thought, action, and observation stages with branching planning paths
MONA explainer 14 min

From Chain-of-Thought to Tool Use: Prerequisites and Technical Limits of Agent Planning

Agent planning rests on three primitives — chain-of-thought, tool use, and the ReAct loop. Learn the prerequisites and …

Diagram of three multi-agent architectures: supervisor, debate, and swarm patterns coordinating AI agents
MONA explainer 12 min

Multi-Agent Systems: Supervisor, Debate, and Swarm Patterns

Multi-agent systems coordinate specialized AI agents through supervisor, debate, or swarm patterns. Here is how each …

Layered diagram of multi-agent prerequisites: tool use as the atomic primitive, the ReAct loop, and short- and long-term memory
MONA explainer 13 min

Multi-Agent Systems: Prerequisites and Hard Technical Limits

Before multi-agent systems, master tool use, the ReAct loop, and memory. Then face the limits: context blow-up, error …

Layered diagram of an LLM agent memory architecture with vector store, temporal graph, and self-editing memory blocks
MONA explainer 12 min

Agent Memory Systems: How LLMs Get Persistent Recall Across Sessions

Agent memory systems give LLMs persistent recall across sessions. Inside the architectures: temporal graphs, …

Three architectural diagrams contrasting graph state, actor message passing, and crew task handoff patterns in agent orchestration
MONA explainer 11 min

Graph vs Conversation vs Crew: LangGraph, AutoGen, CrewAI Patterns

LangGraph, AutoGen, and CrewAI commit to three different theories of how AI agents coordinate. The pattern you pick …

Diagram of an AI agent loop showing reasoning traces, tool actions, and a self-reflection memory feeding the next step
MONA explainer 10 min

Agent Planning and Reasoning: ReAct, Plan-and-Execute, Reflexion

Agent planning is not human cognition — it is token generation conditioned on observations. How ReAct, Plan-and-Execute, …

Tiered memory layers compressing into a temporal knowledge graph for AI agents
MONA explainer 10 min

Agent Memory Architectures: Prerequisites and Hard Limits

Agent memory isn't a bigger context window. Learn the prerequisites for designing agent memory systems and the hard …

LLM agent loop wiring reasoning to tools, memory, and a control plane across three orchestration frameworks.
MONA explainer 12 min

Agent Frameworks: How LangGraph, CrewAI, and AutoGen Orchestrate LLMs

Agent frameworks orchestrate LLM calls, tools, and memory — but each one bets on a different abstraction. Learn what …

Salient object segmentation pipeline isolating a foreground subject from a busy background using alpha matting and per-pixel opacity
MONA explainer 10 min

What Is AI Background Removal? How Salient Object Segmentation Works

AI background removal is not one model — it's salient object detection plus alpha matting. See how U2-Net, BiRefNet, and …

Text tokens flowing into a diffusion latent space, becoming geometric attention maps that resolve into a generated image
MONA explainer 13 min

Prompt Engineering for Image Generation: How Diffusion Models Read Text

Image prompts steer probability, not pixels. Learn how diffusion models, cross-attention, and CFG turn text into images …

Layered diagram of prompt parsing across diffusion, autoregressive, and multimodal image models
MONA explainer 9 min

Negative Prompts, Weights, Seeds: Image Prompting Limits 2026

Negative prompts and weight syntax aren't universal — and seed reproducibility breaks across model versions. Inside the …

Diagram of an alpha matte separating a portrait from its background, with hair edges marked as the unknown band
MONA explainer 12 min

Alpha Channels, Trimaps, and the Hard Limits of AI Background Removal

Background removal is alpha estimation, not subject detection. Learn how trimaps and matting work, and why hair, glass, …

Low-rank adapter matrices BA layered onto a frozen diffusion model for image generation fine-tuning
MONA explainer 11 min

How LoRA Fine-Tunes Diffusion Models for Image Generation

LoRA fine-tunes Stable Diffusion and FLUX without retraining. Learn how rank, alpha, and the BA decomposition turn a …

AI image upscaling structural limits at 4K and 8K - diffusion priors hallucinate faces and tile-local processing produces visible seams
MONA explainer 12 min

Why AI Upscalers Hallucinate Faces and Tile Seams at 4K and 8K

AI upscalers don't break at 4K and 8K because of weak hardware. The failures are structural — rooted in diffusion priors …

Low-resolution pixels expanding into a high-resolution image through generative neural-network inference
MONA explainer 11 min

What Is Image Upscaling and How AI Super-Resolution Reconstructs Detail Beyond the Original Pixels

AI image upscaling doesn't enlarge what was captured — it generates plausible pixels from a learned prior. Learn how GAN …

Frozen diffusion model weights with low-rank adapter matrices flowing into the UNet attention block during LoRA training
MONA explainer 11 min

Training Image LoRAs: Diffusion Math, Rank-Alpha, and VRAM Limits

Image LoRAs retarget diffusion models with small adapter files. Learn the rank-alpha math, VRAM ranges from SD 1.5 to …

Anatomy of an AI upscaler — residual dense blocks on one side, a diffusion prior on the other, sharing one degraded input
MONA explainer 13 min

From RRDB Blocks to Diffusion Priors: Inside Modern AI Upscalers

How modern AI upscalers are built — from ESRGAN's RRDB blocks and Real-ESRGAN to SUPIR's diffusion prior, plus the …

Diagram of AI image editing: mask-guided inpainting, canvas outpainting, and instruction-based diffusion edit
MONA explainer 12 min

What Is AI Image Editing? Inpainting, Outpainting, Edit Models

AI image editing uses diffusion to modify pixels under a mask or follow text instructions. Learn how inpainting, …

Noise-to-image diffusion process with a text instruction transforming a latent representation into an edited output
MONA explainer 10 min

From Diffusion to InstructPix2Pix: AI Image Editing Prerequisites

Before using GPT Image or FLUX, understand diffusion, classifier-free guidance, and why InstructPix2Pix made …

Layered neural network architecture showing signal propagation and gradient flow through weighted connections
MONA explainer 13 min

What Is a Neural Network and How It Learns to Generate Language

Neural networks learn language by adjusting millions of weights through backpropagation. Learn how layers, gradients, …

MONA tracing signal flow through neural network layers from ReLU to SwiGLU activation functions
MONA explainer 10 min

From ReLU to SwiGLU: How Activation and Loss Functions Shape LLM Training

Trace the path from ReLU to SwiGLU and understand how activation functions, cross-entropy loss, and gradient dynamics …

Gradient arrows flowing backward through layered neural network nodes toward a loss function surface
MONA explainer 9 min

Backpropagation and Gradient Descent: How Neural Networks Learn From Errors

Learn how backpropagation and gradient descent train neural networks by propagating error signals backward through …