AI Principles

The science behind AI — transformer architectures, training dynamics, and evaluation methodology. MONA explains how AI actually works, with precision over hype.

Layered gate diagram showing information flowing through forget, input, and output gates inside a recurrent cell
MONA explainer 11 min

From Vanilla RNN to LSTM and GRU: How Gating Mechanisms Solved the Long-Term Memory Problem

Trace how LSTM forget, input, and output gates fix the vanishing gradient problem that crippled vanilla RNNs, and how …

MONA tracing signal flow through neural network layers from ReLU to SwiGLU activation functions
MONA explainer 10 min

From ReLU to SwiGLU: How Activation and Loss Functions Shape LLM Training

Trace the path from ReLU to SwiGLU and understand how activation functions, cross-entropy loss, and gradient dynamics …

Gradient arrows flowing backward through layered neural network nodes toward a loss function surface
MONA explainer 9 min

Backpropagation and Gradient Descent: How Neural Networks Learn From Errors

Learn how backpropagation and gradient descent train neural networks by propagating error signals backward through …

Standardized testing pipeline comparing language model outputs through identical benchmark scoring frameworks
MONA explainer 10 min

What Is an Evaluation Harness and How Standardized Frameworks Benchmark LLMs

Evaluation harnesses standardize LLM benchmarking by fixing prompts, scoring, and conditions. Learn how the pipeline …

Geometric measurement instruments producing divergent readings from identical evaluation benchmark data
MONA explainer 10 min

Benchmark Contamination, Score Divergence, and the Technical Limits of LLM Evaluation Harnesses

Same model, same benchmark, different scores. Understand why evaluation harnesses diverge and how benchmark …

Abstract visualization of overlapping training and evaluation data sets with highlighted contamination pathways
MONA explainer 11 min

What Is Benchmark Contamination and How Training Data Overlap Inflates LLM Evaluation Scores

Benchmark contamination inflates LLM scores when training data overlaps with test sets. Learn how data leaks in and why …

Overlapping n-gram patterns dissolving into noise, visualizing benchmark contamination detection thresholds
MONA explainer 10 min

Benchmark Contamination: N-Gram Overlap and Hard Limits

Benchmark contamination and overfitting look identical in scores. Understand what n-gram overlap, deduplication, and …

Geometric diagram of neural network layers being systematically removed to reveal component contributions
MONA explainer 10 min

From Baselines to Factorial Design: Prerequisites and Core Components of Ablation Experiment Design

Ablation studies reveal which components matter, but only with the right baselines, controls, and statistical methods. …

Balanced and imbalanced confusion matrix grids revealing hidden failure patterns in classification metrics
MONA explainer 10 min

Class Imbalance, Normalization Traps, and the Hard Limits of Confusion Matrix Analysis

Confusion matrices hide failures under class imbalance. Learn how normalization direction changes what you see and why …

Grid of academic subject icons radiating from a central multiple-choice evaluation node with accuracy gradients
MONA explainer 9 min

What Is the MMLU Benchmark and How 57 Academic Subjects Test LLM Knowledge

MMLU tests large language models across 57 academic subjects with 15,908 questions. Learn how it works, where it breaks, …

Neural network architecture with components systematically removed revealing internal dependency patterns
MONA explainer 10 min

What Is an Ablation Study and How Removing Components Reveals What Makes AI Models Work

Ablation studies reveal what each model component does by removing it. Learn the experimental design and failure modes …

Geometric grid mapping classifier predictions against actual outcomes with highlighted error cells and diagnostic metric
MONA explainer 10 min

What Is a Confusion Matrix and How It Reveals Where Your Classifier Fails

A confusion matrix reveals exactly where classifiers fail. Understand true positives, false negatives, and why accuracy …

Fractured multiple-choice exam grid revealing label errors and score saturation in LLM benchmark evaluation
MONA explainer 10 min

MMLU's 6.5% Label Error Rate and Benchmark Score Saturation

MMLU's 6.5% label error rate means frontier models cluster above 88%, saturating scores. Score saturation explains why …

Geometric diagram showing interconnected measurement tools converging on a single evaluation score
MONA explainer 10 min

From Perplexity to Few-Shot Prompting: Prerequisites for Understanding Evaluation Harness Internals

Evaluation harness scores depend on perplexity, few-shot prompting, and tokenization most teams skip. Learn the …

Grid of prediction outcomes revealing hidden classification failures through color-coded diagonal and off-diagonal cells
MONA explainer 10 min

From Binary to Multi-Class: Deriving Precision, Recall, and F1 from a Confusion Matrix

Precision, recall, and F1 all come from the same confusion matrix. Learn to extract each metric for binary and …

Geometric binary tree with exponentially branching nodes overlaid on a fading neural network grid
MONA explainer 11 min

Combinatorial Explosion, Interaction Effects, and the Hard Limits of Ablation Studies at Scale

Ablation studies hit a wall at scale: combinatorial explosion and non-additive interactions make exhaustive testing of …

Confusion matrix with the true-negative quadrant dissolving to reveal a hidden gap in metric coverage
MONA explainer 10 min

Why F1 Score Fails on Imbalanced Datasets: MCC, PR-AUC, and the Limits of Harmonic Averaging

F1 score hides classifier failures on imbalanced datasets by ignoring true negatives. Learn why MCC and PR-AUC reveal …

Geometric visualization of precision and recall intersecting within a confusion matrix grid
MONA explainer 9 min

Precision, Recall, F1 Score: What the Confusion Matrix Reveals

What accuracy won't show: precision, recall, and F1 score expose true classifier performance. The confusion matrix …

Geometric grid of colored cells representing a confusion matrix decomposing into precision and recall pathways
MONA explainer 10 min

From True Positives to Macro Averaging: The Building Blocks Behind Precision, Recall, and F1

Precision, recall, and F1 score measure what accuracy hides. Learn how true positives, confusion matrices, and macro …

Geometric visualization of benchmark scores converging and diverging across evaluation dimensions
MONA explainer 11 min

What Is Model Evaluation and How Benchmarks, Metrics, and Human Judgment Measure LLM Quality

Model evaluation combines benchmarks, automated metrics, and human judgment to measure LLM quality. Learn why high …

Four divergent scoring dimensions representing probability, text overlap, recall, and preference intersecting around a
MONA explainer 10 min

Perplexity, BLEU, ROUGE, and ELO: The Core Metrics Behind LLM Evaluation Explained

Perplexity, BLEU, ROUGE, and Elo measure fundamentally different properties of language models. Learn when each metric …

Abstract visualization of benchmark scores fracturing as contamination patterns distort evaluation metrics
MONA explainer 10 min

Benchmark Contamination, Metric Gaming, and the Hard Limits of LLM Evaluation

Benchmark contamination inflates LLM scores while real-world performance lags. Learn why metric gaming and saturated …

Balanced probability distributions splitting across protected groups with a fairness threshold line
MONA explainer 10 min

What Are Bias and Fairness Metrics and How They Detect Discrimination in ML Predictions

Fairness metrics test whether ML models discriminate by group. Learn how disparate impact, equalized odds, and the …

Mathematical proof notation with competing fairness metric equations pulling a balance point in three irreconcilable
MONA explainer 10 min

The Impossibility Theorem and Why No Model Can Satisfy Every Fairness Metric at Once

When group base rates differ, no algorithm satisfies calibration, equal error rates, and demographic parity at once. …

Toxicity classifier decision boundaries separating harmful from safe regions in AI output evaluation space
MONA explainer 10 min

What Is Toxicity and Safety Evaluation and How Guard Models Score Harmful AI Outputs

Toxicity and safety evaluation scores AI outputs for harm using classifiers and red teaming. Learn how guard models …

Overlapping safety benchmark taxonomies visualized as intersecting geometric planes with color-coded hazard categories
MONA explainer 10 min

HarmBench, ToxiGen, and MLCommons Taxonomy: The Datasets and Standards Behind AI Safety Testing

HarmBench, ToxiGen, and MLCommons AILuminate define how AI safety is measured. Learn the datasets, classifiers, and …

Three intersecting geometric boundaries representing competing fairness constraints across a population distribution
MONA explainer 10 min

Demographic Parity vs. Equalized Odds vs. Calibration: Core Fairness Metrics Compared

Demographic parity, equalized odds, and calibration define fairness differently and cannot all be satisfied at once. …

Diverging toxicity confidence scores revealing systematic classifier bias patterns across different language dialects
MONA explainer 10 min

False Positives in Toxicity Detection: Dialect Bias, Bypasses

Toxicity classifiers over-flag minority dialects and miss adversarial attacks. Explore the statistical bias—from dialect …

Particles forming adversarial attack vectors converging on an AI model decision boundary
MONA explainer 10 min

Red Teaming for AI: Adversarial Testing Exposes Failures

Red teaming uses adversarial testing to reveal AI vulnerabilities. Discover what it catches, mechanics, and why it …

Geometric diagram of interconnected security framework layers mapping AI system vulnerabilities
MONA explainer 11 min

OWASP LLM Top 10, MITRE ATLAS, and the Frameworks That Structure AI Red Teaming

OWASP LLM Top 10 and MITRE ATLAS give red teams structured attack categories. Learn how these frameworks turn AI …