Interactive demonstration of verify-as-you-generate, stealth-edit correction, and concept stability at scale.
Select a theorem from the dropdown below and press Enter to prove it with live step-by-step verification.
Model fails all 32 attempts on or_elim. A perturbation Δ·u is optimized at Layer 10 FFN. Hook fires only during KV-cache prefill.
DLCM concept pooling makes representations ~3x more stable under token-level paraphrasing than vanilla token representations.
Measured as cosine similarity between original and paraphrased internal representations across 4 model scales (23M-206M params).