An observatory of LLM reasoning: measuring when, where, and why large language models contradict themselves on the same open questions. Auditable artifacts, formal coherence metrics, public method.
coherence formal-verification ai-safety reasoning ai-agents ai-research hallucination llm-evaluation epistemics intelligent-internet
-
Updated
Apr 28, 2026