insideLLMs is a Python library and CLI for comparing LLM behaviour across models using shared probes and datasets. The harness is deterministic by design, so you can store run artefacts and reliably diff behaviour in CI.
Multi‑agent AI security testing framework that orchestrates red‑team analyses, consolidates findings with an arbiter, and records an immutable audit ledger—plus a deterministic demo mode for repeatable results.
PhD in Cognitive Neuroscience & AI · 15+ years in human-centred AI · 14 publications · h-index 12
Culture Amp · NEOS Insurance · Source Localisation · University of Sydney



