docs(verify): add Step 1 Orient — think out loud on claim + evidence#22
docs(verify): add Step 1 Orient — think out loud on claim + evidence#22bensonwong merged 6 commits intomainfrom
Conversation
Every /verify call now opens with a short CoT preamble naming the claim(s) and the evidence (or primary sources to look up if none was provided). Gives users scope and progress clarity before the pipeline runs, and surfaces reasoning rather than jumping straight into prepare. Existing Prepare/Respond/Verify steps renumbered to 2/3/4 and inline cross-references updated. https://claude.ai/code/session_01LqnxD512JufmQd1VJn5F1P
prepare latency is asymmetric: URLs and office files take ~20-30s, while images and PDFs finish in ~0.5s after upload. Previously, Step 1 told the model to write the preamble "before running any command," which serialized the CoT gate ahead of slow network fetches and made the user wait on silence during the ~30s prepare window. Now the Orient step instructs the model to fire prepare in the SAME assistant turn as the preamble — text streams to the user first, so they read the orientation while the slow prepare is already running. Added a latency table and three examples (slow URL, fast PDF, mixed) to make the parallelism rule concrete. https://claude.ai/code/session_01LqnxD512JufmQd1VJn5F1P
PR Review: docs(verify): add Step 1 OrientOverall: This is a well-scoped, focused change that meaningfully improves the UX of What works well
Minor issues / suggestions1. Hardcoded latency estimates may go stale | URLs (web pages) | ~20–30s per URL |
| Office files (`.docx`, `.xlsx`, `.pptx`) | ~20–30s per file |These numbers are baked in as constants. If 2. "CoT gate" jargon in a model-facing doc
"CoT" is meaningful to AI researchers but could be opaque in a skill file that the model uses as instructions. Since the sentence already explains the purpose ("user clarity and progress"), consider dropping "CoT gate" or spelling it out: "Chain-of-Thought step." 3. In Step 4, the comment:
Reading "skipped Step 2–3" while in "Step 4" is slightly confusing — the reader has to mentally map back. Spelling it out as "skipped the Prepare and Respond steps" would be unambiguous. 4. Example claim may become stale
This is a good illustrative example, but benchmark numbers get superseded quickly. A more time-stable placeholder (e.g., a legislative or scientific claim) would age better. Low priority — it's just an example. No bugs foundAll step numbers, table entries, and Verdict: Ready to merge with the latency-estimate and jargon points treated as optional follow-up. None are blockers. |
Establishes agent guidance for this repo following the same progressive- disclosure pattern as packages/deepcitation. Includes always-applicable invariants (product name, generic-examples rule, auth command) and a guidance router pointing to SKILL.md and the rules/ files. Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Consistent with auth.md living under rules/. SKILL.md reference updated to match (rules/cloud-sandbox-constraints.md). Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
The CLI renamed the login subcommand to auth. Updates auth.md recovery steps and the --key flag example accordingly. Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
…ange split
- Replace specific model/benchmark claim ("Claude 4.6 Sonnet...SWE-bench")
with a generic placeholder per the AGENTS.md invariant: use domain-
neutral examples that don't carry unintended product implications
- Clarify Step 4 --html skip note: "skipped the Prepare and Respond steps"
instead of "skipped Step 2–3" (unambiguous without mental mapping)
- Add stale Cowork proxy clearing snippet before prepare invocations
- Add page-range split algorithm for parallel agent dispatch (eliminates
f-fabrication by restricting each agent to its own page window)
- Add unique citation ID, Format 2, and bold=k hard rules to agent prompts
Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
Every /verify call now opens with a short CoT preamble naming the
claim(s) and the evidence (or primary sources to look up if none was
provided). Gives users scope and progress clarity before the pipeline
runs, and surfaces reasoning rather than jumping straight into prepare.
Existing Prepare/Respond/Verify steps renumbered to 2/3/4 and inline
cross-references updated.
https://claude.ai/code/session_01LqnxD512JufmQd1VJn5F1P