Who this path serves

Tech leads, EM/TEMs, and program owners who coordinate model rollouts, feature flags, and stakeholder sign-offs. You need a concise map of what to freeze, what to monitor, and what to disclose—without reading three books in series order.

Skim all three experiment hubs

In any order, read evaluation, RAG, and prompts—each is a one-screen orientation before the long essay.

Cross-reference the long essays

Dip into prompts, RAG, and evaluation for sections relevant to your next release (freeze prompts, retrieval freshness, eval gates).

Checklist anchors

  • Pre-freeze: prompt version hashes, model IDs, corpus versions (operations).
  • Week after: drift monitors, refusal deltas, retrieval empty-rate (measurement).
  • Support: known gaps doc linked from safety themes.

Related

Latest experiments hub · Topic map · About for editorial boundaries.

Stakeholder artifacts

Bring a one-page risk tier summary, a table of frozen identifiers (model, corpus, prompt hashes), and a short list of monitors with owners. Security and finance reviews go faster when trade-offs are explicit: what you are not testing yet, and what would trigger a rollback.

Other paths

Foundations-first readers use Path A; document-centric teams often prefer Path B. The paths hub ties them together.

Communication cadence

Send a short pre-release note—what changes, what is monitored, how to report issues—and a post-release summary with metrics versus expectations. Consistency builds trust faster than ad-hoc threads; it also creates an audit trail for compliance questions.