Epic: Part 7 — AI Evaluations
Priority: P1
Role: Engineer
User Story
As an Engineer, I want to instrument Glow CI LLM calls with Cloud Logging and connect to Vertex AI Evaluation Service so that all RAG retrievals and Gemini responses are observable and Vertex AI can run LLM judge evaluations.
Context
All Glow CI LLM calls (RAG retrieval + Gemini synthesis) are logged via Cloud Logging on GCP and fed into Vertex AI Evaluation Service for automated LLM judge evaluation. This replaces any third-party observability tooling — we stay fully within the GCP stack. Sprint timing: chat must be built first so there are real LLM calls to instrument.
Acceptance Criteria
Dependencies
📄 PRD: Part 7 — Glow CI PRD
Epic: Part 7 — AI Evaluations
Priority: P1
Role: Engineer
User Story
As an Engineer, I want to instrument Glow CI LLM calls with Cloud Logging and connect to Vertex AI Evaluation Service so that all RAG retrievals and Gemini responses are observable and Vertex AI can run LLM judge evaluations.
Context
All Glow CI LLM calls (RAG retrieval + Gemini synthesis) are logged via Cloud Logging on GCP and fed into Vertex AI Evaluation Service for automated LLM judge evaluation. This replaces any third-party observability tooling — we stay fully within the GCP stack. Sprint timing: chat must be built first so there are real LLM calls to instrument.
Acceptance Criteria
Dependencies
📄 PRD: Part 7 — Glow CI PRD