Hands-on tutorials for LLM observability and AgentOps tools.
Each directory is a self-contained module with its own notebooks, dependencies, and setup instructions.
agentops-learn/
├── docs/
│ └── market-review.md # Market analysis — 9 vendors compared
│
├── langfuse/ # ✅ Done
│ ├── 01_basic_tracing.ipynb
│ ├── 02_decorator_and_nesting.ipynb
│ ├── 03_prompt_management.ipynb
│ ├── 04_scoring_and_evaluation.ipynb
│ └── 05_datasets_and_experiments.ipynb
│
├── agentops-ai/ # 📋 Planned
├── braintrust/ # 📋 Planned
├── helicone/ # 📋 Planned
├── arize-phoenix/ # 📋 Planned
│
├── .env.example # All API keys in one place
└── .gitignore
| Directory | Tool | What it covers | Status |
|---|---|---|---|
langfuse/ |
Langfuse | Tracing, @observe(), prompt management, scoring, dataset experiments |
Done |
agentops-ai/ |
AgentOps.ai | Multi-agent tracing, time-travel debugging | Planned |
braintrust/ |
Braintrust | Production evals, CI/CD deploy gates | Planned |
helicone/ |
Helicone | Proxy-based integration, cost tracking | Planned |
arize-phoenix/ |
Arize Phoenix | OTel-native tracing, embedding visualization | Planned |
Why these five? Selected from 9 vendors reviewed for high hands-on value and distinct positioning. Excluded: LangSmith (LangChain-tied), W&B Weave (W&B-tied), Datadog (enterprise), Traceloop (instrumentation layer only).
# 1. Copy and fill in your API keys
cp .env.example .env
# 2. Pick a tool directory
cd langfuse/ # (or agentops-ai/, braintrust/, etc.)
# 3. Copy .env and set up
cp ../.env .env
python3 -m venv .venv && source .venv/bin/activate
pip install -r requirements.txt
# 4. Run
jupyter notebook- Market review — vendor comparison, pricing, selection criteria
- OpenTelemetry GenAI Semantic Conventions — the emerging standard all tools converge on