Agentic Workflow

Automated Telemetry Engine

Platform

Gemini 3.1 Pro + 3 Flash

ROLE

Experimentation Specialist

EXPERTISE

Product

YEAR

2026

Project Description

Project Description

Project Description

Turn a product hypothesis into an executive-ready telemetry schema.

When product teams hypothesize an A/B test or experiment, defining the exact KPIs, guardrail metrics, and edge cases usually takes days of alignment.

Even the best product leaders have blind spots.

To solve this, I built an AI-powered telemetry engine that doesn't just agree with the user, it aggressively stress-tests their hypotheses.

See the following video for a masterclass on how this cutting-edge agentic workflow operates

When product teams hypothesize an A/B test or experiment, defining the exact KPIs, guardrail metrics, and edge cases usually takes days of alignment.

Even the best product leaders have blind spots.

To solve this, I built an AI-powered telemetry engine that doesn't just agree with the user, it aggressively stress-tests the hypothesis.

The Business Impact

  • Eliminated Blind Spots: Catches massive revenue traps before an experiment is ever deployed.

  • Massive Time Savings: Reduces telemetry planning and cross-functional swirl from days to hours.

    Agentic Workflow Structure

    A1 · KPI Generator Agent

  • Purpose: Generate primary and secondary KPIs to track for a proposed product experiment.

  • Inputs: Experiment hypothesis.

  • Process: Analyzes the hypothesis, determines the core business objectives, and maps out the primary and secondary tracking metrics needed to accurately measure success.

  • Outputs: Proposed KPIs (e.g., Average Order Value, Conversion Rate).

  • Checks: Relevance to hypothesis, baseline trackability.

A2 · KPI QA Agent

  • Purpose: Evaluate A1’s output to ensure metrics are sound, trackable, and not hallucinated.

  • Inputs: A1’s proposed KPIs + original hypothesis.

  • Process: Acts as a automated "circuit breaker." Audits A1’s work against strict telemetry standards.

    • If it finds major issues or untrackable metrics, it rejects the output, provides feedback, and forces A1 to run again.

  • Outputs: Approved KPIs or Rejection/Feedback Log.

  • Checks: Trackability, business logic validity, zero hallucination.

A3 · Edge Case Generator Agent

  • Purpose: Identify hidden risks, guardrail metrics, and edge cases to ensure no stone is left unturned.

  • Inputs: Approved KPIs + original hypothesis.

  • Process: Hunts for risks and friction points.

    • Formulates guardrail metrics to protect core business health during the experiment.

  • Outputs: Proposed Guardrails and Edge Cases.

  • Checks: Relevance to experiment, severity of impact, gap coverage.

A4 · Edge Case QA Agent

  • Purpose: Filter theoretical concerns from practical, deployable, actionable risks.

  • Inputs: A3’s proposed edge cases and guardrails.

  • Process: Serves as the second "circuit breaker."

    • Strictly filters A3’s output to ensure only highly actionable, realistic edge cases are passed.

    • Forces A3 to regenerate if the output does not meet high-quality standards.

  • Outputs: Approved Guardrails and Edge Cases.

  • Checks: Deployability, actionability, theoretical vs. practical distinction.

A5 · Final Output Agent

  • Purpose: Compile and format the fully QA'd telemetry data into an easy to approach schema.

  • Inputs: Approved KPIs (from A2) + Approved Edge Cases (from A4).

  • Process: Synthesizes the data, applies Gemini-powered explanations for specific metrics, and formats it into an approachable, easy-to-understand UI report.

  • Outputs: Final Telemetry Report (reiterated hypothesis, defined KPIs, guardrails, and edge cases).

  • Checks: Readability, completeness, alignment.


(View on Desktop to see demo video)


The Business Impact

  • Eliminated Blind Spots: Catches massive revenue traps before an experiment is ever deployed.

  • Massive Time Savings: Reduces telemetry planning and cross-functional swirl from days to hours.

    Agentic Workflow Structure

    A1 · KPI Generator Agent

  • Purpose: Generate primary and secondary KPIs to track for a proposed product experiment.

  • Inputs: Experiment hypothesis.

  • Process: Analyzes the hypothesis, determines the core business objectives, and maps out the primary and secondary tracking metrics needed to accurately measure success.

  • Outputs: Proposed KPIs (e.g., Average Order Value, Conversion Rate).

  • Checks: Relevance to hypothesis, baseline trackability.

A2 · KPI QA Agent

  • Purpose: Evaluate A1’s output to ensure metrics are sound, trackable, and not hallucinated.

  • Inputs: A1’s proposed KPIs + original hypothesis.

  • Process: Acts as a automated "circuit breaker." Audits A1’s work against strict telemetry standards.

    • If it finds major issues or untrackable metrics, it rejects the output, provides feedback, and forces A1 to run again.

  • Outputs: Approved KPIs or Rejection/Feedback Log.

  • Checks: Trackability, business logic validity, zero hallucination.

A3 · Edge Case Generator Agent

  • Purpose: Identify hidden risks, guardrail metrics, and edge cases to ensure no stone is left unturned.

  • Inputs: Approved KPIs + original hypothesis.

  • Process: Hunts for risks and friction points.

    • Formulates guardrail metrics to protect core business health during the experiment.

  • Outputs: Proposed Guardrails and Edge Cases.

  • Checks: Relevance to experiment, severity of impact, gap coverage.

A4 · Edge Case QA Agent

  • Purpose: Filter theoretical concerns from practical, deployable, actionable risks.

  • Inputs: A3’s proposed edge cases and guardrails.

  • Process: Serves as the second "circuit breaker."

    • Strictly filters A3’s output to ensure only highly actionable, realistic edge cases are passed.

    • Forces A3 to regenerate if the output does not meet high-quality standards.

  • Outputs: Approved Guardrails and Edge Cases.

  • Checks: Deployability, actionability, theoretical vs. practical distinction.

A5 · Final Output Agent

  • Purpose: Compile and format the fully QA'd telemetry data into an easy to approach schema.

  • Inputs: Approved KPIs (from A2) + Approved Edge Cases (from A4).

  • Process: Synthesizes the data, applies Gemini-powered explanations for specific metrics, and formats it into an approachable, easy-to-understand UI report.

  • Outputs: Final Telemetry Report (reiterated hypothesis, defined KPIs, guardrails, and edge cases).

  • Checks: Readability, completeness, alignment.