Agora.io  ·  2024  ·  9 months

Designing for reducing time to first agent from hours to under 3 minutes.

Reducing time to first agent from roughly 30 to 60 minutes down to 3 minutes.

Role

Principal Product Designer

Team

1 PM, 5 FE, 3 BE, 3 QA, +1 Designer

Timeline

9 months to launch

North star

Time to first agent: 30-60 min → 3 min

Agent Studio: build, test, deploy and monitor voice AI agents in one place

Problem

Agora had the infrastructure. It didn’t have a product around it.

A Conversational AI SDK was live on top of Agora’s real-time audio network. Developers still had to stitch speech-to-text, an LLM, and text-to-speech together on their own before hearing a single reply.

Customers coming from call-center work wanted a faster path to production. Most never reached it.

Baseline TTFD

30-60 min

Sign-up to a first working agent before the Agent Studio redesign.

Objective

Shorten the path from evaluation to a first working agent.

Design a studio that improved confidence in moving beyond a demo. The immediate product objective was narrow: make the early lifecycle easier to follow.

Immediate Goals

Configure an agent with less setup overhead.

Reach a live test earlier.

Surface enough deployment and monitoring context to support real implementation decisions.

Business Goal

Reach the first 1M minutes of usage. Help more developers reach value sooner, improve the chances of reaching a first live agent, and create continuity from testing to paid usage.

Research

Finding out where people actually gave up.

One week of discovery during product definition. Five interviews, one designer, a PM sitting in on half of them. This wasn’t a full study but just lean research to commit and test against.

n=5 interviews·1 week·desk + primary·qualitative

Research approach plan

Secondary Research plan

What information will we gatherWhere will we get itWhat questions does it address
Competitor setup flowsLiveKit, Vapi, Retell: docs + trial sign-upsWhat patterns do users expect for setup, testing, production?
Funnel data on ConvAI SDKInternal product analyticsWhere does drop-off happen after sign-up?
Enterprise voice-AI readinessPublic CCaaS reports, vendor marketingWhat production signals do enterprise buyers expect?
.........
.........

Primary Research plan

Who will we studyWhere will we find themWhat topics will we explore
Integration engineersExisting SDK customersWorkflow pain, demo-to-prod blockers, debug needs
Solution architectsPartner accounts evaluating ConvAIArchitecture decisions, cost/scale, observability asks
Product managersInternal stakeholders and customer-side product leadsNon-technical expectations, onboarding clarity, value communication
.........
.........

Methods used

Competitive analysis

Heuristic walkthrough of LiveKit, Vapi, and Retell setup flows to map expected patterns.

Sales + customer asks

B2B-enterprise input from active deals and customer requests.

Interviews

Five exploratory conversations (n=5) with integration engineers, solutions architects, and product managers.

Secondary research

ConvAI SDK funnel analytics and enterprise voice-AI demand reports.

Insights and opportunities

01

Users trusted the agent only after hearing it.

When the first audible test took hours, most users stalled out. Trust arrived through the speaker, not the setup screen.

02

Users wanted production answers before committing.

Cost, scale, and failure modes lived in docs behind setup. Users needed them upfront to justify moving past a demo.

03

Users lost momentum when they did not know the next step after deploy.

Getting an agent live was not enough. People still had to figure out whether to route to telephony, launch on web, or keep testing. Once the product stopped telling them what to do next, they context-switched into docs and lost momentum.

Goals

Four behaviors the studio had to cause.

Research told us where people fell off. These four principles guided us toward what good would look like if we fixed it, and what number to watch for each one.

Activation path·Sign upOnboarding10K Free min testSelect planCard added = activationScale for production

Adoption

A developer can configure a voice agent without writing integration code.

Time to a first configured agent in about 3 minutes.

Activation

A user reaches paid intent by adding a card after the 10,000-minute test stage.

Card-added activation rate at ~3× baseline.

Retention

People came back to monitor, test and configure agent or campaign performance.

Repeat build rate above 25 percent.

Confidence

The product answers production questions before a developer has to ask.

Observability adoption above 60 percent of deployers.

Growth Loop Hypotesis

Each goal was designed to feed the next. When a developer ships their first agent quickly, that agent runs in production. Production usage shows which models, prompts and configurations actually work, so the platform can surface better starting points. The next developer inherits a smarter default. Held as a bet going in, not a proven mechanic.

  1. 01

    First call

    A developer hears their agent respond in the same session.

  2. 02

    More agents

    Low friction means more builds get past the blank canvas.

  3. 03

    Usage signals

    We learn where developers drop off, what they need, and what is missing.

  4. 04

    Better defaults

    Those signals shape which models, prompts and flows we surface first.

  5. 05

    Shorter path

    The next developer inherits a smarter starting point.

Assumptions

Explicit bets going in. Each shaped the studio’s design and became measurable.

AssumptionRationaleKPISuccess Metric

Users will prefer one unified agent builder to stitching services.

Competitive review showed LiveKit/Vapi users valued integrated UX.TTFD~3 minutes to first working agent

Seeing results early will drive continued exploration and deployment.

Interviews revealed users lose momentum during long setup phases.Test-to-deploy conversion3x baseline rate

Observability will increase deployment confidence.

Enterprise buyers flagged cost/performance visibility as gating factor.Observability adoption60%+ of deployers

A node-based editor will facilitate faster automation.

Early exploration suggested advanced automation users might want visual orchestration.Automation setup speedFaster multi-step automation setup

Solution

What Agent Studio shipped with.

Six surfaces covering the full agent lifecycle. Each one maps to a goal from the table above and the metric badge shows which one it drives.

01  ·  Build

Adoption

Agent configurable in a few quick steps.

  • What it is: A guided form for wiring STT, LLM, and TTS with sensible defaults.
  • Why it exists: Removed the stitch-three-vendors-by-hand step that blocked every first-time user in interviews.
Agent builder showing STT, LLM and TTS configuration

Results

What moved, before and after.

Measured over the first 90 days post-GA, against the baseline funnel from the problem section.

Adoption

Time to a first configured agent

Before30 to 60 minutes

After3 minutes

Activation

Card-added activation rate

Before~1 in 10 (pre-studio funnel)

After~3× baseline

Retention

Repeat build rate (same dev, 2nd agent)

BeforeNot tracked pre-studio

After25%+

Confidence

Observability adoption

BeforeSurface didn’t exist

After60%+ of deployers

Where assumptions failed

One of the clearest assumptions behind the original builder was that users would prefer a node-based graph editor for configuring agents. That assumption failed: the graph added setup overhead, buried the next action, and pulled the product away from the adoption outcome in the table above.

To move time to a first configured agent from roughly 30 to 60 minutes down to 3 minutes, the agent builder had to be redesigned around a simpler setup flow with clearer defaults and a faster path to live testing.

BeforeAfter

Before: node-based graph editor.

After: simpler setup flow with clearer defaults.

What stuck

TTFD dropped from roughly 30 to 60 minutes down to 3 minutes. Observability went from nonexistent to a default surface developers opened on their first deploy. The growth loop from the goals section started compounding: more shipped agents fed better templates, and better templates shortened the next developer’s path.

What changed my thinking

It was making the agent audible within the first three minutes of signup. Once developers could hear their agent respond, trust followed. In voice AI, confidence is an audio problem before it is a screen problem.