u/ord_phreaker

▲ 1 r/tts

Improved Telnyx Ultra voices now out

We just shipped an upgrade to Ultra Voices on Telnyx.

The big thing: the voices sound more natural now, especially in the parts where TTS usually breaks down.

Pauses feel less awkward.

Delivery is less flat.

The voice handles longer sentences better.

And the timing feels closer to an actual phone conversation.

This matters a lot for voice agents because the voice is usually the first thing users judge.

You can have solid STT, good routing, clean prompts, and fast tool calls, but if the voice sounds robotic or weirdly paced, the whole experience feels off.

We put together a quick before and after video using the same voice so the difference is easier to hear.

The improved Ultra Voices are live now on Telnyx.

Give it a shot in the Telnyx portal

reddit.com
u/ord_phreaker — 7 hours ago
▲ 3 r/Telnyx

New in Telnyx: Upgraded Telnyx Ultra Voice

We just shipped an upgrade to Ultra Voices on Telnyx.

The big thing: the voices sound more natural now, especially in the parts where TTS usually breaks down.

Pauses feel less awkward.

Delivery is less flat.

The voice handles longer sentences better.

And the timing feels closer to an actual phone conversation.

This matters a lot for voice agents because the voice is usually the first thing users judge.

You can have solid STT, good routing, clean prompts, and fast tool calls, but if the voice sounds robotic or weirdly paced, the whole experience feels off.

We put together a quick before and after video using the same voice so the difference is easier to hear.

The improved Ultra Voices are live now on Telnyx.

video

reddit.com
u/ord_phreaker — 1 day ago

I’ve been spending a lot of time around production voice AI deployments, and the same patterns keep showing up.

The hard parts usually aren’t the voice model by itself. They’re the system around it.

A few lessons that seem to matter most:

  1. Start with one call type. General support agents usually become vague fast.
  2. Measure resolved calls, not answered calls.
  3. Track time to first audio and full turn latency separately.
  4. Test on real phone audio, not only browser audio.
  5. Word error rate is an incomplete metric. Entity capture matters more.
  6. Let callers interrupt. Turn-taking is where a lot of “AI feel” breaks.
  7. Keep tool responses short and structured.
  8. Confirm before write actions.
  9. Build eval sets from real calls.
  10. Treat handoff as part of the product, not a failure path.
  11. Separate model failures from workflow failures.
  12. Review failed calls every week.

The biggest shift for me is that voice agents are judged inside a live interaction. A caller notices latency, repetition, awkward pauses, bad escalation, and missing context immediately.

So the production question becomes less “can this agent talk?” and more:

  • Can it complete the workflow?
  • Can it recover from messy audio?
  • Can it use the right tools?
  • Can it hand off cleanly?
  • Can the team improve it every week?

For teams building voice agents right now, what has been harder than expected?

reddit.com
u/ord_phreaker — 8 days ago

I’ve been spending a lot of time around production voice AI deployments, and the same patterns keep showing up.

The hard parts usually aren’t the voice model by itself. They’re the system around it.

A few lessons that seem to matter most:

  1. Start with one call type. General support agents usually become vague fast.
  2. Measure resolved calls, not answered calls.
  3. Track time to first audio and full turn latency separately.
  4. Test on real phone audio, not only browser audio.
  5. Word error rate is an incomplete metric. Entity capture matters more.
  6. Let callers interrupt. Turn-taking is where a lot of “AI feel” breaks.
  7. Keep tool responses short and structured.
  8. Confirm before write actions.
  9. Build eval sets from real calls.
  10. Treat handoff as part of the product, not a failure path.
  11. Separate model failures from workflow failures.
  12. Review failed calls every week.

The biggest shift for me is that voice agents are judged inside a live interaction. A caller notices latency, repetition, awkward pauses, bad escalation, and missing context immediately.

So the production question becomes less “can this agent talk?” and more:

  • Can it complete the workflow?
  • Can it recover from messy audio?
  • Can it use the right tools?
  • Can it hand off cleanly?
  • Can the team improve it every week?

For teams building voice agents right now, what has been harder than expected?

reddit.com
u/ord_phreaker — 8 days ago

I’ve been spending a lot of time around production voice AI deployments, and the same patterns keep showing up.

The hard parts usually aren’t the voice model by itself. They’re the system around it.

A few lessons that seem to matter most:

  1. Start with one call type. General support agents usually become vague fast.
  2. Measure resolved calls, not answered calls.
  3. Track time to first audio and full turn latency separately.
  4. Test on real phone audio, not only browser audio.
  5. Word error rate is an incomplete metric. Entity capture matters more.
  6. Let callers interrupt. Turn-taking is where a lot of “AI feel” breaks.
  7. Keep tool responses short and structured.
  8. Confirm before write actions.
  9. Build eval sets from real calls.
  10. Treat handoff as part of the product, not a failure path.
  11. Separate model failures from workflow failures.
  12. Review failed calls every week.

The biggest shift for me is that voice agents are judged inside a live interaction. A caller notices latency, repetition, awkward pauses, bad escalation, and missing context immediately.

So the production question becomes less “can this agent talk?” and more:

  • Can it complete the workflow?
  • Can it recover from messy audio?
  • Can it use the right tools?
  • Can it hand off cleanly?
  • Can the team improve it every week?

For teams building voice agents right now, what has been harder than expected?

reddit.com
u/ord_phreaker — 8 days ago

I’ve been spending a lot of time around production voice AI deployments, and the same patterns keep showing up.

The hard parts usually aren’t the voice model by itself. They’re the system around it.

A few lessons that seem to matter most:

  1. Start with one call type. General support agents usually become vague fast.
  2. Measure resolved calls, not answered calls.
  3. Track time to first audio and full turn latency separately.
  4. Test on real phone audio, not only browser audio.
  5. Word error rate is an incomplete metric. Entity capture matters more.
  6. Let callers interrupt. Turn-taking is where a lot of “AI feel” breaks.
  7. Keep tool responses short and structured.
  8. Confirm before write actions.
  9. Build eval sets from real calls.
  10. Treat handoff as part of the product, not a failure path.
  11. Separate model failures from workflow failures.
  12. Review failed calls every week.

The biggest shift for me is that voice agents are judged inside a live interaction. A caller notices latency, repetition, awkward pauses, bad escalation, and missing context immediately.

So the production question becomes less “can this agent talk?” and more:

  • Can it complete the workflow?
  • Can it recover from messy audio?
  • Can it use the right tools?
  • Can it hand off cleanly?
  • Can the team improve it every week?

For teams building voice agents right now, what has been harder than expected?

reddit.com
u/ord_phreaker — 8 days ago

I’ve been spending a lot of time around production voice AI deployments, and the same patterns keep showing up.

The hard parts usually aren’t the voice model by itself. They’re the system around it.

A few lessons that seem to matter most:

  1. Start with one call type. General support agents usually become vague fast.

  2. Measure resolved calls, not answered calls.

  3. Track time to first audio and full turn latency separately.

  4. Test on real phone audio, not only browser audio.

  5. Word error rate is an incomplete metric. Entity capture matters more.

  6. Let callers interrupt. Turn-taking is where a lot of “AI feel” breaks.

  7. Keep tool responses short and structured.

  8. Confirm before write actions.

  9. Build eval sets from real calls.

  10. Treat handoff as part of the product, not a failure path.

  11. Separate model failures from workflow failures.

  12. Review failed calls every week.

The biggest shift for me is that voice agents are judged inside a live interaction. A caller notices latency, repetition, awkward pauses, bad escalation, and missing context immediately.

So the production question becomes less “can this agent talk?” and more:

  • Can it complete the workflow?
  • Can it recover from messy audio?
  • Can it use the right tools?
  • Can it hand off cleanly?
  • Can the team improve it every week?

For teams building voice agents right now, what has been harder than expected?

reddit.com
u/ord_phreaker — 8 days ago
▲ 5 r/Telnyx

We’ve added multi-participant support for Telnyx AI Assistants.

With this your AI assistant can now bring another person into a live call, understand who is speaking, and stay quiet when the humans are talking to each other.

That opens up workflows that were awkward in a 1:1 voice AI call:

- A customer asks to speak with a specialist, and the assistant brings them in.

- A sales rep pulls in an SE while the AI stays on the call to track next steps.

- A scheduling assistant invites the other person, waits while both sides compare availability, then books the meeting.

- A care coordination assistant connects a patient and provider, then updates records after the call.

What makes it cool is that the assistant does not need to treat every human sentence as a prompt.

With Skip Turn, it can stay silent for a turn when participants are talking to each other, then respond again when someone addresses it.

This is live in Telnyx AI Assistants now.

Docs: https://developers.telnyx.com/docs/inference/ai-assistants/multi-participant-calls

Watch a demo of this in action here: https://youtu.be/aSWxmtD0OCs

u/ord_phreaker — 9 days ago