Contact Sales

All fields are required

FreeSWITCH Heritage: Two Decades of Voice Infrastructure | SignalWire
Built by the FreeSWITCH Team

Trillions of Minutes. Now With an AI Kernel.

The team that built FreeSWITCH extended the same platform with native AI. Same codebase. Same engineers. Two decades of voice infrastructure behind every call.

2.7B
minutes and messages annually
2,000+
companies in production
20 years
continuous development
< 1.2s
typical AI response latency
Why Heritage Matters

Voice Infrastructure Is a Solved Problem (If You Built It)

Media Stack Ownership

The team wrote the media processing engine that powers voice networks worldwide. They can modify the codec pipeline, tune endpointing, and embed AI at the audio frame level. No vendor API stands in the way.

Known Failure Modes

Two decades of production telephony means the edge cases are catalogued: codec negotiation failures, overbuffered audio, overburdened endpoints, overbilled calls. The platform handles them because the team has seen them.

Engine-Level AI

When voice AI emerged, the team embedded orchestration directly in the media stack. The AI kernel coordinates speech recognition, language model inference, and speech synthesis with direct access to the audio stream, eliminating the orchestration overhead of bolt-on pipelines.

End-to-End Changes

When something needs optimization, one team modifies one codebase. No vendor ticket. No API limitation. No waiting for someone else's roadmap.

Bolt-On AI vs. Embedded AI Kernel

Bolt-On Architecture

  • Audio routes out to STT, then LLM, then TTS, then back
  • Each vendor adds latency and a failure mode
  • Barge-in detection happens after network round-trips
  • Error recovery requires per-vendor retry logic
  • State management is your application's problem

Embedded AI Kernel

  • STT, LLM, and TTS orchestrated from inside the media stack
  • 800-1200ms typical response latency, orchestration overhead eliminated
  • Barge-in detection at the media frame level
  • Unified error taxonomy with automatic recovery
  • State is a platform property, not your code

Build a Voice AI Agent

from signalwire_agents import AgentBase
from signalwire_agents.core.function_result import SwaigFunctionResult

class SupportAgent(AgentBase):
    def __init__(self):
        super().__init__(name="Support Agent", route="/support")
        self.prompt_add_section("Instructions",
            body="You are a customer support agent. "
                 "Greet the caller and resolve their issue.")
        self.add_language("English", "en-US", "rime.spore:mistv2")

    @AgentBase.tool(name="check_order")
    def check_order(self, order_id: str):
        """Check the status of a customer order.

        Args:
            order_id: The order ID to look up
        """
        return SwaigFunctionResult(f"Order {order_id}: shipped, ETA April 2nd")

agent = SupportAgent()
agent.run()

From Open-Source Engine to Cloud Platform

1

2006: FreeSWITCH Created

Open-source telephony engine adopted by carriers, contact centers, and UCaaS platforms worldwide.

2

2018: Real-Time WebSocket Control

Persistent session control infrastructure built years before voice AI existed as a category.

3

2020: Cloud Platform Launch

Managed infrastructure layer with state management, orchestration, billing, and security on top of the core engine.

4

2023: AI Kernel Embedded

C-based AI kernel orchestrating STT, LLM, and TTS from inside the media stack.

5

2024: Governed Inference

Production governance with scoped prompts, constrained tools, and enforced state machine transitions.

6

2025: Agents SDK

Python SDK for building, testing, and deploying voice AI agents as standard HTTP microservices.

What It Took to Build the Platform

LayerWhat It Required
State managementPlatform-native call state, conversation context, step transitions
OrchestrationMulti-agent coordination, transfer logic, queue management
AI kernelSTT, LLM, and TTS embedded in the media stack (C)
ReliabilityGlobal redundancy, failover, health monitoring
ObservabilityStructured event streams, per-component latency, error taxonomy
SecurityTenant isolation, encryption, compliance frameworks
API surfaceREST, declarative YAML, real-time WebSocket, function calling, SDKs
💡
FreeSWITCH is open source. Everything built on top is proprietary. A competitor could adopt FreeSWITCH. Replicating the platform layer above it would take years of engineering across state management, orchestration, AI integration, and carrier-grade reliability.

FAQ

What is the relationship between FreeSWITCH and SignalWire?

FreeSWITCH is the open-source telephony engine. SignalWire was built by the same team as a managed cloud platform on top of FreeSWITCH. The AI kernel, state management, orchestration, and developer tools are all proprietary additions.

Why does owning the media stack matter for AI?

The AI kernel sits inside the media stack with direct access to the audio stream. It orchestrates speech recognition, language inference, and audio synthesis without the overhead of routing audio between separate external services. That is why response latency is 800-1200ms instead of 2-4 seconds. With speech-to-speech voice models, latency can be as low as 600ms.

Can I start with a declarative YAML agent and move to the Python SDK later?

Yes. YAML configuration and the Python SDK target the same platform. Start with YAML for rapid prototyping, then adopt the SDK for testing, composition, and programmatic control. Both interfaces compose.

How does pricing work?

$0.16 per minute covers AI processing: STT, LLM, and TTS. Transport (SIP, PSTN) is billed separately at carrier rates. One invoice. No per-token variance, no separate vendor bills.

Trusted by

Build on Two Decades of Voice Infrastructure

One pip install. The same engine that powers carrier-grade telephony worldwide.