Skip to main content
Pipecat’s client SDKs connect your users to Pipecat agents running on the server. They handle the real-time media layer — microphone and camera access, audio playback, transport connections, and the event stream from your bot — so you can focus on building your application.

Quickstart

Build and run your first Pipecat voice application

Available SDKs

JavaScript

Core SDK for web applications

React

Hooks and components for React apps

React Native

iOS and Android via React Native

iOS

Native Swift SDK for iOS

Android

Native Kotlin SDK for Android

C++

Native SDK for desktop and embedded

How It Fits Together

Pipecat’s architecture is split between server and client. The server runs your AI pipeline — speech recognition, LLM, text-to-speech — and the client connects your user to it over a real-time transport.
The client SDKs implement the RTVI standard — an open protocol for real-time AI inference — which means they work with any RTVI-compatible server, not just Pipecat.

What the SDKs Handle

  • Transport management — establishing, maintaining, and tearing down connections
  • Media capture — microphone and camera device access and streaming
  • Audio playback — rendering bot audio output
  • Session state — tracking connection state and bot readiness
  • Event stream — callbacks and events for transcriptions, bot speaking, errors, and more
  • Messaging — sending custom messages to your bot and handling responses

Ready to Build?

Quickstart

Get a voice conversation running in minutes

Core Concepts

Understand transports, sessions, events, and media

API Reference

Full reference for all SDK methods, hooks, and callbacks

Examples

Complete example applications across all platforms