Gradio WebRTC ⚡️
Stream video and audio in real time with Gradio using WebRTC.
Installation
pip install gradio_webrtc
to use built-in pause detection (see ReplyOnPause), install the vad extra:
pip install gradio_webrtc[vad]
For stop word detection (see ReplyOnStopWords), install the stopword extra:
pip install gradio_webrtc[stopword]
Docs
https://freddyaboulton.github.io/gradio-webrtc/
Examples
🗣️ Audio Input/Output with mini-omni2Build a GPT-4o like experience with mini-omni2, an audio-native LLM. |
🗣️ Talk to ClaudeUse the Anthropic and Play.Ht APIs to have an audio conversation with Claude. |
🗣️ Kyutai MoshiKyutai's moshi is a novel speech-to-speech model for modeling human conversations. |
🗣️ Hello Llama: Stop Word DetectionA code editor built with Llama 3.3 70b that is triggered by the phrase "Hello Llama". Build a Siri-like coding assistant in 100 lines of code! |
🤖 Llama Code EditorCreate and edit HTML pages with just your voice! Powered by SambaNova systems. |
🗣️ Talk to UltravoxTalk to Fixie.AI's audio-native Ultravox LLM with the transformers library. |
🗣️ Talk to Llama 3.2 3bUse the Lepton API to make Llama 3.2 talk back to you! |
🤖 Talk to Qwen2-AudioQwen2-Audio is a SOTA audio-to-text LLM developed by Alibaba. |
📷 Yolov10 Object DetectionRun the Yolov10 model on a user webcam stream in real time! |
📷 Video Object Detection with RT-DETRUpload a video and stream out frames with detected objects (powered by RT-DETR) model. |
🔊 Text-to-Speech with ParlerStream out audio generated by Parler TTS! |
Usage
This is an shortened version of the official usage guide.
To get started with WebRTC streams, all that's needed is to import the WebRTC component from this package and implement its stream event.
Reply on Pause
Typically, you want to run an AI model that generates audio when the user has stopped speaking. This can be done by wrapping a python generator with the ReplyOnPause class
and passing it to the stream event of the WebRTC component.
import gradio as gr
from gradio_webrtc import WebRTC, ReplyOnPause
def response(audio: tuple[int, np.ndarray]): # (1)
"""This function must yield audio frames"""
...
for numpy_array in generated_audio:
yield (sampling_rate, numpy_array, "mono") # (2)
with gr.Blocks() as demo:
gr.HTML(
"""
<h1 style='text-align: center'>
Chat (Powered by WebRTC ⚡️)
</h1>
"""
)
with gr.Column():
with gr.Group():
audio = WebRTC(
mode="send-receive", # (3)
modality="audio",
)
audio.stream(fn=ReplyOnPause(response),
inputs=[audio], outputs=[audio], # (4)
time_limit=60) # (5)
demo.launch()
-
The python generator will receive the entire audio up until the user stopped. It will be a tuple of the form (sampling_rate, numpy array of audio). The array will have a shape of (1, num_samples). You can also pass in additional input components.
-
The generator must yield audio chunks as a tuple of (sampling_rate, numpy audio array). Each numpy audio array must have a shape of (1, num_samples).
-
The
modeandmodalityarguments must be set to"send-receive"and"audio". -
The
WebRTCcomponent must be the first input and output component. -
Set a
time_limitto control how long a conversation will last. If theconcurrency_countis 1 (default), only one conversation will be handled at a time.
Reply On Stopwords
You can configure your AI model to run whenever a set of "stop words" are detected, like "Hey Siri" or "computer", with the ReplyOnStopWords class.
The API is similar to ReplyOnPause with the addition of a stop_words parameter.
import gradio as gr
from gradio_webrtc import WebRTC, ReplyOnPause
def response(audio: tuple[int, np.ndarray]):
"""This function must yield audio frames"""
...
for numpy_array in generated_audio:
yield (sampling_rate, numpy_array, "mono")
with gr.Blocks() as demo:
gr.HTML(
"""
<h1 style='text-align: center'>
Chat (Powered by WebRTC ⚡️)
</h1>
"""
)
with gr.Column():
with gr.Group():
audio = WebRTC(
mode="send",
modality="audio",
)
webrtc.stream(ReplyOnStopWords(generate,
input_sample_rate=16000,
stop_words=["computer"]), # (1)
inputs=[webrtc, history, code],
outputs=[webrtc], time_limit=90,
concurrency_limit=10)
demo.launch()
- The
stop_wordscan be single words or pairs of words. Be sure to include common misspellings of your word for more robust detection, e.g. "llama", "lamma". In my experience, it's best to use two very distinct words like "ok computer" or "hello iris".
Audio Server-To-Clien
To stream only from the server to the client, implement a python generator and pass it to the component's stream event. The stream event must also specify a trigger corresponding to a UI interaction that starts the stream. In this case, it's a button click.
import gradio as gr
from gradio_webrtc import WebRTC
from pydub import AudioSegment
def generation(num_steps):
for _ in range(num_steps):
segment = AudioSegment.from_file("audio_file.wav")
array = np.array(segment.get_array_of_samples()).reshape(1, -1)
yield (segment.frame_rate, array)
with gr.Blocks() as demo:
audio = WebRTC(label="Stream", mode="receive", # (1)
modality="audio")
num_steps = gr.Slider(label="Number of Steps", minimum=1,
maximum=10, step=1, value=5)
button = gr.Button("Generate")
audio.stream(
fn=generation, inputs=[num_steps], outputs=[audio],
trigger=button.click # (2)
)
- Set
mode="receive"to only receive audio from the server. - The
streamevent must take atriggerthat corresponds to the gradio event that starts the stream. In this case, it's the button click.
Video Input/Output Streaming
Set up a video Input/Output stream to continuosly receive webcam frames from the user and run an arbitrary python function to return a modified frame.
import gradio as gr
from gradio_webrtc import WebRTC
def detection(image, conf_threshold=0.3): # (1)
... your detection code here ...
return modified_frame # (2)
with gr.Blocks() as demo:
image = WebRTC(label="Stream", mode="send-receive", modality="video") # (3)
conf_threshold = gr.Slider(
label="Confidence Threshold",
minimum=0.0,
maximum=1.0,
step=0.05,
value=0.30,
)
image.stream(
fn=detection,
inputs=[image, conf_threshold], # (4)
outputs=[image], time_limit=10
)
if __name__ == "__main__":
demo.launch()
- The webcam frame will be represented as a numpy array of shape (height, width, RGB).
- The function must return a numpy array. It can take arbitrary values from other components.
- Set the
modality="video"andmode="send-receive" - The
inputsparameter should be a list where the first element is the WebRTC component. The only output allowed is the WebRTC component.
Server-to-Client Only
Set up a server-to-client stream to stream video from an arbitrary user interaction.
import gradio as gr
from gradio_webrtc import WebRTC
import cv2
def generation():
url = "https://download.tsi.telecom-paristech.fr/gpac/dataset/dash/uhd/mux_sources/hevcds_720p30_2M.mp4"
cap = cv2.VideoCapture(url)
iterating = True
while iterating:
iterating, frame = cap.read()
yield frame # (1)
with gr.Blocks() as demo:
output_video = WebRTC(label="Video Stream", mode="receive", # (2)
modality="video")
button = gr.Button("Start", variant="primary")
output_video.stream(
fn=generation, inputs=None, outputs=[output_video],
trigger=button.click # (3)
)
demo.launch()
- The
streamevent'sfnparameter is a generator function that yields the next frame from the video as a numpy array. - Set
mode="receive"to only receive audio from the server. - The
triggerparameter the gradio event that will trigger the stream. In this case, the button click event.
Additional Outputs
In order to modify other components from within the WebRTC stream, you must yield an instance of AdditionalOutputs and add an on_additional_outputs event to the WebRTC component.
This is common for displaying a multimodal text/audio conversation in a Chatbot UI.
from gradio_webrtc import AdditionalOutputs, WebRTC
def transcribe(audio: tuple[int, np.ndarray],
transformers_convo: list[dict],
gradio_convo: list[dict]):
response = model.generate(**inputs, max_length=256)
transformers_convo.append({"role": "assistant", "content": response})
gradio_convo.append({"role": "assistant", "content": response})
yield AdditionalOutputs(transformers_convo, gradio_convo) # (1)
with gr.Blocks() as demo:
gr.HTML(
"""
<h1 style='text-align: center'>
Talk to Qwen2Audio (Powered by WebRTC ⚡️)
</h1>
"""
)
transformers_convo = gr.State(value=[])
with gr.Row():
with gr.Column():
audio = WebRTC(
label="Stream",
mode="send", # (2)
modality="audio",
)
with gr.Column():
transcript = gr.Chatbot(label="transcript", type="messages")
audio.stream(ReplyOnPause(transcribe),
inputs=[audio, transformers_convo, transcript],
outputs=[audio], time_limit=90)
audio.on_additional_outputs(lambda s,a: (s,a), # (3)
outputs=[transformers_convo, transcript],
queue=False, show_progress="hidden")
demo.launch()
1. Pass your data to `AdditionalOutputs` and yield it.
2. In this case, no audio is being returned, so we set `mode="send"`. However, if we set `mode="send-receive"`, we could also yield generated audio and `AdditionalOutputs`.
3. The `on_additional_outputs` event does not take `inputs`. It's common practice to not run this event on the queue since it is just a quick UI update.
=== "Notes"
1. Pass your data to AdditionalOutputs and yield it.
2. In this case, no audio is being returned, so we set mode="send". However, if we set mode="send-receive", we could also yield generated audio and AdditionalOutputs.
3. The on_additional_outputs event does not take inputs. It's common practice to not run this event on the queue since it is just a quick UI update.
Deployment
When deploying in a cloud environment (like Hugging Face Spaces, EC2, etc), you need to set up a TURN server to relay the WebRTC traffic. The easiest way to do this is to use a service like Twilio.
from twilio.rest import Client
import os
account_sid = os.environ.get("TWILIO_ACCOUNT_SID")
auth_token = os.environ.get("TWILIO_AUTH_TOKEN")
client = Client(account_sid, auth_token)
token = client.tokens.create()
rtc_configuration = {
"iceServers": token.ice_servers,
"iceTransportPolicy": "relay",
}
with gr.Blocks() as demo:
...
rtc = WebRTC(rtc_configuration=rtc_configuration, ...)
...