diff --git a/README.md b/README.md
index 9134cb5..5ddb403 100644
--- a/README.md
+++ b/README.md
@@ -2,7 +2,8 @@
+
+
+🗣️ Audio Input/Output with mini-omni2
+Build a GPT-4o like experience with mini-omni2, an audio-native LLM.
+
+
+Demo |
+Code
+
+ |
+
+🗣️ Talk to Claude
+Use the Anthropic and Play.Ht APIs to have an audio conversation with Claude.
+
+
+Demo |
+Code
+
+ |
+
+
+
+
+🗣️ Kyutai Moshi
+Kyutai's moshi is a novel speech-to-speech model for modeling human conversations.
+
+
+Demo |
+Code
+
+ |
+
+🗣️ Hello Llama: Stop Word Detection
+A code editor built with Llama 3.3 70b that is triggered by the phrase "Hello Llama". Build a Siri-like coding assistant in 100 lines of code!
+
+
+Demo |
+Code
+
+ |
+
+
+
+
+🤖 Llama Code Editor
+Create and edit HTML pages with just your voice! Powered by SambaNova systems.
+
+
+Demo |
+Code
+
+ |
+
+🗣️ Talk to Ultravox
+Talk to Fixie.AI's audio-native Ultravox LLM with the transformers library.
+
+
+Demo |
+Code
+
+ |
+
+
+
+
+🗣️ Talk to Llama 3.2 3b
+Use the Lepton API to make Llama 3.2 talk back to you!
+
+
+Demo |
+Code
+
+ |
+
+🤖 Talk to Qwen2-Audio
+Qwen2-Audio is a SOTA audio-to-text LLM developed by Alibaba.
+
+
+Demo |
+Code
+
+ |
+
+
+
+
+📷 Yolov10 Object Detection
+Run the Yolov10 model on a user webcam stream in real time!
+
+
+Demo |
+Code
+
+ |
+
+📷 Video Object Detection with RT-DETR
+Upload a video and stream out frames with detected objects (powered by RT-DETR) model.
+
+Demo |
+Code
+
+ |
+
+
+
+
+🔊 Text-to-Speech with Parler
+Stream out audio generated by Parler TTS!
+
+Demo |
+Code
+
+ |
+
+ |
+
+
## Usage
-The WebRTC component supports the following three use cases:
-1. [Streaming video from the user webcam to the server and back](#h-streaming-video-from-the-user-webcam-to-the-server-and-back)
-2. [Streaming Video from the server to the client](#h-streaming-video-from-the-server-to-the-client)
-3. [Streaming Audio from the server to the client](#h-streaming-audio-from-the-server-to-the-client)
-4. [Streaming Audio from the client to the server and back (conversational AI)](#h-conversational-ai)
+This is an shortened version of the official [usage guide](https://freddyaboulton.github.io/gradio-webrtc/user-guide/).
+To get started with WebRTC streams, all that's needed is to import the `WebRTC` component from this package and implement its `stream` event.
-## Streaming Video from the User Webcam to the Server and Back
+### Reply on Pause
-```python
+Typically, you want to run an AI model that generates audio when the user has stopped speaking. This can be done by wrapping a python generator with the `ReplyOnPause` class
+and passing it to the `stream` event of the `WebRTC` component.
+
+```py
import gradio as gr
-from gradio_webrtc import WebRTC
+from gradio_webrtc import WebRTC, ReplyOnPause
-
-def detection(image, conf_threshold=0.3):
- ... your detection code here ...
+def response(audio: tuple[int, np.ndarray]): # (1)
+ """This function must yield audio frames"""
+ ...
+ for numpy_array in generated_audio:
+ yield (sampling_rate, numpy_array, "mono") # (2)
with gr.Blocks() as demo:
- image = WebRTC(label="Stream", mode="send-receive", modality="video")
- conf_threshold = gr.Slider(
- label="Confidence Threshold",
- minimum=0.0,
- maximum=1.0,
- step=0.05,
- value=0.30,
+ gr.HTML(
+ """
+