mirror of
https://github.com/HumanAIGC-Engineering/gradio-webrtc.git
synced 2026-02-05 18:09:23 +08:00
@@ -37,6 +37,7 @@ from numpy import typing as npt
|
||||
|
||||
from fastrtc.utils import (
|
||||
AdditionalOutputs,
|
||||
CloseStream,
|
||||
DataChannel,
|
||||
WebRTCError,
|
||||
create_message,
|
||||
@@ -54,9 +55,14 @@ VideoNDArray: TypeAlias = Union[
|
||||
]
|
||||
|
||||
VideoEmitType = (
|
||||
VideoNDArray | tuple[VideoNDArray, AdditionalOutputs] | AdditionalOutputs
|
||||
VideoNDArray
|
||||
| tuple[VideoNDArray, AdditionalOutputs]
|
||||
| tuple[VideoNDArray, CloseStream]
|
||||
| AdditionalOutputs
|
||||
| CloseStream
|
||||
)
|
||||
VideoEventHandler = Callable[[npt.ArrayLike], VideoEmitType]
|
||||
VideoEventGenerator = Generator[VideoEmitType, None, None]
|
||||
VideoEventHandler = Callable[[npt.ArrayLike], VideoEmitType | VideoEventGenerator]
|
||||
|
||||
|
||||
@dataclass
|
||||
@@ -172,6 +178,12 @@ class VideoCallback(VideoStreamTrack):
|
||||
|
||||
args = self.add_frame_to_payload(cast(list, self.latest_args), frame_array)
|
||||
array, outputs = split_output(self.event_handler(*args))
|
||||
if isinstance(outputs, CloseStream):
|
||||
cast(DataChannel, self.channel).send(
|
||||
create_message("end_stream", outputs.msg)
|
||||
)
|
||||
self.stop()
|
||||
return None
|
||||
if (
|
||||
isinstance(outputs, AdditionalOutputs)
|
||||
and self.set_additional_outputs
|
||||
@@ -444,6 +456,12 @@ class VideoStreamHandler_(VideoCallback):
|
||||
):
|
||||
self.set_additional_outputs(outputs)
|
||||
self.channel.send(create_message("fetch_output", []))
|
||||
if isinstance(outputs, CloseStream):
|
||||
cast(DataChannel, self.channel).send(
|
||||
create_message("end_stream", outputs.msg)
|
||||
)
|
||||
self.stop()
|
||||
return
|
||||
if array is None and self.mode == "send":
|
||||
return
|
||||
|
||||
@@ -586,6 +604,12 @@ class AudioCallback(AudioStreamTrack):
|
||||
await self.start()
|
||||
|
||||
frame = await self.queue.get()
|
||||
if isinstance(frame, CloseStream):
|
||||
cast(DataChannel, self.channel).send(
|
||||
create_message("end_stream", frame.msg)
|
||||
)
|
||||
self.stop()
|
||||
return
|
||||
logger.debug("frame %s", frame)
|
||||
|
||||
data_time = frame.time
|
||||
@@ -675,6 +699,12 @@ class ServerToClientVideo(VideoStreamTrack):
|
||||
)
|
||||
try:
|
||||
next_array, outputs = split_output(next(self.generator))
|
||||
if isinstance(outputs, CloseStream):
|
||||
cast(DataChannel, self.channel).send(
|
||||
create_message("end_stream", outputs.msg)
|
||||
)
|
||||
self.stop()
|
||||
return
|
||||
if (
|
||||
isinstance(outputs, AdditionalOutputs)
|
||||
and self.set_additional_outputs
|
||||
@@ -770,6 +800,12 @@ class ServerToClientAudio(AudioStreamTrack):
|
||||
|
||||
await self.start()
|
||||
data = await self.queue.get()
|
||||
if isinstance(data, CloseStream):
|
||||
cast(DataChannel, self.channel).send(
|
||||
create_message("end_stream", data.msg)
|
||||
)
|
||||
self.stop()
|
||||
return
|
||||
if data is None:
|
||||
self.stop()
|
||||
return
|
||||
|
||||
Reference in New Issue
Block a user