mirror of
https://gitlab.freedesktop.org/gstreamer/gst-plugins-rs.git
synced 2024-11-25 13:01:07 +00:00
34b791ff5e
This commit adds support for raw payloads such as L24 audio to `webrtcsink` & `webrtcsrc`. Most changes take place within the `Codec` helper structure: * A `Codec` can now advertise a depayloader. This also ensures that a format not only can be decoded when necessary, but it can also be depayloaded in the first place. * It is possible to declare raw `Codec`s, meaning that their caps are compatible with a payloader and a depayloader without the need for an encoder and decoder. * Previous accessor `has_decoder` was renamed as `can_be_received` to account for codecs which can be handled by an available depayloader with or without the need for a decoder. * New codecs were added for the following formats: * L24, L16, L8 audio. * RAW video. The `webrtc-precise-sync` examples were updated to demonstrate streaming of raw audio or video. Part-of: <https://gitlab.freedesktop.org/gstreamer/gst-plugins-rs/-/merge_requests/1501>
479 lines
16 KiB
Rust
479 lines
16 KiB
Rust
use anyhow::{anyhow, bail, Context};
|
|
use futures::prelude::*;
|
|
use gst::prelude::*;
|
|
use gst_rtp::prelude::*;
|
|
use gstrswebrtc::utils::Codecs;
|
|
use itertools::Itertools;
|
|
use tracing::{debug, error, info};
|
|
use url::Url;
|
|
|
|
const VIDEO_PATTERNS: [&str; 3] = ["ball", "smpte", "snow"];
|
|
|
|
#[derive(Debug, Default, Clone, clap::Parser)]
|
|
struct Args {
|
|
#[clap(long, help = "Clock type", default_value = "ntp")]
|
|
pub clock: Clock,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Maximum duration in seconds to wait for clock synchronization",
|
|
default_value = "5"
|
|
)]
|
|
pub clock_sync_timeout: u64,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Enable RFC 7273 PTP or NTP clock & RTP/clock offset signalling"
|
|
)]
|
|
pub do_clock_signalling: bool,
|
|
|
|
#[clap(long, help = "NTP server host", default_value = "pool.ntp.org")]
|
|
pub ntp_server: String,
|
|
|
|
#[clap(long, help = "PTP domain", default_value = "0")]
|
|
pub ptp_domain: u32,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Number of audio streams. Use 0 to disable audio",
|
|
default_value = "1"
|
|
)]
|
|
pub audio_streams: usize,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Number of video streams. Use 0 to disable video",
|
|
default_value = "1"
|
|
)]
|
|
pub video_streams: usize,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Audio codecs that will be proposed in the SDP (ex. 'L24', defaults to ['OPUS']). Accepts several occurrences."
|
|
)]
|
|
pub audio_codecs: Vec<String>,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Use specific audio caps (ex. 'audio/x-raw,rate=48000,channels=2')"
|
|
)]
|
|
pub audio_caps: Option<String>,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Video codecs that will be proposed in the SDP (ex. 'RAW', defaults to ['VP8', 'H264', 'VP9', 'H265']). Accepts several occurrences."
|
|
)]
|
|
pub video_codecs: Vec<String>,
|
|
|
|
#[clap(
|
|
long,
|
|
help = "Use specific video caps (ex. 'video/x-raw,format=I420,width=400')"
|
|
)]
|
|
pub video_caps: Option<String>,
|
|
|
|
#[clap(long, help = "Use RFC 6051 64-bit NTP timestamp RTP header extension.")]
|
|
pub enable_rapid_sync: bool,
|
|
|
|
#[clap(long, help = "Signalling server host", default_value = "localhost")]
|
|
pub server: String,
|
|
|
|
#[clap(long, help = "Signalling server port", default_value = "8443")]
|
|
pub port: u32,
|
|
|
|
#[clap(long, help = "use tls")]
|
|
pub use_tls: bool,
|
|
}
|
|
|
|
impl Args {
|
|
fn scheme(&self) -> &str {
|
|
if self.use_tls {
|
|
"wss"
|
|
} else {
|
|
"ws"
|
|
}
|
|
}
|
|
|
|
async fn get_synced_clock(&self) -> anyhow::Result<gst::Clock> {
|
|
debug!("Syncing to {:?}", self.clock);
|
|
|
|
// Create the requested clock and wait for synchronization.
|
|
let clock = match self.clock {
|
|
Clock::System => gst::SystemClock::obtain(),
|
|
Clock::Ntp => gst_net::NtpClock::new(None, &self.ntp_server, 123, gst::ClockTime::ZERO)
|
|
.upcast::<gst::Clock>(),
|
|
Clock::Ptp => {
|
|
gst_net::PtpClock::init(None, &[])?;
|
|
gst_net::PtpClock::new(None, self.ptp_domain)?.upcast()
|
|
}
|
|
};
|
|
|
|
let clock_sync_timeout = gst::ClockTime::from_seconds(self.clock_sync_timeout);
|
|
let clock =
|
|
tokio::task::spawn_blocking(move || -> Result<gst::Clock, gst::glib::BoolError> {
|
|
clock.wait_for_sync(clock_sync_timeout)?;
|
|
Ok(clock)
|
|
})
|
|
.await
|
|
.with_context(|| format!("Syncing to {:?}", self.clock))?
|
|
.with_context(|| format!("Syncing to {:?}", self.clock))?;
|
|
|
|
info!("Synced to {:?}", self.clock);
|
|
|
|
Ok(clock)
|
|
}
|
|
}
|
|
|
|
#[derive(Copy, Clone, Debug, Default, PartialEq, Eq, clap::ValueEnum)]
|
|
pub enum Clock {
|
|
#[default]
|
|
Ntp,
|
|
Ptp,
|
|
System,
|
|
}
|
|
|
|
#[derive(Debug, Default)]
|
|
struct App {
|
|
args: Args,
|
|
pipeline: Option<gst::Pipeline>,
|
|
}
|
|
|
|
impl App {
|
|
fn new(args: Args) -> Self {
|
|
App {
|
|
args,
|
|
..Default::default()
|
|
}
|
|
}
|
|
|
|
#[inline(always)]
|
|
fn pipeline(&self) -> &gst::Pipeline {
|
|
self.pipeline.as_ref().expect("Set in prepare")
|
|
}
|
|
|
|
async fn prepare_and_run(&mut self) -> anyhow::Result<()> {
|
|
self.prepare().await.context("Preparing")?;
|
|
self.run().await.context("Running")?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn prepare(&mut self) -> anyhow::Result<()> {
|
|
use std::str::FromStr;
|
|
|
|
debug!("Preparing");
|
|
|
|
self.pipeline = Some(gst::Pipeline::new());
|
|
self.pipeline()
|
|
.use_clock(Some(&self.args.get_synced_clock().await?));
|
|
|
|
// Set the base time of the pipeline statically to zero so that running
|
|
// time and clock time are the same and timeoverlay can be used to render
|
|
// the clock time over the video frames.
|
|
//
|
|
// This is needed for no other reasons.
|
|
self.pipeline().set_base_time(gst::ClockTime::ZERO);
|
|
self.pipeline().set_start_time(gst::ClockTime::NONE);
|
|
|
|
let signaller_url = Url::parse(&format!(
|
|
"{}://{}:{}",
|
|
self.args.scheme(),
|
|
self.args.server,
|
|
self.args.port,
|
|
))?;
|
|
|
|
let webrtcsink = gst::ElementFactory::make("webrtcsink")
|
|
// See:
|
|
// * https://gitlab.freedesktop.org/gstreamer/gst-plugins-rs/-/issues/497
|
|
// * https://gitlab.freedesktop.org/gstreamer/gstreamer/-/issues/3301
|
|
.property("do-fec", false)
|
|
.property("do-clock-signalling", self.args.do_clock_signalling)
|
|
.build()
|
|
.context("Creating webrtcsink")?;
|
|
|
|
self.pipeline().add(&webrtcsink).unwrap();
|
|
|
|
let signaller = webrtcsink.property::<gst::glib::Object>("signaller");
|
|
signaller.set_property("uri", signaller_url.as_str());
|
|
|
|
signaller.connect("webrtcbin-ready", false, |args| {
|
|
let webrtcbin = args[2].get::<gst::Element>().unwrap();
|
|
|
|
let rtpbin = webrtcbin
|
|
.downcast_ref::<gst::Bin>()
|
|
.unwrap()
|
|
.by_name("rtpbin")
|
|
.unwrap();
|
|
|
|
// Use local pipeline clock time as RTP NTP time source instead of using
|
|
// the local wallclock time converted to the NTP epoch.
|
|
rtpbin.set_property_from_str("ntp-time-source", "clock-time");
|
|
|
|
// Use the capture time instead of the send time for the RTP / NTP timestamp
|
|
// mapping. The difference between the two options is the capture/encoder/etc.
|
|
// latency that is introduced before sending.
|
|
rtpbin.set_property("rtcp-sync-send-time", false);
|
|
|
|
None
|
|
});
|
|
|
|
webrtcsink.connect("encoder-setup", true, |args| {
|
|
let enc = args[3].get::<gst::Element>().unwrap();
|
|
if enc.is::<gst_audio::AudioEncoder>() {
|
|
// Make sure the audio encoder tracks upstream timestamps.
|
|
enc.set_property("perfect-timestamp", false);
|
|
}
|
|
|
|
Some(true.to_value())
|
|
});
|
|
|
|
if self.args.enable_rapid_sync {
|
|
webrtcsink.connect("payloader-setup", false, |args| {
|
|
let payloader = args[3].get::<gst::Element>().unwrap();
|
|
|
|
// Add RFC6051 64-bit NTP timestamp RTP header extension.
|
|
let hdr_ext = gst_rtp::RTPHeaderExtension::create_from_uri(
|
|
"urn:ietf:params:rtp-hdrext:ntp-64",
|
|
)
|
|
.expect("Creating NTP 64-bit RTP header extension");
|
|
hdr_ext.set_id(1);
|
|
payloader.emit_by_name::<()>("add-extension", &[&hdr_ext]);
|
|
|
|
Some(true.into())
|
|
});
|
|
}
|
|
|
|
if !self.args.audio_codecs.is_empty() {
|
|
let mut audio_caps = gst::Caps::new_empty();
|
|
for codec in self.args.audio_codecs.iter() {
|
|
audio_caps.merge(
|
|
Codecs::audio_codecs()
|
|
.find(|c| &c.name == codec)
|
|
.ok_or_else(|| {
|
|
anyhow!(
|
|
"Unknown audio codec {codec}. Valid values are: {}",
|
|
Codecs::audio_codecs().map(|c| c.name.as_str()).join(", ")
|
|
)
|
|
})?
|
|
.caps
|
|
.clone(),
|
|
);
|
|
}
|
|
|
|
webrtcsink.set_property("audio-caps", audio_caps);
|
|
}
|
|
|
|
for idx in 0..self.args.audio_streams {
|
|
let audiosrc = gst::ElementFactory::make("audiotestsrc")
|
|
.property("is-live", true)
|
|
.property("freq", (idx + 1) as f64 * 440.0)
|
|
.property("volume", 0.2f64)
|
|
.build()
|
|
.context("Creating audiotestsrc")?;
|
|
self.pipeline().add(&audiosrc).context("Adding audiosrc")?;
|
|
|
|
if let Some(ref caps) = self.args.audio_caps {
|
|
audiosrc
|
|
.link_pads_filtered(
|
|
None,
|
|
&webrtcsink,
|
|
Some("audio_%u"),
|
|
&gst::Caps::from_str(caps).context("Parsing audio caps")?,
|
|
)
|
|
.context("Linking audiosrc")?;
|
|
} else {
|
|
audiosrc
|
|
.link_pads(None, &webrtcsink, Some("audio_%u"))
|
|
.context("Linking audiosrc")?;
|
|
}
|
|
}
|
|
|
|
if !self.args.video_codecs.is_empty() {
|
|
let mut video_caps = gst::Caps::new_empty();
|
|
for codec in self.args.video_codecs.iter() {
|
|
video_caps.merge(
|
|
Codecs::video_codecs()
|
|
.find(|c| &c.name == codec)
|
|
.ok_or_else(|| {
|
|
anyhow!(
|
|
"Unknown video codec {codec}. Valid values are: {}",
|
|
Codecs::video_codecs().map(|c| c.name.as_str()).join(", ")
|
|
)
|
|
})?
|
|
.caps
|
|
.clone(),
|
|
);
|
|
}
|
|
|
|
webrtcsink.set_property("video-caps", video_caps);
|
|
}
|
|
|
|
let raw_video_caps = {
|
|
let mut raw_video_caps = if let Some(ref video_caps) = self.args.video_caps {
|
|
gst::Caps::from_str(video_caps).context("Parsing video caps")?
|
|
} else {
|
|
gst::Caps::new_empty_simple("video/x-raw")
|
|
};
|
|
|
|
// If no width / height are specified, set something big enough
|
|
let caps_mut = raw_video_caps.make_mut();
|
|
let s = caps_mut.structure_mut(0).expect("created above");
|
|
match (s.get::<i32>("width").ok(), s.get::<i32>("height").ok()) {
|
|
(None, None) => {
|
|
s.set("width", 800i32);
|
|
s.set("height", 600i32);
|
|
}
|
|
(Some(width), None) => s.set("height", 3 * width / 4),
|
|
(None, Some(height)) => s.set("width", 4 * height / 3),
|
|
_ => (),
|
|
}
|
|
|
|
raw_video_caps
|
|
};
|
|
|
|
for idx in 0..self.args.video_streams {
|
|
let videosrc = gst::ElementFactory::make("videotestsrc")
|
|
.property("is-live", true)
|
|
.property_from_str("pattern", VIDEO_PATTERNS[idx % VIDEO_PATTERNS.len()])
|
|
.build()
|
|
.context("Creating videotestsrc")?;
|
|
let video_overlay = gst::ElementFactory::make("timeoverlay")
|
|
.property_from_str("time-mode", "running-time")
|
|
.build()
|
|
.context("Creating timeoverlay")?;
|
|
|
|
self.pipeline()
|
|
.add_many([&videosrc, &video_overlay])
|
|
.expect("adding video elements");
|
|
|
|
videosrc
|
|
.link_filtered(&video_overlay, &raw_video_caps)
|
|
.context("Linking videosrc to timeoverlay")?;
|
|
|
|
video_overlay
|
|
.link_pads(None, &webrtcsink, Some("video_%u"))
|
|
.context("Linking video overlay")?;
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn run(&mut self) -> anyhow::Result<()> {
|
|
debug!("Running");
|
|
|
|
let bus = self.pipeline().bus().context("Getting the pipeline bus")?;
|
|
let mut bus_stream = bus.stream();
|
|
|
|
self.pipeline()
|
|
.call_async_future(|pipeline| pipeline.set_state(gst::State::Playing))
|
|
.await
|
|
.context("Setting pipeline to Playing")?;
|
|
|
|
while let Some(bus_msg) = bus_stream.next().await {
|
|
use gst::MessageView::*;
|
|
|
|
match bus_msg.view() {
|
|
Error(msg) => {
|
|
let err = msg.error();
|
|
let src_name = msg.src().map(|src| src.name());
|
|
|
|
bail!(
|
|
"Element {} error message: {err:#}",
|
|
src_name.as_deref().unwrap_or("UNKNOWN"),
|
|
);
|
|
}
|
|
Latency(msg) => {
|
|
info!(
|
|
"Latency requirements have changed for element {}",
|
|
msg.src()
|
|
.map(|src| src.name())
|
|
.as_deref()
|
|
.unwrap_or("UNKNOWN"),
|
|
);
|
|
if let Err(err) = self.pipeline().recalculate_latency() {
|
|
error!(%err, "Error recalculating latency");
|
|
}
|
|
}
|
|
_ => (),
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Tears this `App` down and deallocates all its resources by consuming `self`.
|
|
async fn teardown(mut self) {
|
|
debug!("Tearing down");
|
|
|
|
if let Some(pipeline) = self.pipeline.take() {
|
|
let _ = pipeline
|
|
.call_async_future(|pipeline| pipeline.set_state(gst::State::Null))
|
|
.await;
|
|
}
|
|
}
|
|
}
|
|
|
|
#[tokio::main]
|
|
async fn main() -> anyhow::Result<()> {
|
|
use clap::Parser;
|
|
use tracing_subscriber::prelude::*;
|
|
|
|
let args = Args::parse();
|
|
|
|
tracing_log::LogTracer::init().context("Setting logger")?;
|
|
let env_filter = tracing_subscriber::EnvFilter::try_from_env("WEBRTC_PRECISE_SYNC_SEND_LOG")
|
|
.unwrap_or_else(|_| tracing_subscriber::EnvFilter::new("info"));
|
|
let fmt_layer = tracing_subscriber::fmt::layer()
|
|
.with_thread_ids(true)
|
|
.with_target(true)
|
|
.with_span_events(
|
|
tracing_subscriber::fmt::format::FmtSpan::NEW
|
|
| tracing_subscriber::fmt::format::FmtSpan::CLOSE,
|
|
);
|
|
let subscriber = tracing_subscriber::Registry::default()
|
|
.with(env_filter)
|
|
.with(fmt_layer);
|
|
tracing::subscriber::set_global_default(subscriber).context("Setting tracing subscriber")?;
|
|
|
|
gst::init()?;
|
|
gstrswebrtc::plugin_register_static()?;
|
|
gstrsrtp::plugin_register_static()?;
|
|
|
|
debug!("Starting");
|
|
|
|
let mut res = Ok(());
|
|
let mut app = App::new(args);
|
|
|
|
{
|
|
let ctrl_c = tokio::signal::ctrl_c().fuse();
|
|
tokio::pin!(ctrl_c);
|
|
|
|
let prepare_and_run = app.prepare_and_run().fuse();
|
|
tokio::pin!(prepare_and_run);
|
|
|
|
futures::select! {
|
|
_ctrl_c_res = ctrl_c => {
|
|
info!("Shutting down due to user request");
|
|
}
|
|
app_res = prepare_and_run => {
|
|
if let Err(ref err) = app_res {
|
|
error!("Shutting down due to application error: {err:#}");
|
|
} else {
|
|
info!("Shutting down due to application termination");
|
|
}
|
|
|
|
res = app_res;
|
|
}
|
|
}
|
|
}
|
|
|
|
app.teardown().await;
|
|
|
|
debug!("Quitting");
|
|
|
|
unsafe {
|
|
// Needed for certain tracers to write data
|
|
gst::deinit();
|
|
}
|
|
|
|
res
|
|
}
|