Fix code style with cargo fmt

This commit is contained in:
Daniel Vilar 2018-09-18 11:53:12 +02:00
parent 6db4929e08
commit cf8b7db9bd
4 changed files with 816 additions and 747 deletions

View file

@ -22,14 +22,14 @@ extern crate gstreamer_video as gst_video;
#[macro_use] #[macro_use]
extern crate lazy_static; extern crate lazy_static;
mod ndivideosrc;
mod ndiaudiosrc; mod ndiaudiosrc;
pub mod ndisys; pub mod ndisys;
mod ndivideosrc;
use std::{thread, time};
use std::ffi::{CStr, CString};
use ndisys::*;
use gst_plugin::base_src::*; use gst_plugin::base_src::*;
use ndisys::*;
use std::ffi::{CStr, CString};
use std::{thread, time};
use std::collections::HashMap; use std::collections::HashMap;
use std::sync::Mutex; use std::sync::Mutex;
@ -44,7 +44,7 @@ fn plugin_init(plugin: &gst::Plugin) -> bool {
true true
} }
struct ndi_receiver_info{ struct ndi_receiver_info {
stream_name: String, stream_name: String,
ip: String, ip: String,
video: bool, video: bool,
@ -53,12 +53,12 @@ struct ndi_receiver_info{
id: i8, id: i8,
} }
struct Ndi{ struct Ndi {
initial_timestamp: u64, initial_timestamp: u64,
start_pts: gst::ClockTime, start_pts: gst::ClockTime,
} }
static mut ndi_struct: Ndi = Ndi{ static mut ndi_struct: Ndi = Ndi {
initial_timestamp: 0, initial_timestamp: 0,
start_pts: gst::ClockTime(Some(0)), start_pts: gst::ClockTime(Some(0)),
}; };
@ -72,7 +72,7 @@ lazy_static! {
static mut id_receiver: i8 = 0; static mut id_receiver: i8 = 0;
fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream_name: String) -> i8{ fn connect_ndi(cat: gst::DebugCategory, element: &BaseSrc, ip: String, stream_name: String) -> i8 {
gst_debug!(cat, obj: element, "Starting NDI connection..."); gst_debug!(cat, obj: element, "Starting NDI connection...");
let mut receivers = hashmap_receivers.lock().unwrap(); let mut receivers = hashmap_receivers.lock().unwrap();
@ -80,35 +80,33 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
let mut video = false; let mut video = false;
//FIXME Search for another way to know if the source is an audio or a video source //FIXME Search for another way to know if the source is an audio or a video source
if element.get_name().contains("audiosrc"){ if element.get_name().contains("audiosrc") {
audio = true; audio = true;
} } else {
else
{
video = true; video = true;
} }
for val in receivers.values_mut(){ for val in receivers.values_mut() {
if val.ip == ip || val.stream_name == stream_name{ if val.ip == ip || val.stream_name == stream_name {
if (val.audio && val.video) || (val.audio && audio) || (val.video && video){ if (val.audio && val.video) || (val.audio && audio) || (val.video && video) {
continue; continue;
} } else {
else {
if video { if video {
val.video = video; val.video = video;
} } else {
else{
val.audio = audio; val.audio = audio;
} }
return val.id; return val.id;
} }
} }
} }
unsafe { unsafe {
if !NDIlib_initialize() { if !NDIlib_initialize() {
gst_element_error!(element, gst::CoreError::Negotiation, ["Cannot run NDI: NDIlib_initialize error"]); gst_element_error!(
element,
gst::CoreError::Negotiation,
["Cannot run NDI: NDIlib_initialize error"]
);
// return false; // return false;
return 0; return 0;
} }
@ -118,7 +116,11 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
let pNDI_find = NDIlib_find_create_v2(&NDI_find_create_desc); let pNDI_find = NDIlib_find_create_v2(&NDI_find_create_desc);
//let ip_ptr = CString::new(ip.clone()).unwrap(); //let ip_ptr = CString::new(ip.clone()).unwrap();
if pNDI_find.is_null() { if pNDI_find.is_null() {
gst_element_error!(element, gst::CoreError::Negotiation, ["Cannot run NDI: NDIlib_find_create_v2 error"]); gst_element_error!(
element,
gst::CoreError::Negotiation,
["Cannot run NDI: NDIlib_find_create_v2 error"]
);
// return false; // return false;
return 0; return 0;
} }
@ -132,15 +134,26 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
// We need at least one source // We need at least one source
if p_sources.is_null() { if p_sources.is_null() {
gst_element_error!(element, gst::CoreError::Negotiation, ["Error getting NDIlib_find_get_current_sources"]); gst_element_error!(
element,
gst::CoreError::Negotiation,
["Error getting NDIlib_find_get_current_sources"]
);
// return false; // return false;
return 0; return 0;
} }
let mut no_source: isize = -1; let mut no_source: isize = -1;
for i in 0..total_sources as isize{ for i in 0..total_sources as isize {
if CStr::from_ptr((*p_sources.offset(i)).p_ndi_name).to_string_lossy().into_owned() == stream_name || if CStr::from_ptr((*p_sources.offset(i)).p_ndi_name)
CStr::from_ptr((*p_sources.offset(i)).p_ip_address).to_string_lossy().into_owned() == ip{ .to_string_lossy()
.into_owned()
== stream_name
|| CStr::from_ptr((*p_sources.offset(i)).p_ip_address)
.to_string_lossy()
.into_owned()
== ip
{
no_source = i; no_source = i;
break; break;
} }
@ -151,18 +164,27 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
return 0; return 0;
} }
gst_debug!(cat, obj: element, "Total sources in network {}: Connecting to NDI source with name '{}' and address '{}'", total_sources, gst_debug!(
cat,
obj: element,
"Total sources in network {}: Connecting to NDI source with name '{}' and address '{}'",
total_sources,
CStr::from_ptr((*p_sources.offset(no_source)).p_ndi_name) CStr::from_ptr((*p_sources.offset(no_source)).p_ndi_name)
.to_string_lossy() .to_string_lossy()
.into_owned(), .into_owned(),
CStr::from_ptr((*p_sources.offset(no_source)).p_ip_address) CStr::from_ptr((*p_sources.offset(no_source)).p_ip_address)
.to_string_lossy() .to_string_lossy()
.into_owned()); .into_owned()
);
let source = *p_sources.offset(no_source).clone(); let source = *p_sources.offset(no_source).clone();
let source_ip = CStr::from_ptr(source.p_ip_address).to_string_lossy().into_owned(); let source_ip = CStr::from_ptr(source.p_ip_address)
let source_name = CStr::from_ptr(source.p_ndi_name).to_string_lossy().into_owned(); .to_string_lossy()
.into_owned();
let source_name = CStr::from_ptr(source.p_ndi_name)
.to_string_lossy()
.into_owned();
// We now have at least one source, so we create a receiver to look at it. // We now have at least one source, so we create a receiver to look at it.
// We tell it that we prefer YCbCr video since it is more efficient for us. If the source has an alpha channel // We tell it that we prefer YCbCr video since it is more efficient for us. If the source has an alpha channel
@ -177,7 +199,11 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
let pNDI_recv = NDIlib_recv_create_v3(&NDI_recv_create_desc); let pNDI_recv = NDIlib_recv_create_v3(&NDI_recv_create_desc);
if pNDI_recv.is_null() { if pNDI_recv.is_null() {
//println!("Cannot run NDI: NDIlib_recv_create_v3 error."); //println!("Cannot run NDI: NDIlib_recv_create_v3 error.");
gst_element_error!(element, gst::CoreError::Negotiation, ["Cannot run NDI: NDIlib_recv_create_v3 error"]); gst_element_error!(
element,
gst::CoreError::Negotiation,
["Cannot run NDI: NDIlib_recv_create_v3 error"]
);
// return false; // return false;
return 0; return 0;
} }
@ -202,7 +228,17 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
NDIlib_recv_send_metadata(pNDI_recv, &enable_hw_accel); NDIlib_recv_send_metadata(pNDI_recv, &enable_hw_accel);
id_receiver += 1; id_receiver += 1;
receivers.insert(id_receiver, ndi_receiver_info{stream_name: source_name.clone(), ip: source_ip.clone(), video:video, audio: audio, ndi_instance: NdiInstance{recv: pNDI_recv}, id: id_receiver}); receivers.insert(
id_receiver,
ndi_receiver_info {
stream_name: source_name.clone(),
ip: source_ip.clone(),
video: video,
audio: audio,
ndi_instance: NdiInstance { recv: pNDI_recv },
id: id_receiver,
},
);
// let start = SystemTime::now(); // let start = SystemTime::now();
// let since_the_epoch = start.duration_since(UNIX_EPOCH) // let since_the_epoch = start.duration_since(UNIX_EPOCH)
@ -215,16 +251,15 @@ fn connect_ndi(cat: gst::DebugCategory , element: &BaseSrc, ip: String, stream
} }
} }
fn stop_ndi(cat: gst::DebugCategory , element: &BaseSrc, id: i8) -> bool{ fn stop_ndi(cat: gst::DebugCategory, element: &BaseSrc, id: i8) -> bool {
gst_debug!(cat, obj: element, "Closing NDI connection..."); gst_debug!(cat, obj: element, "Closing NDI connection...");
let mut receivers = hashmap_receivers.lock().unwrap(); let mut receivers = hashmap_receivers.lock().unwrap();
{ {
let val = receivers.get_mut(&id).unwrap(); let val = receivers.get_mut(&id).unwrap();
if val.video && val.audio{ if val.video && val.audio {
if element.get_name().contains("audiosrc"){ if element.get_name().contains("audiosrc") {
val.audio = false; val.audio = false;
} } else {
else{
val.video = false; val.video = false;
} }
return true; return true;
@ -232,7 +267,7 @@ fn stop_ndi(cat: gst::DebugCategory , element: &BaseSrc, id: i8) -> bool{
let recv = &val.ndi_instance; let recv = &val.ndi_instance;
let pNDI_recv = recv.recv; let pNDI_recv = recv.recv;
unsafe{ unsafe {
NDIlib_recv_destroy(pNDI_recv); NDIlib_recv_destroy(pNDI_recv);
// ndi_struct.recv = None; // ndi_struct.recv = None;
NDIlib_destroy(); NDIlib_destroy();

View file

@ -6,19 +6,19 @@ use gst::prelude::*;
use gst_audio; use gst_audio;
use gst_base::prelude::*; use gst_base::prelude::*;
use gobject_subclass::object::*;
use gst_plugin::base_src::*; use gst_plugin::base_src::*;
use gst_plugin::element::*; use gst_plugin::element::*;
use gobject_subclass::object::*;
use std::sync::Mutex; use std::sync::Mutex;
use std::{i32, u32}; use std::{i32, u32};
use std::ptr; use std::ptr;
use ndisys::*;
use connect_ndi; use connect_ndi;
use stop_ndi;
use ndi_struct; use ndi_struct;
use ndisys::*;
use stop_ndi;
use hashmap_receivers; use hashmap_receivers;
@ -42,20 +42,20 @@ impl Default for Settings {
// Metadata for the properties // Metadata for the properties
static PROPERTIES: [Property; 2] = [ static PROPERTIES: [Property; 2] = [
Property::String( Property::String(
"stream-name", "stream-name",
"Sream Name", "Sream Name",
"Name of the streaming device", "Name of the streaming device",
None, None,
PropertyMutability::ReadWrite, PropertyMutability::ReadWrite,
), ),
Property::String( Property::String(
"ip", "ip",
"Stream IP", "Stream IP",
"Stream IP", "Stream IP",
None, None,
PropertyMutability::ReadWrite, PropertyMutability::ReadWrite,
), ),
]; ];
// Stream-specific state, i.e. audio format configuration // Stream-specific state, i.e. audio format configuration
@ -66,13 +66,11 @@ struct State {
impl Default for State { impl Default for State {
fn default() -> State { fn default() -> State {
State { State { info: None }
info: None,
}
} }
} }
struct TimestampData{ struct TimestampData {
offset: u64, offset: u64,
} }
@ -100,9 +98,7 @@ impl NdiAudioSrc {
), ),
settings: Mutex::new(Default::default()), settings: Mutex::new(Default::default()),
state: Mutex::new(Default::default()), state: Mutex::new(Default::default()),
timestamp_data: Mutex::new(TimestampData{ timestamp_data: Mutex::new(TimestampData { offset: 0 }),
offset: 0,
}),
}) })
} }
@ -157,11 +153,10 @@ impl NdiAudioSrc {
// Install all our properties // Install all our properties
klass.install_properties(&PROPERTIES); klass.install_properties(&PROPERTIES);
} }
} }
// Virtual methods of GObject itself
// Virtual methods of GObject itself impl ObjectImpl<BaseSrc> for NdiAudioSrc {
impl ObjectImpl<BaseSrc> for NdiAudioSrc {
// Called whenever a value of a property is changed. It can be called // Called whenever a value of a property is changed. It can be called
// at any time from any thread. // at any time from any thread.
fn set_property(&self, obj: &glib::Object, id: u32, value: &glib::Value) { fn set_property(&self, obj: &glib::Object, id: u32, value: &glib::Value) {
@ -184,7 +179,7 @@ impl NdiAudioSrc {
let _ = let _ =
element.post_message(&gst::Message::new_latency().src(Some(&element)).build()); element.post_message(&gst::Message::new_latency().src(Some(&element)).build());
}, }
Property::String("ip", ..) => { Property::String("ip", ..) => {
let mut settings = self.settings.lock().unwrap(); let mut settings = self.settings.lock().unwrap();
let ip = value.get().unwrap(); let ip = value.get().unwrap();
@ -215,7 +210,7 @@ impl NdiAudioSrc {
let settings = self.settings.lock().unwrap(); let settings = self.settings.lock().unwrap();
//TODO to_value supongo que solo funciona con numeros //TODO to_value supongo que solo funciona con numeros
Ok(settings.stream_name.to_value()) Ok(settings.stream_name.to_value())
}, }
Property::String("ip", ..) => { Property::String("ip", ..) => {
let settings = self.settings.lock().unwrap(); let settings = self.settings.lock().unwrap();
//TODO to_value supongo que solo funciona con numeros //TODO to_value supongo que solo funciona con numeros
@ -224,12 +219,16 @@ impl NdiAudioSrc {
_ => unimplemented!(), _ => unimplemented!(),
} }
} }
} }
// Virtual methods of gst::Element. We override none // Virtual methods of gst::Element. We override none
impl ElementImpl<BaseSrc> for NdiAudioSrc { impl ElementImpl<BaseSrc> for NdiAudioSrc {
fn change_state(&self, element: &BaseSrc, transition: gst::StateChange) -> gst::StateChangeReturn { fn change_state(
if transition == gst::StateChange::PausedToPlaying{ &self,
element: &BaseSrc,
transition: gst::StateChange,
) -> gst::StateChangeReturn {
if transition == gst::StateChange::PausedToPlaying {
let receivers = hashmap_receivers.lock().unwrap(); let receivers = hashmap_receivers.lock().unwrap();
let settings = self.settings.lock().unwrap(); let settings = self.settings.lock().unwrap();
@ -240,22 +239,30 @@ impl NdiAudioSrc {
let audio_frame: NDIlib_audio_frame_v2_t = Default::default(); let audio_frame: NDIlib_audio_frame_v2_t = Default::default();
let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none; let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none;
unsafe{ unsafe {
while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_audio{ while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_audio {
frame_type = NDIlib_recv_capture_v2(pNDI_recv, ptr::null(), &audio_frame, ptr::null(), 1000); frame_type = NDIlib_recv_capture_v2(
pNDI_recv,
ptr::null(),
&audio_frame,
ptr::null(),
1000,
);
} }
if ndi_struct.initial_timestamp <= audio_frame.timestamp as u64 || ndi_struct.initial_timestamp == 0{ if ndi_struct.initial_timestamp <= audio_frame.timestamp as u64
|| ndi_struct.initial_timestamp == 0
{
ndi_struct.initial_timestamp = audio_frame.timestamp as u64; ndi_struct.initial_timestamp = audio_frame.timestamp as u64;
} }
} }
} }
element.parent_change_state(transition) element.parent_change_state(transition)
} }
} }
// Virtual methods of gst_base::BaseSrc // Virtual methods of gst_base::BaseSrc
impl BaseSrcImpl<BaseSrc> for NdiAudioSrc { impl BaseSrcImpl<BaseSrc> for NdiAudioSrc {
// Called whenever the input/output caps are changing, i.e. in the very beginning before data // Called whenever the input/output caps are changing, i.e. in the very beginning before data
// flow happens and whenever the situation in the pipeline is changing. All buffers after this // flow happens and whenever the situation in the pipeline is changing. All buffers after this
// call have the caps given here. // call have the caps given here.
@ -263,7 +270,6 @@ impl NdiAudioSrc {
// We simply remember the resulting AudioInfo from the caps to be able to use this for knowing // We simply remember the resulting AudioInfo from the caps to be able to use this for knowing
// the sample rate, etc. when creating buffers // the sample rate, etc. when creating buffers
fn set_caps(&self, element: &BaseSrc, caps: &gst::CapsRef) -> bool { fn set_caps(&self, element: &BaseSrc, caps: &gst::CapsRef) -> bool {
let info = match gst_audio::AudioInfo::from_caps(caps) { let info = match gst_audio::AudioInfo::from_caps(caps) {
None => return false, None => return false,
Some(info) => info, Some(info) => info,
@ -284,11 +290,15 @@ impl NdiAudioSrc {
*self.state.lock().unwrap() = Default::default(); *self.state.lock().unwrap() = Default::default();
let mut settings = self.settings.lock().unwrap(); let mut settings = self.settings.lock().unwrap();
settings.id_receiver = connect_ndi(self.cat, element, settings.ip.clone(), settings.stream_name.clone()); settings.id_receiver = connect_ndi(
if settings.id_receiver == 0{ self.cat,
element,
settings.ip.clone(),
settings.stream_name.clone(),
);
if settings.id_receiver == 0 {
return false; return false;
} } else {
else{
return true; return true;
} }
} }
@ -359,9 +369,10 @@ impl NdiAudioSrc {
let audio_frame: NDIlib_audio_frame_v2_t = Default::default(); let audio_frame: NDIlib_audio_frame_v2_t = Default::default();
let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none; let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none;
while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_audio{ while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_audio {
unsafe{ unsafe {
frame_type = NDIlib_recv_capture_v2(pNDI_recv, ptr::null(), &audio_frame, ptr::null(), 1000); frame_type =
NDIlib_recv_capture_v2(pNDI_recv, ptr::null(), &audio_frame, ptr::null(), 1000);
} }
} }
let mut caps = gst::Caps::truncate(caps); let mut caps = gst::Caps::truncate(caps);
@ -409,37 +420,43 @@ impl NdiAudioSrc {
let pts: u64; let pts: u64;
let audio_frame: NDIlib_audio_frame_v2_t = Default::default(); let audio_frame: NDIlib_audio_frame_v2_t = Default::default();
unsafe{ unsafe {
let time = ndi_struct.initial_timestamp; let time = ndi_struct.initial_timestamp;
let mut skip_frame = true; let mut skip_frame = true;
while skip_frame { while skip_frame {
let frame_type = NDIlib_recv_capture_v2(pNDI_recv, ptr::null(), &audio_frame, ptr::null(), 1000,); let frame_type =
if frame_type == NDIlib_frame_type_e::NDIlib_frame_type_none || frame_type == NDIlib_frame_type_e::NDIlib_frame_type_error { NDIlib_recv_capture_v2(pNDI_recv, ptr::null(), &audio_frame, ptr::null(), 1000);
if frame_type == NDIlib_frame_type_e::NDIlib_frame_type_none
|| frame_type == NDIlib_frame_type_e::NDIlib_frame_type_error
{
gst_element_error!(element, gst::ResourceError::Read, ["NDI frame type none received, assuming that the source closed the stream...."]); gst_element_error!(element, gst::ResourceError::Read, ["NDI frame type none received, assuming that the source closed the stream...."]);
return Err(gst::FlowReturn::CustomError); return Err(gst::FlowReturn::CustomError);
} }
if time >= (audio_frame.timestamp as u64){ if time >= (audio_frame.timestamp as u64) {
gst_debug!(self.cat, obj: element, "Frame timestamp ({:?}) is lower than received in the first frame from NDI ({:?}), so skiping...", (audio_frame.timestamp as u64), time); gst_debug!(self.cat, obj: element, "Frame timestamp ({:?}) is lower than received in the first frame from NDI ({:?}), so skiping...", (audio_frame.timestamp as u64), time);
} } else {
else{
skip_frame = false; skip_frame = false;
} }
} }
pts = audio_frame.timestamp as u64 - time; pts = audio_frame.timestamp as u64 - time;
let buff_size = ((audio_frame.channel_stride_in_bytes)) as usize; let buff_size = (audio_frame.channel_stride_in_bytes) as usize;
let mut buffer = gst::Buffer::with_size(buff_size).unwrap(); let mut buffer = gst::Buffer::with_size(buff_size).unwrap();
{ {
let vec = Vec::from_raw_parts(audio_frame.p_data as *mut u8, buff_size, buff_size); let vec = Vec::from_raw_parts(audio_frame.p_data as *mut u8, buff_size, buff_size);
let pts: gst::ClockTime = (pts * 100).into(); let pts: gst::ClockTime = (pts * 100).into();
let duration: gst::ClockTime = (((audio_frame.no_samples as f64 / audio_frame.sample_rate as f64) * 1000000000.0) as u64).into(); let duration: gst::ClockTime = (((audio_frame.no_samples as f64
/ audio_frame.sample_rate as f64)
* 1000000000.0) as u64)
.into();
let buffer = buffer.get_mut().unwrap(); let buffer = buffer.get_mut().unwrap();
if ndi_struct.start_pts == gst::ClockTime(Some(0)){ if ndi_struct.start_pts == gst::ClockTime(Some(0)) {
ndi_struct.start_pts = element.get_clock().unwrap().get_time() - element.get_base_time(); ndi_struct.start_pts =
element.get_clock().unwrap().get_time() - element.get_base_time();
} }
buffer.set_pts(pts + ndi_struct.start_pts); buffer.set_pts(pts + ndi_struct.start_pts);
@ -455,19 +472,19 @@ impl NdiAudioSrc {
Ok(buffer) Ok(buffer)
} }
} }
} }
// This zero-sized struct is containing the static metadata of our element. It is only necessary to // This zero-sized struct is containing the static metadata of our element. It is only necessary to
// be able to implement traits on it, but e.g. a plugin that registers multiple elements with the // be able to implement traits on it, but e.g. a plugin that registers multiple elements with the
// same code would use this struct to store information about the concrete element. An example of // same code would use this struct to store information about the concrete element. An example of
// this would be a plugin that wraps around a library that has multiple decoders with the same API, // this would be a plugin that wraps around a library that has multiple decoders with the same API,
// but wants (as it should) a separate element registered for each decoder. // but wants (as it should) a separate element registered for each decoder.
struct NdiAudioSrcStatic; struct NdiAudioSrcStatic;
// The basic trait for registering the type: This returns a name for the type and registers the // The basic trait for registering the type: This returns a name for the type and registers the
// instance and class initializations functions with the type system, thus hooking everything // instance and class initializations functions with the type system, thus hooking everything
// together. // together.
impl ImplTypeStatic<BaseSrc> for NdiAudioSrcStatic { impl ImplTypeStatic<BaseSrc> for NdiAudioSrcStatic {
fn get_name(&self) -> &str { fn get_name(&self) -> &str {
"NdiAudioSrc" "NdiAudioSrc"
} }
@ -479,12 +496,12 @@ impl NdiAudioSrc {
fn class_init(&self, klass: &mut BaseSrcClass) { fn class_init(&self, klass: &mut BaseSrcClass) {
NdiAudioSrc::class_init(klass); NdiAudioSrc::class_init(klass);
} }
} }
// Registers the type for our element, and then registers in GStreamer under // Registers the type for our element, and then registers in GStreamer under
// the name NdiAudioSrc for being able to instantiate it via e.g. // the name NdiAudioSrc for being able to instantiate it via e.g.
// gst::ElementFactory::make(). // gst::ElementFactory::make().
pub fn register(plugin: &gst::Plugin) { pub fn register(plugin: &gst::Plugin) {
let type_ = register_type(NdiAudioSrcStatic); let type_ = register_type(NdiAudioSrcStatic);
gst::Element::register(plugin, "ndiaudiosrc", 0, type_); gst::Element::register(plugin, "ndiaudiosrc", 0, type_);
} }

View file

@ -157,7 +157,6 @@ pub struct NdiInstance {
unsafe impl ::std::marker::Send for NdiInstance {} unsafe impl ::std::marker::Send for NdiInstance {}
#[repr(C)] #[repr(C)]
#[derive(Debug, Copy, Clone)] #[derive(Debug, Copy, Clone)]
pub struct NDIlib_tally_t { pub struct NDIlib_tally_t {

View file

@ -3,23 +3,23 @@
use glib; use glib;
use gst; use gst;
use gst::prelude::*; use gst::prelude::*;
use gst_video;
use gst_base::prelude::*;
use gst::Fraction; use gst::Fraction;
use gst_base::prelude::*;
use gst_video;
use gobject_subclass::object::*;
use gst_plugin::base_src::*; use gst_plugin::base_src::*;
use gst_plugin::element::*; use gst_plugin::element::*;
use gobject_subclass::object::*;
use std::sync::Mutex; use std::sync::Mutex;
use std::{i32, u32}; use std::{i32, u32};
use std::ptr; use std::ptr;
use ndisys::*;
use connect_ndi; use connect_ndi;
use stop_ndi;
use ndi_struct; use ndi_struct;
use ndisys::*;
use stop_ndi;
use hashmap_receivers; use hashmap_receivers;
@ -45,20 +45,20 @@ impl Default for Settings {
// Metadata for the properties // Metadata for the properties
static PROPERTIES: [Property; 2] = [ static PROPERTIES: [Property; 2] = [
Property::String( Property::String(
"stream-name", "stream-name",
"Sream Name", "Sream Name",
"Name of the streaming device", "Name of the streaming device",
None, None,
PropertyMutability::ReadWrite, PropertyMutability::ReadWrite,
), ),
Property::String( Property::String(
"ip", "ip",
"Stream IP", "Stream IP",
"Stream IP", "Stream IP",
None, None,
PropertyMutability::ReadWrite, PropertyMutability::ReadWrite,
), ),
]; ];
// Stream-specific state, i.e. audio format configuration // Stream-specific state, i.e. audio format configuration
@ -69,13 +69,11 @@ struct State {
impl Default for State { impl Default for State {
fn default() -> State { fn default() -> State {
State { State { info: None }
info: None,
}
} }
} }
struct TimestampData{ struct TimestampData {
offset: u64, offset: u64,
} }
@ -103,9 +101,7 @@ impl NdiVideoSrc {
), ),
settings: Mutex::new(Default::default()), settings: Mutex::new(Default::default()),
state: Mutex::new(Default::default()), state: Mutex::new(Default::default()),
timestamp_data: Mutex::new(TimestampData{ timestamp_data: Mutex::new(TimestampData { offset: 0 }),
offset: 0,
}),
}) })
} }
@ -166,12 +162,10 @@ impl NdiVideoSrc {
// Install all our properties // Install all our properties
klass.install_properties(&PROPERTIES); klass.install_properties(&PROPERTIES);
} }
} }
// Virtual methods of GObject itself
impl ObjectImpl<BaseSrc> for NdiVideoSrc {
// Virtual methods of GObject itself
impl ObjectImpl<BaseSrc> for NdiVideoSrc {
// Called whenever a value of a property is changed. It can be called // Called whenever a value of a property is changed. It can be called
// at any time from any thread. // at any time from any thread.
fn set_property(&self, obj: &glib::Object, id: u32, value: &glib::Value) { fn set_property(&self, obj: &glib::Object, id: u32, value: &glib::Value) {
@ -194,7 +188,7 @@ impl NdiVideoSrc {
// let _ = // let _ =
// element.post_message(&gst::Message::new_latency().src(Some(&element)).build()); // element.post_message(&gst::Message::new_latency().src(Some(&element)).build());
}, }
Property::String("ip", ..) => { Property::String("ip", ..) => {
let mut settings = self.settings.lock().unwrap(); let mut settings = self.settings.lock().unwrap();
let ip = value.get().unwrap(); let ip = value.get().unwrap();
@ -224,7 +218,7 @@ impl NdiVideoSrc {
Property::String("stream-name", ..) => { Property::String("stream-name", ..) => {
let settings = self.settings.lock().unwrap(); let settings = self.settings.lock().unwrap();
Ok(settings.stream_name.to_value()) Ok(settings.stream_name.to_value())
}, }
Property::String("ip", ..) => { Property::String("ip", ..) => {
let settings = self.settings.lock().unwrap(); let settings = self.settings.lock().unwrap();
Ok(settings.ip.to_value()) Ok(settings.ip.to_value())
@ -232,12 +226,16 @@ impl NdiVideoSrc {
_ => unimplemented!(), _ => unimplemented!(),
} }
} }
} }
// Virtual methods of gst::Element. We override none // Virtual methods of gst::Element. We override none
impl ElementImpl<BaseSrc> for NdiVideoSrc { impl ElementImpl<BaseSrc> for NdiVideoSrc {
fn change_state(&self, element: &BaseSrc, transition: gst::StateChange) -> gst::StateChangeReturn { fn change_state(
if transition == gst::StateChange::PausedToPlaying{ &self,
element: &BaseSrc,
transition: gst::StateChange,
) -> gst::StateChangeReturn {
if transition == gst::StateChange::PausedToPlaying {
let receivers = hashmap_receivers.lock().unwrap(); let receivers = hashmap_receivers.lock().unwrap();
let settings = self.settings.lock().unwrap(); let settings = self.settings.lock().unwrap();
@ -248,23 +246,30 @@ impl NdiVideoSrc {
let video_frame: NDIlib_video_frame_v2_t = Default::default(); let video_frame: NDIlib_video_frame_v2_t = Default::default();
let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none; let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none;
unsafe{ unsafe {
while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_video{ while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_video {
frame_type = NDIlib_recv_capture_v2(pNDI_recv, &video_frame, ptr::null(), ptr::null(), 1000); frame_type = NDIlib_recv_capture_v2(
pNDI_recv,
&video_frame,
ptr::null(),
ptr::null(),
1000,
);
} }
if ndi_struct.initial_timestamp <= video_frame.timestamp as u64 || ndi_struct.initial_timestamp == 0{ if ndi_struct.initial_timestamp <= video_frame.timestamp as u64
|| ndi_struct.initial_timestamp == 0
{
ndi_struct.initial_timestamp = video_frame.timestamp as u64; ndi_struct.initial_timestamp = video_frame.timestamp as u64;
} }
} }
} }
element.parent_change_state(transition) element.parent_change_state(transition)
} }
} }
// Virtual methods of gst_base::BaseSrc
// Virtual methods of gst_base::BaseSrc impl BaseSrcImpl<BaseSrc> for NdiVideoSrc {
impl BaseSrcImpl<BaseSrc> for NdiVideoSrc {
// Called whenever the input/output caps are changing, i.e. in the very beginning before data // Called whenever the input/output caps are changing, i.e. in the very beginning before data
// flow happens and whenever the situation in the pipeline is changing. All buffers after this // flow happens and whenever the situation in the pipeline is changing. All buffers after this
// call have the caps given here. // call have the caps given here.
@ -290,12 +295,16 @@ impl NdiVideoSrc {
// Reset state // Reset state
*self.state.lock().unwrap() = Default::default(); *self.state.lock().unwrap() = Default::default();
let mut settings = self.settings.lock().unwrap(); let mut settings = self.settings.lock().unwrap();
settings.id_receiver = connect_ndi(self.cat, element, settings.ip.clone(), settings.stream_name.clone()); settings.id_receiver = connect_ndi(
self.cat,
element,
settings.ip.clone(),
settings.stream_name.clone(),
);
if settings.id_receiver == 0{ if settings.id_receiver == 0 {
return false; return false;
} } else {
else{
// let _ = element.post_message(&gst::Message::new_latency().src(Some(element)).build()); // let _ = element.post_message(&gst::Message::new_latency().src(Some(element)).build());
return true; return true;
} }
@ -313,7 +322,6 @@ impl NdiVideoSrc {
true true
} }
fn query(&self, element: &BaseSrc, query: &mut gst::QueryRef) -> bool { fn query(&self, element: &BaseSrc, query: &mut gst::QueryRef) -> bool {
use gst::QueryView; use gst::QueryView;
match query.view_mut() { match query.view_mut() {
@ -374,9 +382,10 @@ impl NdiVideoSrc {
let video_frame: NDIlib_video_frame_v2_t = Default::default(); let video_frame: NDIlib_video_frame_v2_t = Default::default();
let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none; let mut frame_type: NDIlib_frame_type_e = NDIlib_frame_type_e::NDIlib_frame_type_none;
while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_video{ while frame_type != NDIlib_frame_type_e::NDIlib_frame_type_video {
unsafe{ unsafe {
frame_type = NDIlib_recv_capture_v2(pNDI_recv, &video_frame, ptr::null(), ptr::null(), 1000); frame_type =
NDIlib_recv_capture_v2(pNDI_recv, &video_frame, ptr::null(), ptr::null(), 1000);
} }
} }
@ -386,7 +395,10 @@ impl NdiVideoSrc {
let s = caps.get_mut_structure(0).unwrap(); let s = caps.get_mut_structure(0).unwrap();
s.fixate_field_nearest_int("width", video_frame.xres); s.fixate_field_nearest_int("width", video_frame.xres);
s.fixate_field_nearest_int("height", video_frame.yres); s.fixate_field_nearest_int("height", video_frame.yres);
s.fixate_field_nearest_fraction("framerate", Fraction::new(video_frame.frame_rate_N, video_frame.frame_rate_D)); s.fixate_field_nearest_fraction(
"framerate",
Fraction::new(video_frame.frame_rate_N, video_frame.frame_rate_D),
);
} }
// Let BaseSrc fixate anything else for us. We could've alternatively have // Let BaseSrc fixate anything else for us. We could've alternatively have
@ -427,20 +439,22 @@ impl NdiVideoSrc {
let pts: u64; let pts: u64;
let video_frame: NDIlib_video_frame_v2_t = Default::default(); let video_frame: NDIlib_video_frame_v2_t = Default::default();
unsafe{ unsafe {
let time = ndi_struct.initial_timestamp; let time = ndi_struct.initial_timestamp;
let mut skip_frame = true; let mut skip_frame = true;
while skip_frame { while skip_frame {
let frame_type = NDIlib_recv_capture_v2(pNDI_recv, &video_frame, ptr::null(), ptr::null(), 1000,); let frame_type =
if frame_type == NDIlib_frame_type_e::NDIlib_frame_type_none || frame_type == NDIlib_frame_type_e::NDIlib_frame_type_error { NDIlib_recv_capture_v2(pNDI_recv, &video_frame, ptr::null(), ptr::null(), 1000);
if frame_type == NDIlib_frame_type_e::NDIlib_frame_type_none
|| frame_type == NDIlib_frame_type_e::NDIlib_frame_type_error
{
gst_element_error!(element, gst::ResourceError::Read, ["NDI frame type none received, assuming that the source closed the stream...."]); gst_element_error!(element, gst::ResourceError::Read, ["NDI frame type none received, assuming that the source closed the stream...."]);
return Err(gst::FlowReturn::CustomError); return Err(gst::FlowReturn::CustomError);
} }
if time >= (video_frame.timestamp as u64){ if time >= (video_frame.timestamp as u64) {
gst_debug!(self.cat, obj: element, "Frame timestamp ({:?}) is lower than received in the first frame from NDI ({:?}), so skiping...", (video_frame.timestamp as u64), time); gst_debug!(self.cat, obj: element, "Frame timestamp ({:?}) is lower than received in the first frame from NDI ({:?}), so skiping...", (video_frame.timestamp as u64), time);
} } else {
else{
skip_frame = false; skip_frame = false;
} }
} }
@ -453,11 +467,15 @@ impl NdiVideoSrc {
let vec = Vec::from_raw_parts(video_frame.p_data as *mut u8, buff_size, buff_size); let vec = Vec::from_raw_parts(video_frame.p_data as *mut u8, buff_size, buff_size);
let pts: gst::ClockTime = (pts * 100).into(); let pts: gst::ClockTime = (pts * 100).into();
let duration: gst::ClockTime = (((video_frame.frame_rate_D as f64 / video_frame.frame_rate_N as f64) * 1000000000.0) as u64).into(); let duration: gst::ClockTime = (((video_frame.frame_rate_D as f64
/ video_frame.frame_rate_N as f64)
* 1000000000.0) as u64)
.into();
let buffer = buffer.get_mut().unwrap(); let buffer = buffer.get_mut().unwrap();
if ndi_struct.start_pts == gst::ClockTime(Some(0)){ if ndi_struct.start_pts == gst::ClockTime(Some(0)) {
ndi_struct.start_pts = element.get_clock().unwrap().get_time() - element.get_base_time(); ndi_struct.start_pts =
element.get_clock().unwrap().get_time() - element.get_base_time();
} }
buffer.set_pts(pts + ndi_struct.start_pts); buffer.set_pts(pts + ndi_struct.start_pts);
@ -473,19 +491,19 @@ impl NdiVideoSrc {
Ok(buffer) Ok(buffer)
} }
} }
} }
// This zero-sized struct is containing the static metadata of our element. It is only necessary to // This zero-sized struct is containing the static metadata of our element. It is only necessary to
// be able to implement traits on it, but e.g. a plugin that registers multiple elements with the // be able to implement traits on it, but e.g. a plugin that registers multiple elements with the
// same code would use this struct to store information about the concrete element. An example of // same code would use this struct to store information about the concrete element. An example of
// this would be a plugin that wraps around a library that has multiple decoders with the same API, // this would be a plugin that wraps around a library that has multiple decoders with the same API,
// but wants (as it should) a separate element registered for each decoder. // but wants (as it should) a separate element registered for each decoder.
struct NdiVideoSrcStatic; struct NdiVideoSrcStatic;
// The basic trait for registering the type: This returns a name for the type and registers the // The basic trait for registering the type: This returns a name for the type and registers the
// instance and class initializations functions with the type system, thus hooking everything // instance and class initializations functions with the type system, thus hooking everything
// together. // together.
impl ImplTypeStatic<BaseSrc> for NdiVideoSrcStatic { impl ImplTypeStatic<BaseSrc> for NdiVideoSrcStatic {
fn get_name(&self) -> &str { fn get_name(&self) -> &str {
"NdiVideoSrc" "NdiVideoSrc"
} }
@ -497,12 +515,12 @@ impl NdiVideoSrc {
fn class_init(&self, klass: &mut BaseSrcClass) { fn class_init(&self, klass: &mut BaseSrcClass) {
NdiVideoSrc::class_init(klass); NdiVideoSrc::class_init(klass);
} }
} }
// Registers the type for our element, and then registers in GStreamer under // Registers the type for our element, and then registers in GStreamer under
// the name NdiVideoSrc for being able to instantiate it via e.g. // the name NdiVideoSrc for being able to instantiate it via e.g.
// gst::ElementFactory::make(). // gst::ElementFactory::make().
pub fn register(plugin: &gst::Plugin) { pub fn register(plugin: &gst::Plugin) {
let type_ = register_type(NdiVideoSrcStatic); let type_ = register_type(NdiVideoSrcStatic);
gst::Element::register(plugin, "ndivideosrc", 0, type_); gst::Element::register(plugin, "ndivideosrc", 0, type_);
} }