mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-23 02:01:12 +00:00
396ef0cbcf
By setting the earliest time to timestamp + 2 * diff there would be a difference of 1 * diff between the current clock time and the earliest time the element would let through in the future. If e.g. a frame is arriving 30s late at the sink, then not just all frames up to that point would be dropped but also 30s of frames after the current clock time. Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/7459>
2301 lines
69 KiB
C
2301 lines
69 KiB
C
/* Generic video mixer plugin
|
|
* Copyright (C) 2004, 2008 Wim Taymans <wim@fluendo.com>
|
|
* Copyright (C) 2010 Sebastian Dröge <sebastian.droege@collabora.co.uk>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-videomixer
|
|
* @title: videomixer
|
|
*
|
|
* IMPORTANT: #videomixer is deprecated in favor of #compositor, please do not
|
|
* use this element in newly-written code!
|
|
*
|
|
* Videomixer can accept AYUV, ARGB and BGRA video streams. For each of the requested
|
|
* sink pads it will compare the incoming geometry and framerate to define the
|
|
* output parameters. Indeed output video frames will have the geometry of the
|
|
* biggest incoming video stream and the framerate of the fastest incoming one.
|
|
*
|
|
* Videomixer will do colorspace conversion.
|
|
*
|
|
* Individual parameters for each input stream can be configured on the
|
|
* videomixer pads.
|
|
*
|
|
* ## Sample pipelines
|
|
* |[
|
|
* gst-launch-1.0 \
|
|
* videotestsrc pattern=1 ! \
|
|
* video/x-raw,format=AYUV,framerate=\(fraction\)10/1,width=100,height=100 ! \
|
|
* videobox border-alpha=0 top=-70 bottom=-70 right=-220 ! \
|
|
* videomixer name=mix sink_0::alpha=0.7 sink_1::alpha=0.5 ! \
|
|
* videoconvert ! xvimagesink \
|
|
* videotestsrc ! \
|
|
* video/x-raw,format=AYUV,framerate=\(fraction\)5/1,width=320,height=240 ! mix.
|
|
* ]| A pipeline to demonstrate videomixer used together with videobox.
|
|
* This should show a 320x240 pixels video test source with some transparency
|
|
* showing the background checker pattern. Another video test source with just
|
|
* the snow pattern of 100x100 pixels is overlaid on top of the first one on
|
|
* the left vertically centered with a small transparency showing the first
|
|
* video test source behind and the checker pattern under it. Note that the
|
|
* framerate of the output video is 10 frames per second.
|
|
* |[
|
|
* gst-launch-1.0 videotestsrc pattern=1 ! \
|
|
* video/x-raw, framerate=\(fraction\)10/1, width=100, height=100 ! \
|
|
* videomixer name=mix ! videoconvert ! ximagesink \
|
|
* videotestsrc ! \
|
|
* video/x-raw, framerate=\(fraction\)5/1, width=320, height=240 ! mix.
|
|
* ]| A pipeline to demonstrate bgra mixing. (This does not demonstrate alpha blending).
|
|
* |[
|
|
* gst-launch-1.0 videotestsrc pattern=1 ! \
|
|
* video/x-raw,format =I420, framerate=\(fraction\)10/1, width=100, height=100 ! \
|
|
* videomixer name=mix ! videoconvert ! ximagesink \
|
|
* videotestsrc ! \
|
|
* video/x-raw,format=I420, framerate=\(fraction\)5/1, width=320, height=240 ! mix.
|
|
* ]| A pipeline to test I420
|
|
* |[
|
|
* gst-launch-1.0 videomixer name=mixer sink_1::alpha=0.5 sink_1::xpos=50 sink_1::ypos=50 ! \
|
|
* videoconvert ! ximagesink \
|
|
* videotestsrc pattern=snow timestamp-offset=3000000000 ! \
|
|
* "video/x-raw,format=AYUV,width=640,height=480,framerate=(fraction)30/1" ! \
|
|
* timeoverlay ! queue2 ! mixer. \
|
|
* videotestsrc pattern=smpte ! \
|
|
* "video/x-raw,format=AYUV,width=800,height=600,framerate=(fraction)10/1" ! \
|
|
* timeoverlay ! queue2 ! mixer.
|
|
* ]| A pipeline to demonstrate synchronized mixing (the second stream starts after 3 seconds)
|
|
*
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include <string.h>
|
|
|
|
#include "videomixer2.h"
|
|
#include "videomixer2pad.h"
|
|
|
|
#ifdef DISABLE_ORC
|
|
#define orc_memset memset
|
|
#else
|
|
#include <orc/orcfunctions.h>
|
|
#endif
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (gst_videomixer2_debug);
|
|
#define GST_CAT_DEFAULT gst_videomixer2_debug
|
|
|
|
#define GST_VIDEO_MIXER2_GET_LOCK(mix) \
|
|
(&GST_VIDEO_MIXER2(mix)->lock)
|
|
#define GST_VIDEO_MIXER2_LOCK(mix) \
|
|
(g_mutex_lock(GST_VIDEO_MIXER2_GET_LOCK (mix)))
|
|
#define GST_VIDEO_MIXER2_UNLOCK(mix) \
|
|
(g_mutex_unlock(GST_VIDEO_MIXER2_GET_LOCK (mix)))
|
|
#define GST_VIDEO_MIXER2_GET_SETCAPS_LOCK(mix) \
|
|
(&GST_VIDEO_MIXER2(mix)->setcaps_lock)
|
|
#define GST_VIDEO_MIXER2_SETCAPS_LOCK(mix) \
|
|
(g_mutex_lock(GST_VIDEO_MIXER2_GET_SETCAPS_LOCK (mix)))
|
|
#define GST_VIDEO_MIXER2_SETCAPS_UNLOCK(mix) \
|
|
(g_mutex_unlock(GST_VIDEO_MIXER2_GET_SETCAPS_LOCK (mix)))
|
|
|
|
#define FORMATS " { AYUV, BGRA, ARGB, RGBA, ABGR, Y444, Y42B, YUY2, UYVY, "\
|
|
" YVYU, I420, YV12, NV12, NV21, Y41B, RGB, BGR, xRGB, xBGR, "\
|
|
" RGBx, BGRx } "
|
|
|
|
static GstStaticPadTemplate src_factory = GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (FORMATS))
|
|
);
|
|
|
|
static GstStaticPadTemplate sink_factory = GST_STATIC_PAD_TEMPLATE ("sink_%u",
|
|
GST_PAD_SINK,
|
|
GST_PAD_REQUEST,
|
|
GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (FORMATS))
|
|
);
|
|
|
|
static void gst_videomixer2_child_proxy_init (gpointer g_iface,
|
|
gpointer iface_data);
|
|
static gboolean gst_videomixer2_push_sink_event (GstVideoMixer2 * mix,
|
|
GstEvent * event);
|
|
static void gst_videomixer2_release_pad (GstElement * element, GstPad * pad);
|
|
static void gst_videomixer2_reset_qos (GstVideoMixer2 * mix);
|
|
|
|
struct _GstVideoMixer2Collect
|
|
{
|
|
GstCollectData collect; /* we extend the CollectData */
|
|
|
|
GstVideoMixer2Pad *mixpad;
|
|
|
|
GstBuffer *queued; /* buffer for which we don't know the end time yet */
|
|
GstVideoInfo queued_vinfo;
|
|
|
|
GstBuffer *buffer; /* buffer that should be blended now */
|
|
GstVideoInfo buffer_vinfo;
|
|
|
|
GstClockTime start_time;
|
|
GstClockTime end_time;
|
|
};
|
|
|
|
#define DEFAULT_PAD_ZORDER 0
|
|
#define DEFAULT_PAD_XPOS 0
|
|
#define DEFAULT_PAD_YPOS 0
|
|
#define DEFAULT_PAD_ALPHA 1.0
|
|
enum
|
|
{
|
|
PROP_PAD_0,
|
|
PROP_PAD_ZORDER,
|
|
PROP_PAD_XPOS,
|
|
PROP_PAD_YPOS,
|
|
PROP_PAD_ALPHA
|
|
};
|
|
|
|
G_DEFINE_TYPE (GstVideoMixer2Pad, gst_videomixer2_pad, GST_TYPE_PAD);
|
|
|
|
static void
|
|
gst_videomixer2_collect_free (GstCollectData * data)
|
|
{
|
|
GstVideoMixer2Collect *cdata = (GstVideoMixer2Collect *) data;
|
|
|
|
gst_buffer_replace (&cdata->buffer, NULL);
|
|
}
|
|
|
|
static gboolean gst_videomixer2_src_setcaps (GstPad * pad, GstVideoMixer2 * mix,
|
|
GstCaps * caps);
|
|
|
|
static gboolean
|
|
gst_videomixer2_update_src_caps (GstVideoMixer2 * mix)
|
|
{
|
|
GSList *l;
|
|
gint best_width = -1, best_height = -1;
|
|
gdouble best_fps = -1, cur_fps;
|
|
gint best_fps_n = -1, best_fps_d = -1;
|
|
gboolean ret = TRUE;
|
|
|
|
GST_VIDEO_MIXER2_SETCAPS_LOCK (mix);
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
|
|
for (l = mix->sinkpads; l; l = l->next) {
|
|
GstVideoMixer2Pad *mpad = l->data;
|
|
gint this_width, this_height;
|
|
gint fps_n, fps_d;
|
|
gint width, height;
|
|
|
|
fps_n = GST_VIDEO_INFO_FPS_N (&mpad->info);
|
|
fps_d = GST_VIDEO_INFO_FPS_D (&mpad->info);
|
|
width = GST_VIDEO_INFO_WIDTH (&mpad->info);
|
|
height = GST_VIDEO_INFO_HEIGHT (&mpad->info);
|
|
|
|
if (width == 0 || height == 0)
|
|
continue;
|
|
|
|
this_width = width + MAX (mpad->xpos, 0);
|
|
this_height = height + MAX (mpad->ypos, 0);
|
|
|
|
if (best_width < this_width)
|
|
best_width = this_width;
|
|
if (best_height < this_height)
|
|
best_height = this_height;
|
|
|
|
if (fps_d == 0)
|
|
cur_fps = 0.0;
|
|
else
|
|
gst_util_fraction_to_double (fps_n, fps_d, &cur_fps);
|
|
|
|
if (best_fps < cur_fps) {
|
|
best_fps = cur_fps;
|
|
best_fps_n = fps_n;
|
|
best_fps_d = fps_d;
|
|
}
|
|
}
|
|
|
|
if (best_fps_n <= 0 || best_fps_d <= 0 || best_fps == 0.0) {
|
|
best_fps_n = 25;
|
|
best_fps_d = 1;
|
|
best_fps = 25.0;
|
|
}
|
|
|
|
if (best_width > 0 && best_height > 0 && best_fps > 0) {
|
|
GstCaps *caps, *peercaps;
|
|
GstStructure *s;
|
|
GstVideoInfo info;
|
|
|
|
if (GST_VIDEO_INFO_FPS_N (&mix->info) != best_fps_n ||
|
|
GST_VIDEO_INFO_FPS_D (&mix->info) != best_fps_d) {
|
|
if (mix->segment.position != -1) {
|
|
mix->ts_offset = mix->segment.position - mix->segment.start;
|
|
mix->nframes = 0;
|
|
}
|
|
}
|
|
gst_video_info_init (&info);
|
|
gst_video_info_set_format (&info, GST_VIDEO_INFO_FORMAT (&mix->info),
|
|
best_width, best_height);
|
|
info.fps_n = best_fps_n;
|
|
info.fps_d = best_fps_d;
|
|
info.par_n = GST_VIDEO_INFO_PAR_N (&mix->info);
|
|
info.par_d = GST_VIDEO_INFO_PAR_D (&mix->info);
|
|
|
|
caps = gst_video_info_to_caps (&info);
|
|
|
|
peercaps = gst_pad_peer_query_caps (mix->srcpad, NULL);
|
|
if (peercaps && !gst_caps_can_intersect (peercaps, caps)) {
|
|
GstCaps *tmp;
|
|
|
|
s = gst_caps_get_structure (caps, 0);
|
|
gst_structure_set (s, "width", GST_TYPE_INT_RANGE, 1, G_MAXINT, "height",
|
|
GST_TYPE_INT_RANGE, 1, G_MAXINT, "framerate", GST_TYPE_FRACTION_RANGE,
|
|
0, 1, G_MAXINT, 1, NULL);
|
|
|
|
tmp = gst_caps_intersect (caps, peercaps);
|
|
gst_caps_unref (caps);
|
|
gst_caps_unref (peercaps);
|
|
peercaps = NULL;
|
|
caps = tmp;
|
|
if (gst_caps_is_empty (caps)) {
|
|
GST_DEBUG_OBJECT (mix, "empty caps");
|
|
ret = FALSE;
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
goto done;
|
|
}
|
|
|
|
caps = gst_caps_truncate (caps);
|
|
s = gst_caps_get_structure (caps, 0);
|
|
gst_structure_fixate_field_nearest_int (s, "width", best_width);
|
|
gst_structure_fixate_field_nearest_int (s, "height", best_height);
|
|
gst_structure_fixate_field_nearest_fraction (s, "framerate", best_fps_n,
|
|
best_fps_d);
|
|
|
|
gst_structure_get_int (s, "width", &info.width);
|
|
gst_structure_get_int (s, "height", &info.height);
|
|
gst_structure_get_fraction (s, "fraction", &info.fps_n, &info.fps_d);
|
|
}
|
|
if (peercaps)
|
|
gst_caps_unref (peercaps);
|
|
|
|
gst_caps_unref (caps);
|
|
caps = gst_video_info_to_caps (&info);
|
|
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
ret = gst_videomixer2_src_setcaps (mix->srcpad, mix, caps);
|
|
gst_caps_unref (caps);
|
|
} else {
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
}
|
|
|
|
done:
|
|
GST_VIDEO_MIXER2_SETCAPS_UNLOCK (mix);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_update_converters (GstVideoMixer2 * mix)
|
|
{
|
|
GSList *tmp;
|
|
GstVideoFormat best_format;
|
|
GstVideoInfo best_info;
|
|
GstVideoMixer2Pad *pad;
|
|
gboolean need_alpha = FALSE;
|
|
gboolean at_least_one_alpha = FALSE;
|
|
GstCaps *downstream_caps;
|
|
GstCaps *possible_caps;
|
|
gchar *best_colorimetry;
|
|
const gchar *best_chroma;
|
|
GHashTable *formats_table;
|
|
gint best_format_number = 0;
|
|
|
|
best_format = GST_VIDEO_FORMAT_UNKNOWN;
|
|
gst_video_info_init (&best_info);
|
|
|
|
downstream_caps = gst_pad_get_allowed_caps (mix->srcpad);
|
|
|
|
if (!downstream_caps || gst_caps_is_empty (downstream_caps)) {
|
|
if (downstream_caps)
|
|
gst_caps_unref (downstream_caps);
|
|
return FALSE;
|
|
}
|
|
|
|
formats_table = g_hash_table_new (g_direct_hash, g_direct_equal);
|
|
|
|
/* first find new preferred format */
|
|
for (tmp = mix->sinkpads; tmp; tmp = tmp->next) {
|
|
GstStructure *s;
|
|
gint format_number;
|
|
|
|
pad = tmp->data;
|
|
|
|
if (!pad->info.finfo)
|
|
continue;
|
|
|
|
if (pad->info.finfo->flags & GST_VIDEO_FORMAT_FLAG_ALPHA)
|
|
at_least_one_alpha = TRUE;
|
|
|
|
/* If we want alpha, disregard all the other formats */
|
|
if (need_alpha && !(pad->info.finfo->flags & GST_VIDEO_FORMAT_FLAG_ALPHA))
|
|
continue;
|
|
|
|
/* This can happen if we release a pad and another pad hasn't been negotiated yet */
|
|
if (GST_VIDEO_INFO_FORMAT (&pad->info) == GST_VIDEO_FORMAT_UNKNOWN)
|
|
continue;
|
|
|
|
possible_caps = gst_video_info_to_caps (&pad->info);
|
|
|
|
s = gst_caps_get_structure (possible_caps, 0);
|
|
gst_structure_remove_fields (s, "width", "height", "framerate",
|
|
"pixel-aspect-ratio", "interlace-mode", NULL);
|
|
|
|
/* Can downstream accept this format ? */
|
|
if (!gst_caps_can_intersect (downstream_caps, possible_caps)) {
|
|
gst_caps_unref (possible_caps);
|
|
continue;
|
|
}
|
|
|
|
gst_caps_unref (possible_caps);
|
|
|
|
format_number =
|
|
GPOINTER_TO_INT (g_hash_table_lookup (formats_table,
|
|
GINT_TO_POINTER (GST_VIDEO_INFO_FORMAT (&pad->info))));
|
|
format_number += 1;
|
|
|
|
g_hash_table_replace (formats_table,
|
|
GINT_TO_POINTER (GST_VIDEO_INFO_FORMAT (&pad->info)),
|
|
GINT_TO_POINTER (format_number));
|
|
|
|
/* If that pad is the first with alpha, set it as the new best format */
|
|
if (!need_alpha && (pad->info.finfo->flags & GST_VIDEO_FORMAT_FLAG_ALPHA)) {
|
|
need_alpha = TRUE;
|
|
best_format = GST_VIDEO_INFO_FORMAT (&pad->info);
|
|
best_info = pad->info;
|
|
best_format_number = format_number;
|
|
} else if (format_number > best_format_number) {
|
|
best_format = GST_VIDEO_INFO_FORMAT (&pad->info);
|
|
best_info = pad->info;
|
|
best_format_number = format_number;
|
|
}
|
|
}
|
|
|
|
g_hash_table_unref (formats_table);
|
|
|
|
if (best_format == GST_VIDEO_FORMAT_UNKNOWN) {
|
|
downstream_caps = gst_caps_fixate (downstream_caps);
|
|
gst_video_info_from_caps (&best_info, downstream_caps);
|
|
best_format = GST_VIDEO_INFO_FORMAT (&best_info);
|
|
}
|
|
|
|
gst_caps_unref (downstream_caps);
|
|
|
|
if (at_least_one_alpha
|
|
&& !(best_info.finfo->flags & GST_VIDEO_FORMAT_FLAG_ALPHA)) {
|
|
GST_ELEMENT_ERROR (mix, CORE, NEGOTIATION,
|
|
("At least one of the input pads contains alpha, but downstream can't support alpha."),
|
|
("Either convert your inputs to not contain alpha or add a videoconvert after the mixer"));
|
|
return FALSE;
|
|
}
|
|
|
|
best_colorimetry = gst_video_colorimetry_to_string (&(best_info.colorimetry));
|
|
best_chroma = gst_video_chroma_to_string (best_info.chroma_site);
|
|
|
|
if (GST_VIDEO_INFO_FPS_N (&mix->info) != GST_VIDEO_INFO_FPS_N (&best_info) ||
|
|
GST_VIDEO_INFO_FPS_D (&mix->info) != GST_VIDEO_INFO_FPS_D (&best_info)) {
|
|
if (mix->segment.position != -1) {
|
|
mix->ts_offset = mix->segment.position - mix->segment.start;
|
|
mix->nframes = 0;
|
|
} else {
|
|
mix->ts_offset += gst_util_uint64_scale_round (mix->nframes,
|
|
GST_SECOND * GST_VIDEO_INFO_FPS_D (&mix->info),
|
|
GST_VIDEO_INFO_FPS_N (&mix->info));
|
|
mix->nframes = 0;
|
|
}
|
|
}
|
|
|
|
mix->info = best_info;
|
|
|
|
GST_DEBUG_OBJECT (mix,
|
|
"The output format will now be : %d with colorimetry : %s and chroma : %s",
|
|
best_format, best_colorimetry, best_chroma);
|
|
|
|
/* Then browse the sinks once more, setting or unsetting conversion if needed */
|
|
for (tmp = mix->sinkpads; tmp; tmp = tmp->next) {
|
|
gchar *colorimetry;
|
|
const gchar *chroma;
|
|
|
|
pad = tmp->data;
|
|
|
|
if (!pad->info.finfo)
|
|
continue;
|
|
|
|
if (GST_VIDEO_INFO_FORMAT (&pad->info) == GST_VIDEO_FORMAT_UNKNOWN)
|
|
continue;
|
|
|
|
if (pad->convert)
|
|
gst_video_converter_free (pad->convert);
|
|
|
|
pad->convert = NULL;
|
|
|
|
colorimetry = gst_video_colorimetry_to_string (&(pad->info.colorimetry));
|
|
chroma = gst_video_chroma_to_string (pad->info.chroma_site);
|
|
|
|
if (best_format != GST_VIDEO_INFO_FORMAT (&pad->info) ||
|
|
g_strcmp0 (colorimetry, best_colorimetry) ||
|
|
g_strcmp0 (chroma, best_chroma)) {
|
|
GstVideoInfo tmp_info = pad->info;
|
|
tmp_info.finfo = best_info.finfo;
|
|
tmp_info.chroma_site = best_info.chroma_site;
|
|
tmp_info.colorimetry = best_info.colorimetry;
|
|
|
|
GST_DEBUG_OBJECT (pad, "This pad will be converted from %d to %d",
|
|
GST_VIDEO_INFO_FORMAT (&pad->info),
|
|
GST_VIDEO_INFO_FORMAT (&best_info));
|
|
pad->convert = gst_video_converter_new (&pad->info, &tmp_info, NULL);
|
|
pad->need_conversion_update = TRUE;
|
|
if (!pad->convert) {
|
|
g_free (colorimetry);
|
|
g_free (best_colorimetry);
|
|
GST_WARNING ("No path found for conversion");
|
|
return FALSE;
|
|
}
|
|
} else {
|
|
GST_DEBUG_OBJECT (pad, "This pad will not need conversion");
|
|
}
|
|
g_free (colorimetry);
|
|
}
|
|
|
|
g_free (best_colorimetry);
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_pad_sink_setcaps (GstPad * pad, GstObject * parent,
|
|
GstCaps * caps)
|
|
{
|
|
GstVideoMixer2 *mix;
|
|
GstVideoMixer2Pad *mixpad;
|
|
GstVideoInfo info;
|
|
gboolean ret = FALSE;
|
|
|
|
GST_INFO_OBJECT (pad, "Setting caps %" GST_PTR_FORMAT, caps);
|
|
|
|
mix = GST_VIDEO_MIXER2 (parent);
|
|
mixpad = GST_VIDEO_MIXER2_PAD (pad);
|
|
|
|
if (!gst_video_info_from_caps (&info, caps)) {
|
|
GST_ERROR_OBJECT (pad, "Failed to parse caps");
|
|
goto beach;
|
|
}
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
if (GST_VIDEO_INFO_FORMAT (&mix->info) != GST_VIDEO_FORMAT_UNKNOWN) {
|
|
if (GST_VIDEO_INFO_PAR_N (&mix->info) != GST_VIDEO_INFO_PAR_N (&info)
|
|
|| GST_VIDEO_INFO_PAR_D (&mix->info) != GST_VIDEO_INFO_PAR_D (&info) ||
|
|
GST_VIDEO_INFO_INTERLACE_MODE (&mix->info) !=
|
|
GST_VIDEO_INFO_INTERLACE_MODE (&info)) {
|
|
GST_DEBUG_OBJECT (pad,
|
|
"got input caps %" GST_PTR_FORMAT ", but " "current caps are %"
|
|
GST_PTR_FORMAT, caps, mix->current_caps);
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
mixpad->info = info;
|
|
|
|
GST_COLLECT_PADS_STREAM_LOCK (mix->collect);
|
|
|
|
ret = gst_videomixer2_update_converters (mix);
|
|
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
if (ret)
|
|
ret = gst_videomixer2_update_src_caps (mix);
|
|
GST_COLLECT_PADS_STREAM_UNLOCK (mix->collect);
|
|
|
|
beach:
|
|
return ret;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_videomixer2_pad_sink_getcaps (GstPad * pad, GstVideoMixer2 * mix,
|
|
GstCaps * filter)
|
|
{
|
|
GstCaps *srccaps;
|
|
GstCaps *template_caps;
|
|
GstCaps *filtered_caps;
|
|
GstCaps *returned_caps;
|
|
GstStructure *s;
|
|
gboolean had_current_caps = TRUE;
|
|
gint i, n;
|
|
|
|
template_caps = gst_pad_get_pad_template_caps (GST_PAD (mix->srcpad));
|
|
|
|
srccaps = gst_pad_get_current_caps (GST_PAD (mix->srcpad));
|
|
if (srccaps == NULL) {
|
|
had_current_caps = FALSE;
|
|
srccaps = template_caps;
|
|
}
|
|
|
|
srccaps = gst_caps_make_writable (srccaps);
|
|
|
|
n = gst_caps_get_size (srccaps);
|
|
for (i = 0; i < n; i++) {
|
|
s = gst_caps_get_structure (srccaps, i);
|
|
gst_structure_set (s, "width", GST_TYPE_INT_RANGE, 1, G_MAXINT,
|
|
"height", GST_TYPE_INT_RANGE, 1, G_MAXINT,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL);
|
|
if (!gst_structure_has_field (s, "pixel-aspect-ratio"))
|
|
gst_structure_set (s, "pixel-aspect-ratio", GST_TYPE_FRACTION, 1, 1,
|
|
NULL);
|
|
|
|
gst_structure_remove_fields (s, "colorimetry", "chroma-site", "format",
|
|
NULL);
|
|
}
|
|
|
|
filtered_caps = srccaps;
|
|
if (filter)
|
|
filtered_caps = gst_caps_intersect (srccaps, filter);
|
|
returned_caps = gst_caps_intersect (filtered_caps, template_caps);
|
|
|
|
gst_caps_unref (srccaps);
|
|
if (filter)
|
|
gst_caps_unref (filtered_caps);
|
|
if (had_current_caps)
|
|
gst_caps_unref (template_caps);
|
|
|
|
return returned_caps;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_pad_sink_acceptcaps (GstPad * pad, GstVideoMixer2 * mix,
|
|
GstCaps * caps)
|
|
{
|
|
gboolean ret;
|
|
GstCaps *modified_caps;
|
|
GstCaps *accepted_caps;
|
|
GstCaps *template_caps;
|
|
gboolean had_current_caps = TRUE;
|
|
gint i, n;
|
|
GstStructure *s;
|
|
|
|
GST_DEBUG_OBJECT (pad, "%" GST_PTR_FORMAT, caps);
|
|
|
|
accepted_caps = gst_pad_get_current_caps (GST_PAD (mix->srcpad));
|
|
|
|
template_caps = gst_pad_get_pad_template_caps (GST_PAD (mix->srcpad));
|
|
|
|
if (accepted_caps == NULL) {
|
|
accepted_caps = template_caps;
|
|
had_current_caps = FALSE;
|
|
}
|
|
|
|
accepted_caps = gst_caps_make_writable (accepted_caps);
|
|
|
|
GST_LOG_OBJECT (pad, "src caps %" GST_PTR_FORMAT, accepted_caps);
|
|
|
|
n = gst_caps_get_size (accepted_caps);
|
|
for (i = 0; i < n; i++) {
|
|
s = gst_caps_get_structure (accepted_caps, i);
|
|
gst_structure_set (s, "width", GST_TYPE_INT_RANGE, 1, G_MAXINT,
|
|
"height", GST_TYPE_INT_RANGE, 1, G_MAXINT,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL);
|
|
if (!gst_structure_has_field (s, "pixel-aspect-ratio"))
|
|
gst_structure_set (s, "pixel-aspect-ratio", GST_TYPE_FRACTION, 1, 1,
|
|
NULL);
|
|
|
|
gst_structure_remove_fields (s, "colorimetry", "chroma-site", "format",
|
|
NULL);
|
|
}
|
|
|
|
modified_caps = gst_caps_intersect (accepted_caps, template_caps);
|
|
|
|
ret = gst_caps_can_intersect (caps, accepted_caps);
|
|
GST_DEBUG_OBJECT (pad, "%saccepted caps %" GST_PTR_FORMAT,
|
|
(ret ? "" : "not "), caps);
|
|
GST_DEBUG_OBJECT (pad, "acceptable caps are %" GST_PTR_FORMAT, accepted_caps);
|
|
gst_caps_unref (accepted_caps);
|
|
gst_caps_unref (modified_caps);
|
|
if (had_current_caps)
|
|
gst_caps_unref (template_caps);
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_sink_query (GstCollectPads * pads, GstCollectData * cdata,
|
|
GstQuery * query, GstVideoMixer2 * mix)
|
|
{
|
|
GstVideoMixer2Pad *pad = GST_VIDEO_MIXER2_PAD (cdata->pad);
|
|
gboolean ret = FALSE;
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
case GST_QUERY_CAPS:
|
|
{
|
|
GstCaps *filter, *caps;
|
|
|
|
gst_query_parse_caps (query, &filter);
|
|
caps = gst_videomixer2_pad_sink_getcaps (GST_PAD (pad), mix, filter);
|
|
gst_query_set_caps_result (query, caps);
|
|
gst_caps_unref (caps);
|
|
ret = TRUE;
|
|
break;
|
|
}
|
|
case GST_QUERY_ACCEPT_CAPS:
|
|
{
|
|
GstCaps *caps;
|
|
|
|
gst_query_parse_accept_caps (query, &caps);
|
|
ret = gst_videomixer2_pad_sink_acceptcaps (GST_PAD (pad), mix, caps);
|
|
gst_query_set_accept_caps_result (query, ret);
|
|
ret = TRUE;
|
|
break;
|
|
}
|
|
default:
|
|
ret = gst_collect_pads_query_default (pads, cdata, query, FALSE);
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_pad_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstVideoMixer2Pad *pad = GST_VIDEO_MIXER2_PAD (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_PAD_ZORDER:
|
|
g_value_set_uint (value, pad->zorder);
|
|
break;
|
|
case PROP_PAD_XPOS:
|
|
g_value_set_int (value, pad->xpos);
|
|
break;
|
|
case PROP_PAD_YPOS:
|
|
g_value_set_int (value, pad->ypos);
|
|
break;
|
|
case PROP_PAD_ALPHA:
|
|
g_value_set_double (value, pad->alpha);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static int
|
|
pad_zorder_compare (const GstVideoMixer2Pad * pad1,
|
|
const GstVideoMixer2Pad * pad2)
|
|
{
|
|
return pad1->zorder - pad2->zorder;
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_pad_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstVideoMixer2Pad *pad = GST_VIDEO_MIXER2_PAD (object);
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (gst_pad_get_parent (GST_PAD (pad)));
|
|
|
|
switch (prop_id) {
|
|
case PROP_PAD_ZORDER:
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
pad->zorder = g_value_get_uint (value);
|
|
|
|
mix->sinkpads = g_slist_sort (mix->sinkpads,
|
|
(GCompareFunc) pad_zorder_compare);
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
break;
|
|
case PROP_PAD_XPOS:
|
|
pad->xpos = g_value_get_int (value);
|
|
break;
|
|
case PROP_PAD_YPOS:
|
|
pad->ypos = g_value_get_int (value);
|
|
break;
|
|
case PROP_PAD_ALPHA:
|
|
pad->alpha = g_value_get_double (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
|
|
gst_object_unref (mix);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_pad_class_init (GstVideoMixer2PadClass * klass)
|
|
{
|
|
GObjectClass *gobject_class = (GObjectClass *) klass;
|
|
|
|
gobject_class->set_property = gst_videomixer2_pad_set_property;
|
|
gobject_class->get_property = gst_videomixer2_pad_get_property;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_PAD_ZORDER,
|
|
g_param_spec_uint ("zorder", "Z-Order", "Z Order of the picture",
|
|
0, 10000, DEFAULT_PAD_ZORDER,
|
|
G_PARAM_READWRITE | GST_PARAM_CONTROLLABLE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_PAD_XPOS,
|
|
g_param_spec_int ("xpos", "X Position", "X Position of the picture",
|
|
G_MININT, G_MAXINT, DEFAULT_PAD_XPOS,
|
|
G_PARAM_READWRITE | GST_PARAM_CONTROLLABLE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_PAD_YPOS,
|
|
g_param_spec_int ("ypos", "Y Position", "Y Position of the picture",
|
|
G_MININT, G_MAXINT, DEFAULT_PAD_YPOS,
|
|
G_PARAM_READWRITE | GST_PARAM_CONTROLLABLE | G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_PAD_ALPHA,
|
|
g_param_spec_double ("alpha", "Alpha", "Alpha of the picture", 0.0, 1.0,
|
|
DEFAULT_PAD_ALPHA,
|
|
G_PARAM_READWRITE | GST_PARAM_CONTROLLABLE | G_PARAM_STATIC_STRINGS));
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_pad_init (GstVideoMixer2Pad * mixerpad)
|
|
{
|
|
mixerpad->zorder = DEFAULT_PAD_ZORDER;
|
|
mixerpad->xpos = DEFAULT_PAD_XPOS;
|
|
mixerpad->ypos = DEFAULT_PAD_YPOS;
|
|
mixerpad->alpha = DEFAULT_PAD_ALPHA;
|
|
mixerpad->convert = NULL;
|
|
mixerpad->need_conversion_update = FALSE;
|
|
}
|
|
|
|
/* GstVideoMixer2 */
|
|
#define DEFAULT_BACKGROUND VIDEO_MIXER2_BACKGROUND_CHECKER
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_BACKGROUND
|
|
};
|
|
|
|
#define GST_TYPE_VIDEO_MIXER2_BACKGROUND (gst_videomixer2_background_get_type())
|
|
static GType
|
|
gst_videomixer2_background_get_type (void)
|
|
{
|
|
static GType video_mixer_background_type = 0;
|
|
|
|
static const GEnumValue video_mixer_background[] = {
|
|
{VIDEO_MIXER2_BACKGROUND_CHECKER, "Checker pattern", "checker"},
|
|
{VIDEO_MIXER2_BACKGROUND_BLACK, "Black", "black"},
|
|
{VIDEO_MIXER2_BACKGROUND_WHITE, "White", "white"},
|
|
{VIDEO_MIXER2_BACKGROUND_TRANSPARENT,
|
|
"Transparent Background to enable further mixing", "transparent"},
|
|
{0, NULL, NULL},
|
|
};
|
|
|
|
if (!video_mixer_background_type) {
|
|
video_mixer_background_type =
|
|
g_enum_register_static ("GstVideoMixer2Background",
|
|
video_mixer_background);
|
|
}
|
|
return video_mixer_background_type;
|
|
}
|
|
|
|
#define gst_videomixer2_parent_class parent_class
|
|
G_DEFINE_TYPE_WITH_CODE (GstVideoMixer2, gst_videomixer2, GST_TYPE_ELEMENT,
|
|
G_IMPLEMENT_INTERFACE (GST_TYPE_CHILD_PROXY,
|
|
gst_videomixer2_child_proxy_init));
|
|
|
|
static void
|
|
gst_videomixer2_update_qos (GstVideoMixer2 * mix, gdouble proportion,
|
|
GstClockTimeDiff diff, GstClockTime timestamp)
|
|
{
|
|
GST_DEBUG_OBJECT (mix,
|
|
"Updating QoS: proportion %lf, diff %" GST_STIME_FORMAT ", timestamp %"
|
|
GST_TIME_FORMAT, proportion, GST_STIME_ARGS (diff),
|
|
GST_TIME_ARGS (timestamp));
|
|
|
|
GST_OBJECT_LOCK (mix);
|
|
mix->proportion = proportion;
|
|
if (G_LIKELY (timestamp != GST_CLOCK_TIME_NONE)) {
|
|
if (!mix->live && G_UNLIKELY (diff > 0))
|
|
mix->earliest_time =
|
|
timestamp + MIN (2 * diff,
|
|
GST_SECOND) + gst_util_uint64_scale_int_round (GST_SECOND,
|
|
GST_VIDEO_INFO_FPS_D (&mix->info), GST_VIDEO_INFO_FPS_N (&mix->info));
|
|
else
|
|
mix->earliest_time = timestamp + diff;
|
|
} else {
|
|
mix->earliest_time = GST_CLOCK_TIME_NONE;
|
|
}
|
|
GST_OBJECT_UNLOCK (mix);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_reset_qos (GstVideoMixer2 * mix)
|
|
{
|
|
gst_videomixer2_update_qos (mix, 0.5, 0, GST_CLOCK_TIME_NONE);
|
|
mix->qos_processed = mix->qos_dropped = 0;
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_read_qos (GstVideoMixer2 * mix, gdouble * proportion,
|
|
GstClockTime * time)
|
|
{
|
|
GST_OBJECT_LOCK (mix);
|
|
*proportion = mix->proportion;
|
|
*time = mix->earliest_time;
|
|
GST_OBJECT_UNLOCK (mix);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_reset (GstVideoMixer2 * mix)
|
|
{
|
|
GSList *l;
|
|
|
|
gst_video_info_init (&mix->info);
|
|
mix->ts_offset = 0;
|
|
mix->nframes = 0;
|
|
|
|
gst_segment_init (&mix->segment, GST_FORMAT_TIME);
|
|
mix->segment.position = -1;
|
|
|
|
gst_videomixer2_reset_qos (mix);
|
|
|
|
for (l = mix->sinkpads; l; l = l->next) {
|
|
GstVideoMixer2Pad *p = l->data;
|
|
GstVideoMixer2Collect *mixcol = p->mixcol;
|
|
|
|
gst_buffer_replace (&mixcol->buffer, NULL);
|
|
mixcol->start_time = -1;
|
|
mixcol->end_time = -1;
|
|
|
|
gst_video_info_init (&p->info);
|
|
}
|
|
|
|
mix->newseg_pending = TRUE;
|
|
}
|
|
|
|
/* 1 == OK
|
|
* 0 == need more data
|
|
* -1 == EOS
|
|
* -2 == error
|
|
*/
|
|
static gint
|
|
gst_videomixer2_fill_queues (GstVideoMixer2 * mix,
|
|
GstClockTime output_start_time, GstClockTime output_end_time)
|
|
{
|
|
GSList *l;
|
|
gboolean eos = TRUE;
|
|
gboolean need_more_data = FALSE;
|
|
|
|
for (l = mix->sinkpads; l; l = l->next) {
|
|
GstVideoMixer2Pad *pad = l->data;
|
|
GstVideoMixer2Collect *mixcol = pad->mixcol;
|
|
GstSegment *segment = &pad->mixcol->collect.segment;
|
|
GstBuffer *buf;
|
|
GstVideoInfo *vinfo;
|
|
|
|
buf = gst_collect_pads_peek (mix->collect, &mixcol->collect);
|
|
if (buf) {
|
|
GstClockTime start_time, end_time;
|
|
|
|
start_time = GST_BUFFER_TIMESTAMP (buf);
|
|
if (start_time == -1) {
|
|
gst_buffer_unref (buf);
|
|
GST_ERROR_OBJECT (pad, "Need timestamped buffers!");
|
|
return -2;
|
|
}
|
|
|
|
vinfo = &pad->info;
|
|
|
|
/* FIXME: Make all this work with negative rates */
|
|
|
|
if ((mixcol->buffer && start_time < GST_BUFFER_TIMESTAMP (mixcol->buffer))
|
|
|| (mixcol->queued
|
|
&& start_time < GST_BUFFER_TIMESTAMP (mixcol->queued))) {
|
|
GST_WARNING_OBJECT (pad, "Buffer from the past, dropping");
|
|
gst_buffer_unref (buf);
|
|
buf = gst_collect_pads_pop (mix->collect, &mixcol->collect);
|
|
gst_buffer_unref (buf);
|
|
need_more_data = TRUE;
|
|
continue;
|
|
}
|
|
|
|
if (mixcol->queued) {
|
|
end_time = start_time - GST_BUFFER_TIMESTAMP (mixcol->queued);
|
|
start_time = GST_BUFFER_TIMESTAMP (mixcol->queued);
|
|
gst_buffer_unref (buf);
|
|
buf = gst_buffer_ref (mixcol->queued);
|
|
vinfo = &mixcol->queued_vinfo;
|
|
} else {
|
|
end_time = GST_BUFFER_DURATION (buf);
|
|
|
|
if (end_time == -1) {
|
|
mixcol->queued = buf;
|
|
buf = gst_collect_pads_pop (mix->collect, &mixcol->collect);
|
|
gst_buffer_unref (buf);
|
|
mixcol->queued_vinfo = pad->info;
|
|
need_more_data = TRUE;
|
|
continue;
|
|
}
|
|
}
|
|
|
|
g_assert (start_time != -1 && end_time != -1);
|
|
end_time += start_time; /* convert from duration to position */
|
|
|
|
/* Check if it's inside the segment */
|
|
if (start_time >= segment->stop || end_time < segment->start) {
|
|
GST_DEBUG_OBJECT (pad, "Buffer outside the segment");
|
|
|
|
if (buf == mixcol->queued) {
|
|
gst_buffer_unref (buf);
|
|
gst_buffer_replace (&mixcol->queued, NULL);
|
|
} else {
|
|
gst_buffer_unref (buf);
|
|
buf = gst_collect_pads_pop (mix->collect, &mixcol->collect);
|
|
gst_buffer_unref (buf);
|
|
}
|
|
|
|
need_more_data = TRUE;
|
|
continue;
|
|
}
|
|
|
|
/* Clip to segment and convert to running time */
|
|
start_time = MAX (start_time, segment->start);
|
|
if (segment->stop != -1)
|
|
end_time = MIN (end_time, segment->stop);
|
|
start_time =
|
|
gst_segment_to_running_time (segment, GST_FORMAT_TIME, start_time);
|
|
end_time =
|
|
gst_segment_to_running_time (segment, GST_FORMAT_TIME, end_time);
|
|
g_assert (start_time != -1 && end_time != -1);
|
|
|
|
/* Convert to the output segment rate */
|
|
if (ABS (mix->segment.rate) != 1.0) {
|
|
start_time *= ABS (mix->segment.rate);
|
|
end_time *= ABS (mix->segment.rate);
|
|
}
|
|
|
|
if (mixcol->end_time != -1 && mixcol->end_time > end_time) {
|
|
GST_DEBUG_OBJECT (pad, "Buffer from the past, dropping");
|
|
if (buf == mixcol->queued) {
|
|
gst_buffer_unref (buf);
|
|
gst_buffer_replace (&mixcol->queued, NULL);
|
|
} else {
|
|
gst_buffer_unref (buf);
|
|
buf = gst_collect_pads_pop (mix->collect, &mixcol->collect);
|
|
gst_buffer_unref (buf);
|
|
}
|
|
|
|
need_more_data = TRUE;
|
|
continue;
|
|
}
|
|
|
|
if (end_time >= output_start_time && start_time < output_end_time) {
|
|
GST_DEBUG_OBJECT (pad,
|
|
"Taking new buffer with start time %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (start_time));
|
|
gst_buffer_replace (&mixcol->buffer, buf);
|
|
mixcol->buffer_vinfo = *vinfo;
|
|
mixcol->start_time = start_time;
|
|
mixcol->end_time = end_time;
|
|
|
|
if (buf == mixcol->queued) {
|
|
gst_buffer_unref (buf);
|
|
gst_buffer_replace (&mixcol->queued, NULL);
|
|
} else {
|
|
gst_buffer_unref (buf);
|
|
buf = gst_collect_pads_pop (mix->collect, &mixcol->collect);
|
|
gst_buffer_unref (buf);
|
|
}
|
|
eos = FALSE;
|
|
} else if (start_time >= output_end_time) {
|
|
GST_DEBUG_OBJECT (pad, "Keeping buffer until %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (start_time));
|
|
gst_buffer_unref (buf);
|
|
eos = FALSE;
|
|
} else {
|
|
GST_DEBUG_OBJECT (pad, "Too old buffer -- dropping");
|
|
if (buf == mixcol->queued) {
|
|
gst_buffer_unref (buf);
|
|
gst_buffer_replace (&mixcol->queued, NULL);
|
|
} else {
|
|
gst_buffer_unref (buf);
|
|
buf = gst_collect_pads_pop (mix->collect, &mixcol->collect);
|
|
gst_buffer_unref (buf);
|
|
}
|
|
|
|
need_more_data = TRUE;
|
|
continue;
|
|
}
|
|
} else {
|
|
if (mixcol->end_time != -1) {
|
|
if (mixcol->end_time <= output_start_time) {
|
|
gst_buffer_replace (&mixcol->buffer, NULL);
|
|
mixcol->start_time = mixcol->end_time = -1;
|
|
if (!GST_COLLECT_PADS_STATE_IS_SET (mixcol,
|
|
GST_COLLECT_PADS_STATE_EOS))
|
|
need_more_data = TRUE;
|
|
} else if (!GST_COLLECT_PADS_STATE_IS_SET (mixcol,
|
|
GST_COLLECT_PADS_STATE_EOS)) {
|
|
eos = FALSE;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
if (need_more_data)
|
|
return 0;
|
|
if (eos)
|
|
return -1;
|
|
|
|
return 1;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_videomixer2_blend_buffers (GstVideoMixer2 * mix,
|
|
GstClockTime output_start_time, GstClockTime output_end_time,
|
|
GstBuffer ** outbuf)
|
|
{
|
|
GSList *l;
|
|
guint outsize;
|
|
BlendFunction composite;
|
|
GstVideoFrame outframe;
|
|
static GstAllocationParams params = { 0, 15, 0, 0, };
|
|
|
|
outsize = GST_VIDEO_INFO_SIZE (&mix->info);
|
|
|
|
*outbuf = gst_buffer_new_allocate (NULL, outsize, ¶ms);
|
|
GST_BUFFER_TIMESTAMP (*outbuf) = output_start_time;
|
|
GST_BUFFER_DURATION (*outbuf) = output_end_time - output_start_time;
|
|
|
|
gst_video_frame_map (&outframe, &mix->info, *outbuf, GST_MAP_READWRITE);
|
|
|
|
/* default to blending */
|
|
composite = mix->blend;
|
|
switch (mix->background) {
|
|
case VIDEO_MIXER2_BACKGROUND_CHECKER:
|
|
mix->fill_checker (&outframe);
|
|
break;
|
|
case VIDEO_MIXER2_BACKGROUND_BLACK:
|
|
mix->fill_color (&outframe, 16, 128, 128);
|
|
break;
|
|
case VIDEO_MIXER2_BACKGROUND_WHITE:
|
|
mix->fill_color (&outframe, 240, 128, 128);
|
|
break;
|
|
case VIDEO_MIXER2_BACKGROUND_TRANSPARENT:
|
|
{
|
|
guint i, plane, num_planes, height;
|
|
|
|
num_planes = GST_VIDEO_FRAME_N_PLANES (&outframe);
|
|
for (plane = 0; plane < num_planes; ++plane) {
|
|
guint8 *pdata;
|
|
gsize rowsize, plane_stride;
|
|
|
|
pdata = GST_VIDEO_FRAME_PLANE_DATA (&outframe, plane);
|
|
plane_stride = GST_VIDEO_FRAME_PLANE_STRIDE (&outframe, plane);
|
|
rowsize = GST_VIDEO_FRAME_COMP_WIDTH (&outframe, plane)
|
|
* GST_VIDEO_FRAME_COMP_PSTRIDE (&outframe, plane);
|
|
height = GST_VIDEO_FRAME_COMP_HEIGHT (&outframe, plane);
|
|
for (i = 0; i < height; ++i) {
|
|
memset (pdata, 0, rowsize);
|
|
pdata += plane_stride;
|
|
}
|
|
}
|
|
|
|
/* use overlay to keep background transparent */
|
|
composite = mix->overlay;
|
|
break;
|
|
}
|
|
}
|
|
|
|
for (l = mix->sinkpads; l; l = l->next) {
|
|
GstVideoMixer2Pad *pad = l->data;
|
|
GstVideoMixer2Collect *mixcol = pad->mixcol;
|
|
|
|
if (mixcol->buffer != NULL) {
|
|
GstClockTime timestamp;
|
|
gint64 stream_time;
|
|
GstSegment *seg;
|
|
GstVideoFrame converted_frame;
|
|
GstBuffer *converted_buf = NULL;
|
|
GstVideoFrame frame;
|
|
|
|
seg = &mixcol->collect.segment;
|
|
|
|
timestamp = GST_BUFFER_TIMESTAMP (mixcol->buffer);
|
|
|
|
stream_time =
|
|
gst_segment_to_stream_time (seg, GST_FORMAT_TIME, timestamp);
|
|
|
|
/* sync object properties on stream time */
|
|
if (GST_CLOCK_TIME_IS_VALID (stream_time))
|
|
gst_object_sync_values (GST_OBJECT (pad), stream_time);
|
|
|
|
gst_video_frame_map (&frame, &mixcol->buffer_vinfo, mixcol->buffer,
|
|
GST_MAP_READ);
|
|
|
|
if (pad->convert) {
|
|
gint converted_size;
|
|
|
|
/* We wait until here to set the conversion infos, in case mix->info changed */
|
|
if (pad->need_conversion_update) {
|
|
pad->conversion_info = mix->info;
|
|
gst_video_info_set_format (&(pad->conversion_info),
|
|
GST_VIDEO_INFO_FORMAT (&mix->info), pad->info.width,
|
|
pad->info.height);
|
|
pad->need_conversion_update = FALSE;
|
|
}
|
|
|
|
converted_size = pad->conversion_info.size;
|
|
converted_size = converted_size > outsize ? converted_size : outsize;
|
|
converted_buf = gst_buffer_new_allocate (NULL, converted_size, ¶ms);
|
|
|
|
gst_video_frame_map (&converted_frame, &(pad->conversion_info),
|
|
converted_buf, GST_MAP_READWRITE);
|
|
gst_video_converter_frame (pad->convert, &frame, &converted_frame);
|
|
gst_video_frame_unmap (&frame);
|
|
} else {
|
|
converted_frame = frame;
|
|
}
|
|
|
|
composite (&converted_frame, pad->xpos, pad->ypos, pad->alpha, &outframe);
|
|
|
|
if (pad->convert)
|
|
gst_buffer_unref (converted_buf);
|
|
|
|
gst_video_frame_unmap (&converted_frame);
|
|
}
|
|
}
|
|
gst_video_frame_unmap (&outframe);
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
/* Perform qos calculations before processing the next frame. Returns TRUE if
|
|
* the frame should be processed, FALSE if the frame can be dropped entirely */
|
|
static gint64
|
|
gst_videomixer2_do_qos (GstVideoMixer2 * mix, GstClockTime timestamp)
|
|
{
|
|
GstClockTime qostime, earliest_time;
|
|
gdouble proportion;
|
|
gint64 jitter;
|
|
|
|
/* no timestamp, can't do QoS => process frame */
|
|
if (G_UNLIKELY (!GST_CLOCK_TIME_IS_VALID (timestamp))) {
|
|
GST_LOG_OBJECT (mix, "invalid timestamp, can't do QoS, process frame");
|
|
return -1;
|
|
}
|
|
|
|
/* get latest QoS observation values */
|
|
gst_videomixer2_read_qos (mix, &proportion, &earliest_time);
|
|
|
|
/* skip qos if we have no observation (yet) => process frame */
|
|
if (G_UNLIKELY (!GST_CLOCK_TIME_IS_VALID (earliest_time))) {
|
|
GST_LOG_OBJECT (mix, "no observation yet, process frame");
|
|
return -1;
|
|
}
|
|
|
|
/* qos is done on running time */
|
|
qostime =
|
|
gst_segment_to_running_time (&mix->segment, GST_FORMAT_TIME, timestamp);
|
|
|
|
/* see how our next timestamp relates to the latest qos timestamp */
|
|
GST_LOG_OBJECT (mix, "qostime %" GST_TIME_FORMAT ", earliest %"
|
|
GST_TIME_FORMAT, GST_TIME_ARGS (qostime), GST_TIME_ARGS (earliest_time));
|
|
|
|
jitter = GST_CLOCK_DIFF (qostime, earliest_time);
|
|
if (qostime != GST_CLOCK_TIME_NONE && jitter > 0) {
|
|
GST_DEBUG_OBJECT (mix, "we are late, drop frame");
|
|
return jitter;
|
|
}
|
|
|
|
GST_LOG_OBJECT (mix, "process frame");
|
|
return jitter;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_videomixer2_collected (GstCollectPads * pads, GstVideoMixer2 * mix)
|
|
{
|
|
GstFlowReturn ret;
|
|
GstClockTime output_start_time, output_end_time;
|
|
GstBuffer *outbuf = NULL;
|
|
gint res;
|
|
gint64 jitter;
|
|
|
|
/* If we're not negotiated yet... */
|
|
if (GST_VIDEO_INFO_FORMAT (&mix->info) == GST_VIDEO_FORMAT_UNKNOWN)
|
|
return GST_FLOW_NOT_NEGOTIATED;
|
|
|
|
if (mix->send_stream_start) {
|
|
gchar s_id[32];
|
|
|
|
/* stream-start (FIXME: create id based on input ids) */
|
|
g_snprintf (s_id, sizeof (s_id), "mix-%08x", g_random_int ());
|
|
if (!gst_pad_push_event (mix->srcpad, gst_event_new_stream_start (s_id))) {
|
|
GST_WARNING_OBJECT (mix->srcpad, "Sending stream start event failed");
|
|
}
|
|
mix->send_stream_start = FALSE;
|
|
}
|
|
|
|
if (gst_pad_check_reconfigure (mix->srcpad))
|
|
gst_videomixer2_update_src_caps (mix);
|
|
|
|
if (mix->send_caps) {
|
|
if (!gst_pad_push_event (mix->srcpad,
|
|
gst_event_new_caps (mix->current_caps))) {
|
|
GST_WARNING_OBJECT (mix->srcpad, "Sending caps event failed");
|
|
}
|
|
mix->send_caps = FALSE;
|
|
}
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
|
|
if (mix->newseg_pending) {
|
|
GST_DEBUG_OBJECT (mix, "Sending NEWSEGMENT event");
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
if (!gst_pad_push_event (mix->srcpad,
|
|
gst_event_new_segment (&mix->segment))) {
|
|
ret = GST_FLOW_ERROR;
|
|
goto done_unlocked;
|
|
}
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
mix->newseg_pending = FALSE;
|
|
}
|
|
|
|
if (mix->segment.position == -1)
|
|
output_start_time = mix->segment.start;
|
|
else
|
|
output_start_time = mix->segment.position;
|
|
|
|
output_end_time =
|
|
mix->ts_offset + gst_util_uint64_scale_round (mix->nframes + 1,
|
|
GST_SECOND * GST_VIDEO_INFO_FPS_D (&mix->info),
|
|
GST_VIDEO_INFO_FPS_N (&mix->info)) + mix->segment.start;
|
|
|
|
if (output_end_time >= mix->segment.stop) {
|
|
GST_DEBUG_OBJECT (mix, "Segment done");
|
|
if (!(mix->segment.flags & GST_SEGMENT_FLAG_SEGMENT)) {
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
gst_pad_push_event (mix->srcpad, gst_event_new_eos ());
|
|
|
|
ret = GST_FLOW_EOS;
|
|
goto done_unlocked;
|
|
}
|
|
}
|
|
|
|
if (G_UNLIKELY (mix->pending_tags)) {
|
|
gst_pad_push_event (mix->srcpad, gst_event_new_tag (mix->pending_tags));
|
|
mix->pending_tags = NULL;
|
|
}
|
|
|
|
if (mix->segment.stop != -1)
|
|
output_end_time = MIN (output_end_time, mix->segment.stop);
|
|
|
|
res = gst_videomixer2_fill_queues (mix, output_start_time, output_end_time);
|
|
|
|
if (res == 0) {
|
|
GST_DEBUG_OBJECT (mix, "Need more data for decisions");
|
|
ret = GST_FLOW_OK;
|
|
goto done;
|
|
} else if (res == -1) {
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
GST_DEBUG_OBJECT (mix, "All sinkpads are EOS -- forwarding");
|
|
gst_pad_push_event (mix->srcpad, gst_event_new_eos ());
|
|
ret = GST_FLOW_EOS;
|
|
goto done_unlocked;
|
|
} else if (res == -2) {
|
|
GST_ERROR_OBJECT (mix, "Error collecting buffers");
|
|
ret = GST_FLOW_ERROR;
|
|
goto done;
|
|
}
|
|
|
|
jitter = gst_videomixer2_do_qos (mix, output_start_time);
|
|
if (jitter <= 0) {
|
|
ret =
|
|
gst_videomixer2_blend_buffers (mix, output_start_time,
|
|
output_end_time, &outbuf);
|
|
mix->qos_processed++;
|
|
} else {
|
|
GstMessage *msg;
|
|
|
|
mix->qos_dropped++;
|
|
|
|
/* TODO: live */
|
|
msg =
|
|
gst_message_new_qos (GST_OBJECT_CAST (mix), FALSE,
|
|
gst_segment_to_running_time (&mix->segment, GST_FORMAT_TIME,
|
|
output_start_time), gst_segment_to_stream_time (&mix->segment,
|
|
GST_FORMAT_TIME, output_start_time), output_start_time,
|
|
output_end_time - output_start_time);
|
|
gst_message_set_qos_values (msg, jitter, mix->proportion, 1000000);
|
|
gst_message_set_qos_stats (msg, GST_FORMAT_BUFFERS, mix->qos_processed,
|
|
mix->qos_dropped);
|
|
gst_element_post_message (GST_ELEMENT_CAST (mix), msg);
|
|
|
|
ret = GST_FLOW_OK;
|
|
}
|
|
|
|
mix->segment.position = output_end_time;
|
|
mix->nframes++;
|
|
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
if (outbuf) {
|
|
GST_LOG_OBJECT (mix,
|
|
"Pushing buffer with ts %" GST_TIME_FORMAT " and duration %"
|
|
GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (outbuf)),
|
|
GST_TIME_ARGS (GST_BUFFER_DURATION (outbuf)));
|
|
ret = gst_pad_push (mix->srcpad, outbuf);
|
|
}
|
|
goto done_unlocked;
|
|
|
|
done:
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
|
|
done_unlocked:
|
|
return ret;
|
|
}
|
|
|
|
/* FIXME, the duration query should reflect how long you will produce
|
|
* data, that is the amount of stream time until you will emit EOS.
|
|
*
|
|
* For synchronized mixing this is always the max of all the durations
|
|
* of upstream since we emit EOS when all of them finished.
|
|
*
|
|
* We don't do synchronized mixing so this really depends on where the
|
|
* streams where punched in and what their relative offsets are against
|
|
* each other which we can get from the first timestamps we see.
|
|
*
|
|
* When we add a new stream (or remove a stream) the duration might
|
|
* also become invalid again and we need to post a new DURATION
|
|
* message to notify this fact to the parent.
|
|
* For now we take the max of all the upstream elements so the simple
|
|
* cases work at least somewhat.
|
|
*/
|
|
static gboolean
|
|
gst_videomixer2_query_duration (GstVideoMixer2 * mix, GstQuery * query)
|
|
{
|
|
GValue item = { 0 };
|
|
gint64 max;
|
|
gboolean res;
|
|
GstFormat format;
|
|
GstIterator *it;
|
|
gboolean done;
|
|
|
|
/* parse format */
|
|
gst_query_parse_duration (query, &format, NULL);
|
|
|
|
max = -1;
|
|
res = TRUE;
|
|
done = FALSE;
|
|
|
|
/* Take maximum of all durations */
|
|
it = gst_element_iterate_sink_pads (GST_ELEMENT_CAST (mix));
|
|
while (!done) {
|
|
switch (gst_iterator_next (it, &item)) {
|
|
case GST_ITERATOR_DONE:
|
|
done = TRUE;
|
|
break;
|
|
case GST_ITERATOR_OK:
|
|
{
|
|
GstPad *pad;
|
|
gint64 duration;
|
|
|
|
pad = g_value_get_object (&item);
|
|
|
|
/* ask sink peer for duration */
|
|
res &= gst_pad_peer_query_duration (pad, format, &duration);
|
|
/* take max from all valid return values */
|
|
if (res) {
|
|
/* valid unknown length, stop searching */
|
|
if (duration == -1) {
|
|
max = duration;
|
|
done = TRUE;
|
|
}
|
|
/* else see if bigger than current max */
|
|
else if (duration > max)
|
|
max = duration;
|
|
}
|
|
g_value_reset (&item);
|
|
break;
|
|
}
|
|
case GST_ITERATOR_RESYNC:
|
|
max = -1;
|
|
res = TRUE;
|
|
gst_iterator_resync (it);
|
|
break;
|
|
default:
|
|
res = FALSE;
|
|
done = TRUE;
|
|
break;
|
|
}
|
|
}
|
|
g_value_unset (&item);
|
|
gst_iterator_free (it);
|
|
|
|
if (res) {
|
|
/* and store the max */
|
|
GST_DEBUG_OBJECT (mix, "Total duration in format %s: %"
|
|
GST_TIME_FORMAT, gst_format_get_name (format), GST_TIME_ARGS (max));
|
|
gst_query_set_duration (query, format, max);
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_src_query (GstPad * pad, GstObject * parent, GstQuery * query)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (parent);
|
|
gboolean res = FALSE;
|
|
|
|
switch (GST_QUERY_TYPE (query)) {
|
|
case GST_QUERY_POSITION:
|
|
{
|
|
GstFormat format;
|
|
|
|
gst_query_parse_position (query, &format, NULL);
|
|
|
|
switch (format) {
|
|
case GST_FORMAT_TIME:
|
|
gst_query_set_position (query, format,
|
|
gst_segment_to_stream_time (&mix->segment, GST_FORMAT_TIME,
|
|
mix->segment.position));
|
|
res = TRUE;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
break;
|
|
}
|
|
case GST_QUERY_DURATION:
|
|
res = gst_videomixer2_query_duration (mix, query);
|
|
break;
|
|
case GST_QUERY_CAPS:
|
|
res = gst_pad_query_default (pad, parent, query);
|
|
break;
|
|
default:
|
|
/* FIXME, needs a custom query handler because we have multiple
|
|
* sinkpads */
|
|
res = FALSE;
|
|
break;
|
|
}
|
|
return res;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_src_event (GstPad * pad, GstObject * parent, GstEvent * event)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (parent);
|
|
gboolean result;
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_QOS:
|
|
{
|
|
GstQOSType type;
|
|
GstClockTimeDiff diff;
|
|
GstClockTime timestamp;
|
|
gdouble proportion;
|
|
|
|
gst_event_parse_qos (event, &type, &proportion, &diff, ×tamp);
|
|
|
|
gst_videomixer2_update_qos (mix, proportion, diff, timestamp);
|
|
|
|
result = gst_videomixer2_push_sink_event (mix, event);
|
|
break;
|
|
}
|
|
case GST_EVENT_SEEK:
|
|
{
|
|
gdouble rate;
|
|
GstFormat fmt;
|
|
GstSeekFlags flags;
|
|
GstSeekType start_type, stop_type;
|
|
gint64 start, stop;
|
|
GSList *l;
|
|
gdouble abs_rate;
|
|
|
|
/* parse the seek parameters */
|
|
gst_event_parse_seek (event, &rate, &fmt, &flags, &start_type,
|
|
&start, &stop_type, &stop);
|
|
|
|
if (rate <= 0.0) {
|
|
GST_ERROR_OBJECT (mix, "Negative rates not supported yet");
|
|
result = FALSE;
|
|
gst_event_unref (event);
|
|
break;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (mix, "Handling SEEK event");
|
|
|
|
abs_rate = ABS (rate);
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
for (l = mix->sinkpads; l; l = l->next) {
|
|
GstVideoMixer2Pad *p = l->data;
|
|
|
|
if (flags & GST_SEEK_FLAG_FLUSH) {
|
|
gst_buffer_replace (&p->mixcol->buffer, NULL);
|
|
p->mixcol->start_time = p->mixcol->end_time = -1;
|
|
continue;
|
|
}
|
|
|
|
/* Convert to the output segment rate */
|
|
if (ABS (mix->segment.rate) != abs_rate) {
|
|
if (ABS (mix->segment.rate) != 1.0 && p->mixcol->buffer) {
|
|
p->mixcol->start_time /= ABS (mix->segment.rate);
|
|
p->mixcol->end_time /= ABS (mix->segment.rate);
|
|
}
|
|
if (abs_rate != 1.0 && p->mixcol->buffer) {
|
|
p->mixcol->start_time *= abs_rate;
|
|
p->mixcol->end_time *= abs_rate;
|
|
}
|
|
}
|
|
}
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
|
|
gst_segment_do_seek (&mix->segment, rate, fmt, flags, start_type, start,
|
|
stop_type, stop, NULL);
|
|
mix->segment.position = -1;
|
|
mix->ts_offset = 0;
|
|
mix->nframes = 0;
|
|
mix->newseg_pending = TRUE;
|
|
|
|
gst_videomixer2_reset_qos (mix);
|
|
|
|
result = gst_collect_pads_src_event_default (mix->collect, pad, event);
|
|
break;
|
|
}
|
|
case GST_EVENT_NAVIGATION:
|
|
/* navigation is rather pointless. */
|
|
result = FALSE;
|
|
gst_event_unref (event);
|
|
break;
|
|
default:
|
|
/* just forward the rest for now */
|
|
result = gst_videomixer2_push_sink_event (mix, event);
|
|
break;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_src_setcaps (GstPad * pad, GstVideoMixer2 * mix, GstCaps * caps)
|
|
{
|
|
gboolean ret = FALSE;
|
|
GstVideoInfo info;
|
|
|
|
GST_INFO_OBJECT (pad, "set src caps: %" GST_PTR_FORMAT, caps);
|
|
|
|
if (!gst_video_info_from_caps (&info, caps))
|
|
goto done;
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
|
|
mix->blend = NULL;
|
|
mix->overlay = NULL;
|
|
mix->fill_checker = NULL;
|
|
mix->fill_color = NULL;
|
|
|
|
if (GST_VIDEO_INFO_FPS_N (&mix->info) != GST_VIDEO_INFO_FPS_N (&info) ||
|
|
GST_VIDEO_INFO_FPS_D (&mix->info) != GST_VIDEO_INFO_FPS_D (&info)) {
|
|
if (mix->segment.position != -1) {
|
|
mix->ts_offset = mix->segment.position - mix->segment.start;
|
|
mix->nframes = 0;
|
|
}
|
|
gst_videomixer2_reset_qos (mix);
|
|
}
|
|
|
|
mix->info = info;
|
|
|
|
switch (GST_VIDEO_INFO_FORMAT (&mix->info)) {
|
|
case GST_VIDEO_FORMAT_AYUV:
|
|
mix->blend = gst_video_mixer_blend_ayuv;
|
|
mix->overlay = gst_video_mixer_overlay_ayuv;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_ayuv;
|
|
mix->fill_color = gst_video_mixer_fill_color_ayuv;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_ARGB:
|
|
mix->blend = gst_video_mixer_blend_argb;
|
|
mix->overlay = gst_video_mixer_overlay_argb;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_argb;
|
|
mix->fill_color = gst_video_mixer_fill_color_argb;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_BGRA:
|
|
mix->blend = gst_video_mixer_blend_bgra;
|
|
mix->overlay = gst_video_mixer_overlay_bgra;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_bgra;
|
|
mix->fill_color = gst_video_mixer_fill_color_bgra;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_ABGR:
|
|
mix->blend = gst_video_mixer_blend_abgr;
|
|
mix->overlay = gst_video_mixer_overlay_abgr;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_abgr;
|
|
mix->fill_color = gst_video_mixer_fill_color_abgr;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_RGBA:
|
|
mix->blend = gst_video_mixer_blend_rgba;
|
|
mix->overlay = gst_video_mixer_overlay_rgba;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_rgba;
|
|
mix->fill_color = gst_video_mixer_fill_color_rgba;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_Y444:
|
|
mix->blend = gst_video_mixer_blend_y444;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_y444;
|
|
mix->fill_color = gst_video_mixer_fill_color_y444;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_Y42B:
|
|
mix->blend = gst_video_mixer_blend_y42b;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_y42b;
|
|
mix->fill_color = gst_video_mixer_fill_color_y42b;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_YUY2:
|
|
mix->blend = gst_video_mixer_blend_yuy2;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_yuy2;
|
|
mix->fill_color = gst_video_mixer_fill_color_yuy2;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_UYVY:
|
|
mix->blend = gst_video_mixer_blend_uyvy;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_uyvy;
|
|
mix->fill_color = gst_video_mixer_fill_color_uyvy;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_YVYU:
|
|
mix->blend = gst_video_mixer_blend_yvyu;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_yvyu;
|
|
mix->fill_color = gst_video_mixer_fill_color_yvyu;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_I420:
|
|
mix->blend = gst_video_mixer_blend_i420;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_i420;
|
|
mix->fill_color = gst_video_mixer_fill_color_i420;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_YV12:
|
|
mix->blend = gst_video_mixer_blend_yv12;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_yv12;
|
|
mix->fill_color = gst_video_mixer_fill_color_yv12;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_NV12:
|
|
mix->blend = gst_video_mixer_blend_nv12;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_nv12;
|
|
mix->fill_color = gst_video_mixer_fill_color_nv12;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_NV21:
|
|
mix->blend = gst_video_mixer_blend_nv21;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_nv21;
|
|
mix->fill_color = gst_video_mixer_fill_color_nv21;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_Y41B:
|
|
mix->blend = gst_video_mixer_blend_y41b;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_y41b;
|
|
mix->fill_color = gst_video_mixer_fill_color_y41b;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_RGB:
|
|
mix->blend = gst_video_mixer_blend_rgb;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_rgb;
|
|
mix->fill_color = gst_video_mixer_fill_color_rgb;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_BGR:
|
|
mix->blend = gst_video_mixer_blend_bgr;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_bgr;
|
|
mix->fill_color = gst_video_mixer_fill_color_bgr;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_xRGB:
|
|
mix->blend = gst_video_mixer_blend_xrgb;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_xrgb;
|
|
mix->fill_color = gst_video_mixer_fill_color_xrgb;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_xBGR:
|
|
mix->blend = gst_video_mixer_blend_xbgr;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_xbgr;
|
|
mix->fill_color = gst_video_mixer_fill_color_xbgr;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_RGBx:
|
|
mix->blend = gst_video_mixer_blend_rgbx;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_rgbx;
|
|
mix->fill_color = gst_video_mixer_fill_color_rgbx;
|
|
ret = TRUE;
|
|
break;
|
|
case GST_VIDEO_FORMAT_BGRx:
|
|
mix->blend = gst_video_mixer_blend_bgrx;
|
|
mix->overlay = mix->blend;
|
|
mix->fill_checker = gst_video_mixer_fill_checker_bgrx;
|
|
mix->fill_color = gst_video_mixer_fill_color_bgrx;
|
|
ret = TRUE;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
|
|
if (mix->current_caps == NULL ||
|
|
gst_caps_is_equal (caps, mix->current_caps) == FALSE) {
|
|
gst_caps_replace (&mix->current_caps, caps);
|
|
mix->send_caps = TRUE;
|
|
}
|
|
|
|
done:
|
|
return ret;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_videomixer2_sink_clip (GstCollectPads * pads,
|
|
GstCollectData * data, GstBuffer * buf, GstBuffer ** outbuf,
|
|
GstVideoMixer2 * mix)
|
|
{
|
|
GstVideoMixer2Pad *pad = GST_VIDEO_MIXER2_PAD (data->pad);
|
|
GstVideoMixer2Collect *mixcol = pad->mixcol;
|
|
GstClockTime start_time, end_time;
|
|
|
|
start_time = GST_BUFFER_TIMESTAMP (buf);
|
|
if (start_time == -1) {
|
|
GST_ERROR_OBJECT (pad, "Timestamped buffers required!");
|
|
gst_buffer_unref (buf);
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
end_time = GST_BUFFER_DURATION (buf);
|
|
if (end_time == -1 && GST_VIDEO_INFO_FPS_N (&pad->info) != 0)
|
|
end_time =
|
|
gst_util_uint64_scale_int_round (GST_SECOND,
|
|
GST_VIDEO_INFO_FPS_D (&pad->info), GST_VIDEO_INFO_FPS_N (&pad->info));
|
|
if (end_time == -1) {
|
|
*outbuf = buf;
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
start_time = MAX (start_time, mixcol->collect.segment.start);
|
|
start_time =
|
|
gst_segment_to_running_time (&mixcol->collect.segment,
|
|
GST_FORMAT_TIME, start_time);
|
|
|
|
end_time += GST_BUFFER_TIMESTAMP (buf);
|
|
if (mixcol->collect.segment.stop != -1)
|
|
end_time = MIN (end_time, mixcol->collect.segment.stop);
|
|
end_time =
|
|
gst_segment_to_running_time (&mixcol->collect.segment,
|
|
GST_FORMAT_TIME, end_time);
|
|
|
|
/* Convert to the output segment rate */
|
|
if (ABS (mix->segment.rate) != 1.0) {
|
|
start_time *= ABS (mix->segment.rate);
|
|
end_time *= ABS (mix->segment.rate);
|
|
}
|
|
|
|
if (mixcol->buffer != NULL && end_time < mixcol->end_time) {
|
|
gst_buffer_unref (buf);
|
|
*outbuf = NULL;
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
*outbuf = buf;
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_flush (GstCollectPads * pads, GstVideoMixer2 * mix)
|
|
{
|
|
if (mix->pending_tags) {
|
|
gst_tag_list_unref (mix->pending_tags);
|
|
mix->pending_tags = NULL;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_sink_event (GstCollectPads * pads, GstCollectData * cdata,
|
|
GstEvent * event, GstVideoMixer2 * mix)
|
|
{
|
|
GstVideoMixer2Pad *pad = GST_VIDEO_MIXER2_PAD (cdata->pad);
|
|
gboolean ret = TRUE, discard = FALSE;
|
|
|
|
GST_DEBUG_OBJECT (pad, "Got %s event: %" GST_PTR_FORMAT,
|
|
GST_EVENT_TYPE_NAME (event), event);
|
|
|
|
switch (GST_EVENT_TYPE (event)) {
|
|
case GST_EVENT_CAPS:
|
|
{
|
|
GstCaps *caps;
|
|
|
|
gst_event_parse_caps (event, &caps);
|
|
ret =
|
|
gst_videomixer2_pad_sink_setcaps (GST_PAD (pad), GST_OBJECT (mix),
|
|
caps);
|
|
gst_event_unref (event);
|
|
event = NULL;
|
|
break;
|
|
}
|
|
case GST_EVENT_SEGMENT:{
|
|
GstSegment seg;
|
|
gst_event_copy_segment (event, &seg);
|
|
|
|
g_assert (seg.format == GST_FORMAT_TIME);
|
|
gst_videomixer2_reset_qos (mix);
|
|
break;
|
|
}
|
|
case GST_EVENT_FLUSH_STOP:
|
|
mix->newseg_pending = TRUE;
|
|
|
|
gst_videomixer2_reset_qos (mix);
|
|
gst_buffer_replace (&pad->mixcol->buffer, NULL);
|
|
pad->mixcol->start_time = -1;
|
|
pad->mixcol->end_time = -1;
|
|
|
|
mix->segment.position = -1;
|
|
mix->ts_offset = 0;
|
|
mix->nframes = 0;
|
|
break;
|
|
case GST_EVENT_TAG:
|
|
{
|
|
/* collect tags here so we can push them out when we collect data */
|
|
GstTagList *tags;
|
|
|
|
gst_event_parse_tag (event, &tags);
|
|
tags = gst_tag_list_merge (mix->pending_tags, tags, GST_TAG_MERGE_APPEND);
|
|
if (mix->pending_tags)
|
|
gst_tag_list_unref (mix->pending_tags);
|
|
mix->pending_tags = tags;
|
|
event = NULL;
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (event != NULL)
|
|
return gst_collect_pads_event_default (pads, cdata, event, discard);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
forward_event_func (GValue * item, GValue * ret, GstEvent * event)
|
|
{
|
|
GstPad *pad = g_value_get_object (item);
|
|
gst_event_ref (event);
|
|
GST_LOG_OBJECT (pad, "About to send event %s", GST_EVENT_TYPE_NAME (event));
|
|
if (!gst_pad_push_event (pad, event)) {
|
|
g_value_set_boolean (ret, FALSE);
|
|
GST_WARNING_OBJECT (pad, "Sending event %p (%s) failed.",
|
|
event, GST_EVENT_TYPE_NAME (event));
|
|
} else {
|
|
GST_LOG_OBJECT (pad, "Sent event %p (%s).",
|
|
event, GST_EVENT_TYPE_NAME (event));
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_videomixer2_push_sink_event (GstVideoMixer2 * mix, GstEvent * event)
|
|
{
|
|
GstIterator *it;
|
|
GValue vret = { 0 };
|
|
|
|
GST_LOG_OBJECT (mix, "Forwarding event %p (%s)", event,
|
|
GST_EVENT_TYPE_NAME (event));
|
|
|
|
g_value_init (&vret, G_TYPE_BOOLEAN);
|
|
g_value_set_boolean (&vret, TRUE);
|
|
it = gst_element_iterate_sink_pads (GST_ELEMENT_CAST (mix));
|
|
gst_iterator_fold (it, (GstIteratorFoldFunction) forward_event_func, &vret,
|
|
event);
|
|
gst_iterator_free (it);
|
|
gst_event_unref (event);
|
|
|
|
return g_value_get_boolean (&vret);
|
|
}
|
|
|
|
/* GstElement vmethods */
|
|
static GstStateChangeReturn
|
|
gst_videomixer2_change_state (GstElement * element, GstStateChange transition)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (element);
|
|
GstStateChangeReturn ret;
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_READY_TO_PAUSED:
|
|
mix->send_stream_start = TRUE;
|
|
mix->send_caps = TRUE;
|
|
gst_segment_init (&mix->segment, GST_FORMAT_TIME);
|
|
gst_caps_replace (&mix->current_caps, NULL);
|
|
GST_LOG_OBJECT (mix, "starting collectpads");
|
|
gst_collect_pads_start (mix->collect);
|
|
break;
|
|
case GST_STATE_CHANGE_PAUSED_TO_READY:
|
|
GST_LOG_OBJECT (mix, "stopping collectpads");
|
|
gst_collect_pads_stop (mix->collect);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_PAUSED_TO_READY:
|
|
gst_videomixer2_reset (mix);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstPad *
|
|
gst_videomixer2_request_new_pad (GstElement * element,
|
|
GstPadTemplate * templ, const gchar * req_name, const GstCaps * caps)
|
|
{
|
|
GstVideoMixer2 *mix;
|
|
GstVideoMixer2Pad *mixpad;
|
|
GstElementClass *klass = GST_ELEMENT_GET_CLASS (element);
|
|
|
|
mix = GST_VIDEO_MIXER2 (element);
|
|
|
|
if (templ == gst_element_class_get_pad_template (klass, "sink_%u")) {
|
|
guint serial = 0;
|
|
gchar *name = NULL;
|
|
GstVideoMixer2Collect *mixcol = NULL;
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
if (req_name == NULL || strlen (req_name) < 6
|
|
|| !g_str_has_prefix (req_name, "sink_")) {
|
|
/* no name given when requesting the pad, use next available int */
|
|
serial = mix->next_sinkpad++;
|
|
} else {
|
|
/* parse serial number from requested padname */
|
|
serial = g_ascii_strtoull (&req_name[5], NULL, 10);
|
|
if (serial >= mix->next_sinkpad)
|
|
mix->next_sinkpad = serial + 1;
|
|
}
|
|
/* create new pad with the name */
|
|
name = g_strdup_printf ("sink_%u", serial);
|
|
mixpad = g_object_new (GST_TYPE_VIDEO_MIXER2_PAD, "name", name, "direction",
|
|
templ->direction, "template", templ, NULL);
|
|
g_free (name);
|
|
|
|
mixpad->zorder = mix->numpads;
|
|
mixpad->xpos = DEFAULT_PAD_XPOS;
|
|
mixpad->ypos = DEFAULT_PAD_YPOS;
|
|
mixpad->alpha = DEFAULT_PAD_ALPHA;
|
|
|
|
mixcol = (GstVideoMixer2Collect *)
|
|
gst_collect_pads_add_pad (mix->collect, GST_PAD (mixpad),
|
|
sizeof (GstVideoMixer2Collect),
|
|
(GstCollectDataDestroyNotify) gst_videomixer2_collect_free, TRUE);
|
|
|
|
/* Keep track of each other */
|
|
mixcol->mixpad = mixpad;
|
|
mixpad->mixcol = mixcol;
|
|
|
|
mixcol->start_time = -1;
|
|
mixcol->end_time = -1;
|
|
|
|
/* Keep an internal list of mixpads for zordering */
|
|
mix->sinkpads = g_slist_insert_sorted (mix->sinkpads, mixpad,
|
|
(GCompareFunc) pad_zorder_compare);
|
|
mix->numpads++;
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
} else {
|
|
return NULL;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (element, "Adding pad %s", GST_PAD_NAME (mixpad));
|
|
|
|
/* add the pad to the element */
|
|
gst_element_add_pad (element, GST_PAD (mixpad));
|
|
gst_child_proxy_child_added (GST_CHILD_PROXY (mix), G_OBJECT (mixpad),
|
|
GST_OBJECT_NAME (mixpad));
|
|
|
|
return GST_PAD (mixpad);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_release_pad (GstElement * element, GstPad * pad)
|
|
{
|
|
GstVideoMixer2 *mix = NULL;
|
|
GstVideoMixer2Pad *mixpad;
|
|
gboolean update_caps;
|
|
|
|
mix = GST_VIDEO_MIXER2 (element);
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
if (G_UNLIKELY (g_slist_find (mix->sinkpads, pad) == NULL)) {
|
|
g_warning ("Unknown pad %s", GST_PAD_NAME (pad));
|
|
goto error;
|
|
}
|
|
|
|
mixpad = GST_VIDEO_MIXER2_PAD (pad);
|
|
|
|
if (mixpad->convert)
|
|
gst_video_converter_free (mixpad->convert);
|
|
mixpad->convert = NULL;
|
|
|
|
mix->sinkpads = g_slist_remove (mix->sinkpads, pad);
|
|
gst_child_proxy_child_removed (GST_CHILD_PROXY (mix), G_OBJECT (mixpad),
|
|
GST_OBJECT_NAME (mixpad));
|
|
mix->numpads--;
|
|
|
|
GST_COLLECT_PADS_STREAM_LOCK (mix->collect);
|
|
gst_videomixer2_update_converters (mix);
|
|
GST_COLLECT_PADS_STREAM_UNLOCK (mix->collect);
|
|
|
|
update_caps = GST_VIDEO_INFO_FORMAT (&mix->info) != GST_VIDEO_FORMAT_UNKNOWN;
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
|
|
gst_collect_pads_remove_pad (mix->collect, pad);
|
|
|
|
if (update_caps)
|
|
gst_videomixer2_update_src_caps (mix);
|
|
|
|
gst_element_remove_pad (element, pad);
|
|
return;
|
|
error:
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
}
|
|
|
|
/* GObject vmethods */
|
|
static void
|
|
gst_videomixer2_finalize (GObject * o)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (o);
|
|
|
|
gst_object_unref (mix->collect);
|
|
g_mutex_clear (&mix->lock);
|
|
g_mutex_clear (&mix->setcaps_lock);
|
|
|
|
G_OBJECT_CLASS (parent_class)->finalize (o);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_dispose (GObject * o)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (o);
|
|
GSList *tmp;
|
|
|
|
for (tmp = mix->sinkpads; tmp; tmp = tmp->next) {
|
|
GstVideoMixer2Pad *mixpad = tmp->data;
|
|
|
|
if (mixpad->convert)
|
|
gst_video_converter_free (mixpad->convert);
|
|
mixpad->convert = NULL;
|
|
}
|
|
|
|
if (mix->pending_tags) {
|
|
gst_tag_list_unref (mix->pending_tags);
|
|
mix->pending_tags = NULL;
|
|
}
|
|
|
|
gst_caps_replace (&mix->current_caps, NULL);
|
|
|
|
G_OBJECT_CLASS (parent_class)->dispose (o);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_get_property (GObject * object,
|
|
guint prop_id, GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_BACKGROUND:
|
|
g_value_set_enum (value, mix->background);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_set_property (GObject * object,
|
|
guint prop_id, const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_BACKGROUND:
|
|
mix->background = g_value_get_enum (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* GstChildProxy implementation */
|
|
static GObject *
|
|
gst_videomixer2_child_proxy_get_child_by_index (GstChildProxy * child_proxy,
|
|
guint index)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (child_proxy);
|
|
GObject *obj;
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
if ((obj = g_slist_nth_data (mix->sinkpads, index)))
|
|
g_object_ref (obj);
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
return obj;
|
|
}
|
|
|
|
static guint
|
|
gst_videomixer2_child_proxy_get_children_count (GstChildProxy * child_proxy)
|
|
{
|
|
guint count = 0;
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (child_proxy);
|
|
|
|
GST_VIDEO_MIXER2_LOCK (mix);
|
|
count = mix->numpads;
|
|
GST_VIDEO_MIXER2_UNLOCK (mix);
|
|
GST_INFO_OBJECT (mix, "Children Count: %d", count);
|
|
return count;
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_child_proxy_init (gpointer g_iface, gpointer iface_data)
|
|
{
|
|
GstChildProxyInterface *iface = g_iface;
|
|
|
|
GST_INFO ("initializing child proxy interface");
|
|
iface->get_child_by_index = gst_videomixer2_child_proxy_get_child_by_index;
|
|
iface->get_children_count = gst_videomixer2_child_proxy_get_children_count;
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_constructed (GObject * obj)
|
|
{
|
|
GstVideoMixer2 *mix = GST_VIDEO_MIXER2 (obj);
|
|
gchar *cp_name;
|
|
|
|
cp_name = g_strconcat (GST_OBJECT_NAME (obj), "-collectpads", NULL);
|
|
gst_object_set_name (GST_OBJECT (mix->collect), cp_name);
|
|
g_free (cp_name);
|
|
|
|
G_OBJECT_CLASS (gst_videomixer2_parent_class)->constructed (obj);
|
|
}
|
|
|
|
/* GObject boilerplate */
|
|
static void
|
|
gst_videomixer2_class_init (GstVideoMixer2Class * klass)
|
|
{
|
|
GObjectClass *gobject_class = (GObjectClass *) klass;
|
|
GstElementClass *gstelement_class = (GstElementClass *) klass;
|
|
|
|
gobject_class->constructed = gst_videomixer2_constructed;
|
|
gobject_class->finalize = gst_videomixer2_finalize;
|
|
gobject_class->dispose = gst_videomixer2_dispose;
|
|
|
|
gobject_class->get_property = gst_videomixer2_get_property;
|
|
gobject_class->set_property = gst_videomixer2_set_property;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_BACKGROUND,
|
|
g_param_spec_enum ("background", "Background", "Background type",
|
|
GST_TYPE_VIDEO_MIXER2_BACKGROUND,
|
|
DEFAULT_BACKGROUND, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
|
|
gstelement_class->request_new_pad =
|
|
GST_DEBUG_FUNCPTR (gst_videomixer2_request_new_pad);
|
|
gstelement_class->release_pad =
|
|
GST_DEBUG_FUNCPTR (gst_videomixer2_release_pad);
|
|
gstelement_class->change_state =
|
|
GST_DEBUG_FUNCPTR (gst_videomixer2_change_state);
|
|
|
|
gst_element_class_add_static_pad_template (gstelement_class, &src_factory);
|
|
gst_element_class_add_static_pad_template (gstelement_class, &sink_factory);
|
|
|
|
gst_element_class_set_static_metadata (gstelement_class, "Video mixer 2",
|
|
"Filter/Editor/Video/Compositor",
|
|
"Deprecated by compositor. Mix multiple video streams",
|
|
"Wim Taymans <wim@fluendo.com>, "
|
|
"Sebastian Dröge <sebastian.droege@collabora.co.uk>");
|
|
|
|
/* Register the pad class */
|
|
g_type_class_ref (GST_TYPE_VIDEO_MIXER2_PAD);
|
|
|
|
gst_type_mark_as_plugin_api (GST_TYPE_VIDEO_MIXER2_BACKGROUND, 0);
|
|
}
|
|
|
|
static void
|
|
gst_videomixer2_init (GstVideoMixer2 * mix)
|
|
{
|
|
GstElementClass *klass = GST_ELEMENT_GET_CLASS (mix);
|
|
|
|
mix->srcpad =
|
|
gst_pad_new_from_template (gst_element_class_get_pad_template (klass,
|
|
"src"), "src");
|
|
gst_pad_set_query_function (GST_PAD (mix->srcpad),
|
|
GST_DEBUG_FUNCPTR (gst_videomixer2_src_query));
|
|
gst_pad_set_event_function (GST_PAD (mix->srcpad),
|
|
GST_DEBUG_FUNCPTR (gst_videomixer2_src_event));
|
|
gst_element_add_pad (GST_ELEMENT (mix), mix->srcpad);
|
|
|
|
mix->collect = gst_collect_pads_new ();
|
|
gst_collect_pads_set_flush_function (mix->collect,
|
|
(GstCollectPadsFlushFunction) gst_videomixer2_flush, mix);
|
|
mix->background = DEFAULT_BACKGROUND;
|
|
mix->current_caps = NULL;
|
|
mix->pending_tags = NULL;
|
|
|
|
gst_collect_pads_set_function (mix->collect,
|
|
(GstCollectPadsFunction) GST_DEBUG_FUNCPTR (gst_videomixer2_collected),
|
|
mix);
|
|
gst_collect_pads_set_event_function (mix->collect,
|
|
(GstCollectPadsEventFunction) gst_videomixer2_sink_event, mix);
|
|
gst_collect_pads_set_query_function (mix->collect,
|
|
(GstCollectPadsQueryFunction) gst_videomixer2_sink_query, mix);
|
|
gst_collect_pads_set_clip_function (mix->collect,
|
|
(GstCollectPadsClipFunction) gst_videomixer2_sink_clip, mix);
|
|
|
|
g_mutex_init (&mix->lock);
|
|
g_mutex_init (&mix->setcaps_lock);
|
|
/* initialize variables */
|
|
gst_videomixer2_reset (mix);
|
|
}
|
|
|
|
/* Element registration */
|
|
static gboolean
|
|
plugin_init (GstPlugin * plugin)
|
|
{
|
|
GST_DEBUG_CATEGORY_INIT (gst_videomixer2_debug, "videomixer", 0,
|
|
"video mixer");
|
|
|
|
gst_video_mixer_init_blend ();
|
|
|
|
return gst_element_register (plugin, "videomixer", GST_RANK_PRIMARY,
|
|
GST_TYPE_VIDEO_MIXER2);
|
|
}
|
|
|
|
GST_PLUGIN_DEFINE (GST_VERSION_MAJOR,
|
|
GST_VERSION_MINOR,
|
|
videomixer,
|
|
"Video mixer", plugin_init, VERSION, GST_LICENSE, GST_PACKAGE_NAME,
|
|
GST_PACKAGE_ORIGIN)
|