/* GStreamer * Copyright (C) <1999> Erik Walthinsen * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:element-videorate * @title: videorate * * This element takes an incoming stream of timestamped video frames. * It will produce a perfect stream that matches the source pad's framerate. * * The correction is performed by dropping and duplicating frames, no fancy * algorithm is used to interpolate frames (yet). * * By default the element will simply negotiate the same framerate on its * source and sink pad. * * This operation is useful to link to elements that require a perfect stream. * Typical examples are formats that do not store timestamps for video frames, * but only store a framerate, like Ogg and AVI. * * A conversion to a specific framerate can be forced by using filtered caps on * the source pad. * * The properties #GstVideoRate:in, #GstVideoRate:out, #GstVideoRate:duplicate * and #GstVideoRate:drop can be read to obtain information about number of * input frames, output frames, dropped frames (i.e. the number of unused input * frames) and duplicated frames (i.e. the number of times an input frame was * duplicated, beside being used normally). * * An input stream that needs no adjustments will thus never have dropped or * duplicated frames. * * When the #GstVideoRate:silent property is set to FALSE, a GObject property * notification will be emitted whenever one of the #GstVideoRate:duplicate or * #GstVideoRate:drop values changes. * This can potentially cause performance degradation. * Note that property notification will happen from the streaming thread, so * applications should be prepared for this. * * The property #GstVideoRate:rate allows the modification of video speed by a * certain factor. It must not be confused with framerate. Think of rate as * speed and framerate as flow. * * ## Example pipelines * |[ * gst-launch-1.0 -v uridecodebin uri=file:///path/to/video.ogg ! videoconvert ! videoscale ! videorate ! video/x-raw,framerate=15/1 ! autovideosink * ]| * Decode a video file and adjust the framerate to 15 fps before playing. * To create a test Ogg/Theora file refer to the documentation of theoraenc. * |[ * gst-launch-1.0 -v v4l2src ! videorate ! video/x-raw,framerate=25/2 ! theoraenc ! oggmux ! filesink location=recording.ogg * ]| * Capture video from a V4L device, and adjust the stream to 12.5 fps before * encoding to Ogg/Theora. * |[ * gst-launch-1.0 -v uridecodebin uri=file:///path/to/video.ogg ! videoconvert ! videoscale ! videorate ! video/x-raw,framerate=1/5 ! jpegenc ! multifilesink location=snapshot-%05d.jpg * ]| * Decode a video file and save a snapshot every 5 seconds as consecutively numbered jpeg file. * */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstvideorate.h" #include GST_DEBUG_CATEGORY_STATIC (video_rate_debug); #define GST_CAT_DEFAULT video_rate_debug /* GstVideoRate signals and args */ enum { /* FILL ME */ LAST_SIGNAL }; #define DEFAULT_SILENT TRUE #define DEFAULT_NEW_PREF 1.0 #define DEFAULT_SKIP_TO_FIRST FALSE #define DEFAULT_DROP_ONLY FALSE #define DEFAULT_AVERAGE_PERIOD 0 #define DEFAULT_MAX_RATE G_MAXINT #define DEFAULT_RATE 1.0 #define DEFAULT_MAX_DUPLICATION_TIME 0 enum { PROP_0, PROP_IN, PROP_OUT, PROP_DUP, PROP_DROP, PROP_SILENT, PROP_NEW_PREF, PROP_SKIP_TO_FIRST, PROP_DROP_ONLY, PROP_AVERAGE_PERIOD, PROP_MAX_RATE, PROP_RATE, PROP_MAX_DUPLICATION_TIME }; static GstStaticPadTemplate gst_video_rate_src_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS ("video/x-raw(ANY);" "video/x-bayer(ANY);" "image/jpeg(ANY);" "image/png(ANY)") ); static GstStaticPadTemplate gst_video_rate_sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS ("video/x-raw(ANY);" "video/x-bayer(ANY);" "image/jpeg(ANY);" "image/png(ANY)") ); static void gst_video_rate_swap_prev (GstVideoRate * videorate, GstBuffer * buffer, gint64 time); static gboolean gst_video_rate_sink_event (GstBaseTransform * trans, GstEvent * event); static gboolean gst_video_rate_src_event (GstBaseTransform * trans, GstEvent * event); static gboolean gst_video_rate_query (GstBaseTransform * trans, GstPadDirection direction, GstQuery * query); static gboolean gst_video_rate_setcaps (GstBaseTransform * trans, GstCaps * in_caps, GstCaps * out_caps); static GstCaps *gst_video_rate_transform_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * caps, GstCaps * filter); static GstCaps *gst_video_rate_fixate_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * caps, GstCaps * othercaps); static GstFlowReturn gst_video_rate_transform_ip (GstBaseTransform * trans, GstBuffer * buf); static gboolean gst_video_rate_propose_allocation (GstBaseTransform * trans, GstQuery * decide_query, GstQuery * query); static gboolean gst_video_rate_start (GstBaseTransform * trans); static gboolean gst_video_rate_stop (GstBaseTransform * trans); static void gst_video_rate_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_video_rate_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static GParamSpec *pspec_drop = NULL; static GParamSpec *pspec_duplicate = NULL; #define gst_video_rate_parent_class parent_class G_DEFINE_TYPE (GstVideoRate, gst_video_rate, GST_TYPE_BASE_TRANSFORM); static void gst_video_rate_class_init (GstVideoRateClass * klass) { GObjectClass *object_class = G_OBJECT_CLASS (klass); GstElementClass *element_class = GST_ELEMENT_CLASS (klass); GstBaseTransformClass *base_class = GST_BASE_TRANSFORM_CLASS (klass); object_class->set_property = gst_video_rate_set_property; object_class->get_property = gst_video_rate_get_property; base_class->set_caps = GST_DEBUG_FUNCPTR (gst_video_rate_setcaps); base_class->transform_caps = GST_DEBUG_FUNCPTR (gst_video_rate_transform_caps); base_class->transform_ip = GST_DEBUG_FUNCPTR (gst_video_rate_transform_ip); base_class->sink_event = GST_DEBUG_FUNCPTR (gst_video_rate_sink_event); base_class->src_event = GST_DEBUG_FUNCPTR (gst_video_rate_src_event); base_class->start = GST_DEBUG_FUNCPTR (gst_video_rate_start); base_class->stop = GST_DEBUG_FUNCPTR (gst_video_rate_stop); base_class->fixate_caps = GST_DEBUG_FUNCPTR (gst_video_rate_fixate_caps); base_class->query = GST_DEBUG_FUNCPTR (gst_video_rate_query); base_class->propose_allocation = GST_DEBUG_FUNCPTR (gst_video_rate_propose_allocation); g_object_class_install_property (object_class, PROP_IN, g_param_spec_uint64 ("in", "In", "Number of input frames", 0, G_MAXUINT64, 0, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (object_class, PROP_OUT, g_param_spec_uint64 ("out", "Out", "Number of output frames", 0, G_MAXUINT64, 0, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS)); pspec_duplicate = g_param_spec_uint64 ("duplicate", "Duplicate", "Number of duplicated frames", 0, G_MAXUINT64, 0, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_DUP, pspec_duplicate); pspec_drop = g_param_spec_uint64 ("drop", "Drop", "Number of dropped frames", 0, G_MAXUINT64, 0, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_property (object_class, PROP_DROP, pspec_drop); g_object_class_install_property (object_class, PROP_SILENT, g_param_spec_boolean ("silent", "silent", "Don't emit notify for dropped and duplicated frames", DEFAULT_SILENT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (object_class, PROP_NEW_PREF, g_param_spec_double ("new-pref", "New Pref", "Value indicating how much to prefer new frames (unused)", 0.0, 1.0, DEFAULT_NEW_PREF, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVideoRate:skip-to-first: * * Don't produce buffers before the first one we receive. */ g_object_class_install_property (object_class, PROP_SKIP_TO_FIRST, g_param_spec_boolean ("skip-to-first", "Skip to first buffer", "Don't produce buffers before the first one we receive", DEFAULT_SKIP_TO_FIRST, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVideoRate:drop-only: * * Only drop frames, no duplicates are produced. */ g_object_class_install_property (object_class, PROP_DROP_ONLY, g_param_spec_boolean ("drop-only", "Only Drop", "Only drop frames, no duplicates are produced", DEFAULT_DROP_ONLY, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVideoRate:average-period: * * Arrange for maximum framerate by dropping frames beyond a certain framerate, * where the framerate is calculated using a moving average over the * configured. */ g_object_class_install_property (object_class, PROP_AVERAGE_PERIOD, g_param_spec_uint64 ("average-period", "Period over which to average", "Period over which to average the framerate (in ns) (0 = disabled)", 0, G_MAXINT64, DEFAULT_AVERAGE_PERIOD, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVideoRate:max-rate: * * maximum framerate to pass through */ g_object_class_install_property (object_class, PROP_MAX_RATE, g_param_spec_int ("max-rate", "maximum framerate", "Maximum framerate allowed to pass through " "(in frames per second, implies drop-only)", 1, G_MAXINT, DEFAULT_MAX_RATE, G_PARAM_READWRITE | G_PARAM_CONSTRUCT | G_PARAM_STATIC_STRINGS)); /** * GstVideoRate:rate: * * Factor of speed for frame displaying * * Since: 1.12 */ g_object_class_install_property (object_class, PROP_RATE, g_param_spec_double ("rate", "Rate", "Factor of speed for frame displaying", 0.0, G_MAXDOUBLE, DEFAULT_RATE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_MUTABLE_READY)); /** * GstVideoRate:max-duplication-time: * * Duplicate frames only if the gap between two consecutive frames does not * exceed this duration. * * Since: 1.16 */ g_object_class_install_property (object_class, PROP_MAX_DUPLICATION_TIME, g_param_spec_uint64 ("max-duplication-time", "Maximum time to duplicate a frame", "Do not duplicate frames if the gap exceeds this period " "(in ns) (0 = disabled)", 0, G_MAXUINT64, DEFAULT_MAX_DUPLICATION_TIME, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); gst_element_class_set_static_metadata (element_class, "Video rate adjuster", "Filter/Effect/Video", "Drops/duplicates/adjusts timestamps on video frames to make a perfect stream", "Wim Taymans "); gst_element_class_add_static_pad_template (element_class, &gst_video_rate_sink_template); gst_element_class_add_static_pad_template (element_class, &gst_video_rate_src_template); } static void gst_value_fraction_get_extremes (const GValue * v, gint * min_num, gint * min_denom, gint * max_num, gint * max_denom) { if (GST_VALUE_HOLDS_FRACTION (v)) { *min_num = *max_num = gst_value_get_fraction_numerator (v); *min_denom = *max_denom = gst_value_get_fraction_denominator (v); } else if (GST_VALUE_HOLDS_FRACTION_RANGE (v)) { const GValue *min, *max; min = gst_value_get_fraction_range_min (v); *min_num = gst_value_get_fraction_numerator (min); *min_denom = gst_value_get_fraction_denominator (min); max = gst_value_get_fraction_range_max (v); *max_num = gst_value_get_fraction_numerator (max); *max_denom = gst_value_get_fraction_denominator (max); } else if (GST_VALUE_HOLDS_LIST (v)) { gint min_n = G_MAXINT, min_d = 1, max_n = 0, max_d = 1; int i, n; *min_num = G_MAXINT; *min_denom = 1; *max_num = 0; *max_denom = 1; n = gst_value_list_get_size (v); g_assert (n > 0); for (i = 0; i < n; i++) { const GValue *t = gst_value_list_get_value (v, i); gst_value_fraction_get_extremes (t, &min_n, &min_d, &max_n, &max_d); if (gst_util_fraction_compare (min_n, min_d, *min_num, *min_denom) < 0) { *min_num = min_n; *min_denom = min_d; } if (gst_util_fraction_compare (max_n, max_d, *max_num, *max_denom) > 0) { *max_num = max_n; *max_denom = max_d; } } } else { g_warning ("Unknown type for framerate"); *min_num = 0; *min_denom = 1; *max_num = G_MAXINT; *max_denom = 1; } } /* Clamp the framerate in a caps structure to be a smaller range then * [1...max_rate], otherwise return false */ static gboolean gst_video_max_rate_clamp_structure (GstStructure * s, gint maxrate, gint * min_num, gint * min_denom, gint * max_num, gint * max_denom) { gboolean ret = FALSE; if (!gst_structure_has_field (s, "framerate")) { /* No framerate field implies any framerate, clamping would result in * [1..max_rate] so not a real subset */ goto out; } else { const GValue *v; GValue intersection = { 0, }; GValue clamp = { 0, }; gint tmp_num, tmp_denom; g_value_init (&clamp, GST_TYPE_FRACTION_RANGE); gst_value_set_fraction_range_full (&clamp, 0, 1, maxrate, 1); v = gst_structure_get_value (s, "framerate"); ret = gst_value_intersect (&intersection, v, &clamp); g_value_unset (&clamp); if (!ret) goto out; gst_value_fraction_get_extremes (&intersection, min_num, min_denom, max_num, max_denom); gst_value_fraction_get_extremes (v, &tmp_num, &tmp_denom, max_num, max_denom); if (gst_util_fraction_compare (*max_num, *max_denom, maxrate, 1) > 0) { *max_num = maxrate; *max_denom = 1; } gst_structure_take_value (s, "framerate", &intersection); } out: return ret; } static GstCaps * gst_video_rate_transform_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * caps, GstCaps * filter) { GstVideoRate *videorate = GST_VIDEO_RATE (trans); GstCaps *ret; GstStructure *s, *s1, *s2, *s3 = NULL; int maxrate = g_atomic_int_get (&videorate->max_rate); gint i; ret = gst_caps_new_empty (); for (i = 0; i < gst_caps_get_size (caps); i++) { s = gst_caps_get_structure (caps, i); s1 = gst_structure_copy (s); if (videorate->updating_caps && direction == GST_PAD_SINK) { GST_INFO_OBJECT (trans, "Only updating caps %" GST_PTR_FORMAT " with framerate" " %d/%d", caps, videorate->to_rate_numerator, videorate->to_rate_denominator); gst_structure_set (s1, "framerate", GST_TYPE_FRACTION, videorate->to_rate_numerator, videorate->to_rate_denominator, NULL); ret = gst_caps_merge_structure (ret, s1); continue; } s2 = gst_structure_copy (s); s3 = NULL; if (videorate->drop_only) { gint min_num = 0, min_denom = 1; gint max_num = G_MAXINT, max_denom = 1; /* Clamp the caps to our maximum rate as the first caps if possible */ if (!gst_video_max_rate_clamp_structure (s1, maxrate, &min_num, &min_denom, &max_num, &max_denom)) { min_num = 0; min_denom = 1; max_num = maxrate; max_denom = 1; /* clamp wouldn't be a real subset of 1..maxrate, in this case the sink * caps should become [1..maxrate], [1..maxint] and the src caps just * [1..maxrate]. In case there was a caps incompatibility things will * explode later as appropriate :) * * In case [X..maxrate] == [X..maxint], skip as we'll set it later */ if (direction == GST_PAD_SRC && maxrate != G_MAXINT) gst_structure_set (s1, "framerate", GST_TYPE_FRACTION_RANGE, min_num, min_denom, maxrate, 1, NULL); else { gst_structure_free (s1); s1 = NULL; } } if (direction == GST_PAD_SRC) { /* We can accept anything as long as it's at least the minimal framerate * the the sink needs */ gst_structure_set (s2, "framerate", GST_TYPE_FRACTION_RANGE, min_num, min_denom, G_MAXINT, 1, NULL); /* Also allow unknown framerate, if it isn't already */ if (min_num != 0 || min_denom != 1) { s3 = gst_structure_copy (s); gst_structure_set (s3, "framerate", GST_TYPE_FRACTION, 0, 1, NULL); } } else if (max_num != 0 || max_denom != 1) { /* We can provide everything up to the maximum framerate at the src */ gst_structure_set (s2, "framerate", GST_TYPE_FRACTION_RANGE, 0, 1, max_num, max_denom, NULL); } } else if (direction == GST_PAD_SINK) { gint min_num = 0, min_denom = 1; gint max_num = G_MAXINT, max_denom = 1; if (!gst_video_max_rate_clamp_structure (s1, maxrate, &min_num, &min_denom, &max_num, &max_denom)) { gst_structure_free (s1); s1 = NULL; } gst_structure_set (s2, "framerate", GST_TYPE_FRACTION_RANGE, 0, 1, maxrate, 1, NULL); } else { /* set the framerate as a range */ gst_structure_set (s2, "framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL); } if (s1 != NULL) ret = gst_caps_merge_structure_full (ret, s1, gst_caps_features_copy (gst_caps_get_features (caps, i))); ret = gst_caps_merge_structure_full (ret, s2, gst_caps_features_copy (gst_caps_get_features (caps, i))); if (s3 != NULL) ret = gst_caps_merge_structure_full (ret, s3, gst_caps_features_copy (gst_caps_get_features (caps, i))); } if (filter) { GstCaps *intersection; intersection = gst_caps_intersect_full (filter, ret, GST_CAPS_INTERSECT_FIRST); gst_caps_unref (ret); ret = intersection; } return ret; } static GstCaps * gst_video_rate_fixate_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * caps, GstCaps * othercaps) { GstStructure *s; gint num, denom; const GValue *par; s = gst_caps_get_structure (caps, 0); if (G_UNLIKELY (!gst_structure_get_fraction (s, "framerate", &num, &denom))) return othercaps; othercaps = gst_caps_truncate (othercaps); othercaps = gst_caps_make_writable (othercaps); s = gst_caps_get_structure (othercaps, 0); gst_structure_fixate_field_nearest_fraction (s, "framerate", num, denom); if ((par = gst_structure_get_value (s, "pixel-aspect-ratio"))) gst_structure_fixate_field_nearest_fraction (s, "pixel-aspect-ratio", 1, 1); return othercaps; } static gboolean gst_video_rate_setcaps (GstBaseTransform * trans, GstCaps * in_caps, GstCaps * out_caps) { GstVideoRate *videorate = GST_VIDEO_RATE (trans); GstStructure *structure; gboolean ret = TRUE; gint rate_numerator, rate_denominator; GST_DEBUG_OBJECT (trans, "setcaps called in: %" GST_PTR_FORMAT " out: %" GST_PTR_FORMAT, in_caps, out_caps); structure = gst_caps_get_structure (in_caps, 0); if (!gst_structure_get_fraction (structure, "framerate", &rate_numerator, &rate_denominator)) goto no_framerate; videorate->from_rate_numerator = rate_numerator; videorate->from_rate_denominator = rate_denominator; structure = gst_caps_get_structure (out_caps, 0); if (!gst_structure_get_fraction (structure, "framerate", &rate_numerator, &rate_denominator)) goto no_framerate; /* out_frame_count is scaled by the frame rate caps when calculating next_ts. * when the frame rate caps change, we must update base_ts and reset * out_frame_count */ if (videorate->to_rate_numerator) { videorate->base_ts += gst_util_uint64_scale (videorate->out_frame_count + (videorate->segment.rate < 0.0 ? 1 : 0), videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); } videorate->out_frame_count = 0; videorate->to_rate_numerator = rate_numerator; videorate->to_rate_denominator = rate_denominator; if (rate_numerator) videorate->wanted_diff = gst_util_uint64_scale_int (GST_SECOND, rate_denominator, rate_numerator); else videorate->wanted_diff = 0; done: /* After a setcaps, our caps may have changed. In that case, we can't use * the old buffer, if there was one (it might have different dimensions) */ GST_DEBUG_OBJECT (videorate, "swapping old buffers"); gst_video_rate_swap_prev (videorate, NULL, GST_CLOCK_TIME_NONE); videorate->last_ts = GST_CLOCK_TIME_NONE; videorate->average = 0; return ret; no_framerate: { GST_DEBUG_OBJECT (videorate, "no framerate specified"); ret = FALSE; goto done; } } static void gst_video_rate_reset (GstVideoRate * videorate) { GST_DEBUG_OBJECT (videorate, "resetting internal variables"); videorate->in = 0; videorate->out = 0; videorate->base_ts = 0; videorate->out_frame_count = 0; videorate->drop = 0; videorate->dup = 0; videorate->next_ts = GST_CLOCK_TIME_NONE; videorate->last_ts = GST_CLOCK_TIME_NONE; videorate->discont = TRUE; videorate->average = 0; videorate->force_variable_rate = FALSE; gst_video_rate_swap_prev (videorate, NULL, 0); gst_segment_init (&videorate->segment, GST_FORMAT_TIME); } static void gst_video_rate_init (GstVideoRate * videorate) { gst_video_rate_reset (videorate); videorate->silent = DEFAULT_SILENT; videorate->new_pref = DEFAULT_NEW_PREF; videorate->drop_only = DEFAULT_DROP_ONLY; videorate->average_period = DEFAULT_AVERAGE_PERIOD; videorate->average_period_set = DEFAULT_AVERAGE_PERIOD; videorate->max_rate = DEFAULT_MAX_RATE; videorate->rate = DEFAULT_RATE; videorate->pending_rate = DEFAULT_RATE; videorate->max_duplication_time = DEFAULT_MAX_DUPLICATION_TIME; videorate->from_rate_numerator = 0; videorate->from_rate_denominator = 0; videorate->to_rate_numerator = 0; videorate->to_rate_denominator = 0; gst_base_transform_set_gap_aware (GST_BASE_TRANSFORM (videorate), TRUE); } /* @outbuf: (transfer full) needs to be writable */ static GstFlowReturn gst_video_rate_push_buffer (GstVideoRate * videorate, GstBuffer * outbuf, gboolean duplicate, GstClockTime next_intime) { GstFlowReturn res; GstClockTime push_ts; GST_BUFFER_OFFSET (outbuf) = videorate->out; GST_BUFFER_OFFSET_END (outbuf) = videorate->out + 1; if (videorate->discont) { GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_DISCONT); videorate->discont = FALSE; } else GST_BUFFER_FLAG_UNSET (outbuf, GST_BUFFER_FLAG_DISCONT); if (duplicate) GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_GAP); else GST_BUFFER_FLAG_UNSET (outbuf, GST_BUFFER_FLAG_GAP); /* this is the timestamp we put on the buffer */ push_ts = videorate->next_ts; videorate->out++; videorate->out_frame_count++; if (videorate->segment.rate < 0.0) { if (videorate->to_rate_numerator) { /* interpolate next expected timestamp in the segment */ GstClockTimeDiff next_ts = videorate->segment.base + videorate->segment.stop - videorate->base_ts - gst_util_uint64_scale (videorate->out_frame_count + 1, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); videorate->next_ts = next_ts < 0 ? GST_CLOCK_TIME_NONE : next_ts; GST_BUFFER_DURATION (outbuf) = gst_util_uint64_scale (videorate->out_frame_count, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator) - gst_util_uint64_scale (videorate->out_frame_count - 1, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); } else if (next_intime != GST_CLOCK_TIME_NONE) { videorate->next_ts = next_intime; } else { GST_FIXME_OBJECT (videorate, "No next intime for reverse playback"); } } else { if (videorate->to_rate_numerator) { /* interpolate next expected timestamp in the segment */ videorate->next_ts = videorate->segment.base + videorate->segment.start + videorate->base_ts + gst_util_uint64_scale (videorate->out_frame_count, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); GST_BUFFER_DURATION (outbuf) = videorate->next_ts - push_ts; } else if (GST_CLOCK_TIME_IS_VALID (GST_BUFFER_DURATION (outbuf))) { videorate->next_ts = GST_BUFFER_PTS (outbuf) + GST_BUFFER_DURATION (outbuf); } else { /* There must always be a valid duration on prevbuf if rate > 0, * it is ensured in the transform_ip function */ GST_FIXME_OBJECT (videorate, "No buffer duration known"); } } /* We do not need to update time in VFR (variable frame rate) mode */ if (!videorate->drop_only) { /* adapt for looping, bring back to time in current segment. */ GST_BUFFER_TIMESTAMP (outbuf) = push_ts - videorate->segment.base; } GST_LOG_OBJECT (videorate, "old is best, dup, pushing buffer outgoing ts %" GST_TIME_FORMAT, GST_TIME_ARGS (push_ts)); res = gst_pad_push (GST_BASE_TRANSFORM_SRC_PAD (videorate), outbuf); return res; } /* flush the oldest buffer */ static GstFlowReturn gst_video_rate_flush_prev (GstVideoRate * videorate, gboolean duplicate, GstClockTime next_intime) { GstBuffer *outbuf; if (!videorate->prevbuf) goto eos_before_buffers; outbuf = gst_buffer_ref (videorate->prevbuf); /* make sure we can write to the metadata */ outbuf = gst_buffer_make_writable (outbuf); return gst_video_rate_push_buffer (videorate, outbuf, duplicate, next_intime); /* WARNINGS */ eos_before_buffers: { GST_INFO_OBJECT (videorate, "got EOS before any buffer was received"); return GST_FLOW_OK; } } static void gst_video_rate_swap_prev (GstVideoRate * videorate, GstBuffer * buffer, gint64 time) { GST_LOG_OBJECT (videorate, "swap_prev: storing buffer %p in prev", buffer); if (videorate->prevbuf) gst_buffer_unref (videorate->prevbuf); videorate->prevbuf = buffer != NULL ? gst_buffer_ref (buffer) : NULL; videorate->prev_ts = time; } static void gst_video_rate_notify_drop (GstVideoRate * videorate) { g_object_notify_by_pspec ((GObject *) videorate, pspec_drop); } static void gst_video_rate_notify_duplicate (GstVideoRate * videorate) { g_object_notify_by_pspec ((GObject *) videorate, pspec_duplicate); } #define MAGIC_LIMIT 25 static gboolean gst_video_rate_sink_event (GstBaseTransform * trans, GstEvent * event) { GstVideoRate *videorate; videorate = GST_VIDEO_RATE (trans); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEGMENT: { GstSegment segment; gint seqnum; gst_event_copy_segment (event, &segment); if (segment.format != GST_FORMAT_TIME) goto format_error; GST_DEBUG_OBJECT (videorate, "handle NEWSEGMENT"); /* close up the previous segment, if appropriate */ if (videorate->prevbuf) { gint count = 0; GstFlowReturn res; res = GST_FLOW_OK; /* fill up to the end of current segment, * or only send out the stored buffer if there is no specific stop. * regardless, prevent going loopy in strange cases */ while (res == GST_FLOW_OK && count <= MAGIC_LIMIT && !videorate->drop_only && ((videorate->segment.rate > 0.0 && GST_CLOCK_TIME_IS_VALID (videorate->segment.stop) && GST_CLOCK_TIME_IS_VALID (videorate->next_ts) && videorate->next_ts - videorate->segment.base < videorate->segment.stop) || (videorate->segment.rate < 0.0 && GST_CLOCK_TIME_IS_VALID (videorate->segment.start) && GST_CLOCK_TIME_IS_VALID (videorate->next_ts) && videorate->next_ts - videorate->segment.base >= videorate->segment.start) || count < 1)) { res = gst_video_rate_flush_prev (videorate, count > 0, GST_CLOCK_TIME_NONE); count++; } if (count > 1) { videorate->dup += count - 1; if (!videorate->silent) gst_video_rate_notify_duplicate (videorate); } /* clean up for the new one; _chain will resume from the new start */ gst_video_rate_swap_prev (videorate, NULL, 0); } videorate->base_ts = 0; videorate->out_frame_count = 0; videorate->next_ts = GST_CLOCK_TIME_NONE; /* We just want to update the accumulated stream_time */ segment.start = (gint64) (segment.start / videorate->rate); segment.position = (gint64) (segment.position / videorate->rate); if (GST_CLOCK_TIME_IS_VALID (segment.stop)) segment.stop = (gint64) (segment.stop / videorate->rate); segment.time = (gint64) (segment.time / videorate->rate); gst_segment_copy_into (&segment, &videorate->segment); GST_DEBUG_OBJECT (videorate, "updated segment: %" GST_SEGMENT_FORMAT, &videorate->segment); seqnum = gst_event_get_seqnum (event); gst_event_unref (event); event = gst_event_new_segment (&segment); gst_event_set_seqnum (event, seqnum); break; } case GST_EVENT_SEGMENT_DONE: case GST_EVENT_EOS:{ gint count = 0; GstFlowReturn res = GST_FLOW_OK; GST_DEBUG_OBJECT (videorate, "Got %s", gst_event_type_get_name (GST_EVENT_TYPE (event))); /* If the segment has a stop position, fill the segment */ if (GST_CLOCK_TIME_IS_VALID (videorate->segment.stop)) { /* fill up to the end of current segment, * or only send out the stored buffer if there is no specific stop. * regardless, prevent going loopy in strange cases */ while (res == GST_FLOW_OK && count <= MAGIC_LIMIT && !videorate->drop_only && ((videorate->segment.rate > 0.0 && GST_CLOCK_TIME_IS_VALID (videorate->segment.stop) && GST_CLOCK_TIME_IS_VALID (videorate->next_ts) && videorate->next_ts - videorate->segment.base < videorate->segment.stop) || (videorate->segment.rate < 0.0 && GST_CLOCK_TIME_IS_VALID (videorate->segment.start) && GST_CLOCK_TIME_IS_VALID (videorate->next_ts) && videorate->next_ts - videorate->segment.base >= videorate->segment.start) )) { res = gst_video_rate_flush_prev (videorate, count > 0, GST_CLOCK_TIME_NONE); count++; } } else if (!videorate->drop_only && videorate->prevbuf) { /* Output at least one frame but if the buffer duration is valid, output * enough frames to use the complete buffer duration */ if (GST_BUFFER_DURATION_IS_VALID (videorate->prevbuf)) { GstClockTime end_ts = videorate->next_ts + GST_BUFFER_DURATION (videorate->prevbuf); while (res == GST_FLOW_OK && count <= MAGIC_LIMIT && ((videorate->segment.rate > 0.0 && GST_CLOCK_TIME_IS_VALID (videorate->segment.stop) && GST_CLOCK_TIME_IS_VALID (videorate->next_ts) && videorate->next_ts - videorate->segment.base < end_ts) || count < 1)) { res = gst_video_rate_flush_prev (videorate, count > 0, GST_CLOCK_TIME_NONE); count++; } } else { res = gst_video_rate_flush_prev (videorate, FALSE, GST_CLOCK_TIME_NONE); count = 1; } } if (count > 1) { videorate->dup += count - 1; if (!videorate->silent) gst_video_rate_notify_duplicate (videorate); } else if (count == 0 && !GST_CLOCK_TIME_IS_VALID (videorate->segment.stop)) { videorate->drop++; if (!videorate->silent) gst_video_rate_notify_drop (videorate); } break; } case GST_EVENT_FLUSH_STOP: /* also resets the segment */ GST_DEBUG_OBJECT (videorate, "Got FLUSH_STOP"); gst_video_rate_reset (videorate); break; case GST_EVENT_GAP: /* no gaps after videorate, ignore the event */ gst_event_unref (event); return TRUE; default: break; } return GST_BASE_TRANSFORM_CLASS (parent_class)->sink_event (trans, event); /* ERRORS */ format_error: { GST_WARNING_OBJECT (videorate, "Got segment but doesn't have GST_FORMAT_TIME value"); return FALSE; } } static gboolean gst_video_rate_src_event (GstBaseTransform * trans, GstEvent * event) { GstVideoRate *videorate; GstPad *sinkpad; gboolean res = FALSE; videorate = GST_VIDEO_RATE (trans); sinkpad = GST_BASE_TRANSFORM_SINK_PAD (trans); switch (GST_EVENT_TYPE (event)) { case GST_EVENT_SEEK: { gdouble srate; GstSeekFlags flags; GstSeekType start_type, stop_type; gint64 start, stop; gint seqnum = gst_event_get_seqnum (event); gst_event_parse_seek (event, &srate, NULL, &flags, &start_type, &start, &stop_type, &stop); start = (gint64) (start * videorate->rate); if (GST_CLOCK_TIME_IS_VALID (stop)) { stop = (gint64) (stop * videorate->rate); } gst_event_unref (event); event = gst_event_new_seek (srate, GST_FORMAT_TIME, flags, start_type, start, stop_type, stop); gst_event_set_seqnum (event, seqnum); res = gst_pad_push_event (sinkpad, event); break; } case GST_EVENT_QOS: { GstQOSType type; gdouble proportion; GstClockTimeDiff diff; GstClockTime timestamp; gst_event_parse_qos (event, &type, &proportion, &diff, ×tamp); if (GST_CLOCK_TIME_IS_VALID (timestamp) && videorate->rate != 1.0) { GST_OBJECT_LOCK (trans); GST_DEBUG_OBJECT (trans, "Rescaling QoS event taking our rate into" "account. Timestamp: %" GST_TIME_FORMAT " -> %" GST_TIME_FORMAT " - diff %" G_GINT64_FORMAT "-> %" G_GINT64_FORMAT, GST_TIME_ARGS (timestamp), GST_TIME_ARGS (videorate->base_ts + ((timestamp - videorate->base_ts) * videorate->rate)), diff, (GstClockTimeDiff) (diff * videorate->rate)); if (videorate->segment.rate < 0.0) timestamp = (videorate->segment.stop - videorate->base_ts) - ((videorate->segment.stop - videorate->base_ts - timestamp) * videorate->rate); else timestamp = videorate->base_ts + ((timestamp - videorate->base_ts) * videorate->rate); diff *= videorate->rate; GST_OBJECT_UNLOCK (trans); gst_event_unref (event); event = gst_event_new_qos (type, proportion, diff, timestamp); } /* Fallthrough */ } default: res = gst_pad_push_event (sinkpad, event); break; } return res; } static gboolean gst_video_rate_query (GstBaseTransform * trans, GstPadDirection direction, GstQuery * query) { GstVideoRate *videorate = GST_VIDEO_RATE (trans); gboolean res = FALSE; GstPad *otherpad; otherpad = (direction == GST_PAD_SRC) ? GST_BASE_TRANSFORM_SINK_PAD (trans) : GST_BASE_TRANSFORM_SRC_PAD (trans); switch (GST_QUERY_TYPE (query)) { case GST_QUERY_LATENCY: { GstClockTime min, max; gboolean live; guint64 latency; guint64 avg_period; gboolean drop_only; GstPad *peer; GST_OBJECT_LOCK (videorate); avg_period = videorate->average_period_set; drop_only = videorate->drop_only; GST_OBJECT_UNLOCK (videorate); if (avg_period == 0 && (peer = gst_pad_get_peer (otherpad))) { if ((res = gst_pad_query (peer, query))) { gst_query_parse_latency (query, &live, &min, &max); GST_DEBUG_OBJECT (videorate, "Peer latency: min %" GST_TIME_FORMAT " max %" GST_TIME_FORMAT, GST_TIME_ARGS (min), GST_TIME_ARGS (max)); /* Drop only has no latency, other modes have one frame latency */ if (!drop_only && videorate->from_rate_numerator != 0) { /* add latency. We don't really know since we hold on to the frames * until we get a next frame, which can be anything. We assume * however that this will take from_rate time. */ latency = gst_util_uint64_scale (GST_SECOND, videorate->from_rate_denominator, videorate->from_rate_numerator); } else { /* no input framerate, we don't know */ latency = 0; } GST_DEBUG_OBJECT (videorate, "Our latency: %" GST_TIME_FORMAT, GST_TIME_ARGS (latency)); min += latency; if (max != -1) max += latency; GST_DEBUG_OBJECT (videorate, "Calculated total latency : min %" GST_TIME_FORMAT " max %" GST_TIME_FORMAT, GST_TIME_ARGS (min), GST_TIME_ARGS (max)); gst_query_set_latency (query, live, min, max); } gst_object_unref (peer); break; } /* Simple fall back if we don't have a latency or a peer that we * can ask about its latency yet.. */ res = GST_BASE_TRANSFORM_CLASS (parent_class)->query (trans, direction, query); break; } case GST_QUERY_DURATION: { GstFormat format; gint64 duration; gdouble rate; res = GST_BASE_TRANSFORM_CLASS (parent_class)->query (trans, direction, query); if (!res) break; GST_OBJECT_LOCK (videorate); rate = videorate->pending_rate; GST_OBJECT_UNLOCK (videorate); if (rate == 1.0) break; gst_query_parse_duration (query, &format, &duration); if (format != GST_FORMAT_TIME) { GST_DEBUG_OBJECT (videorate, "not TIME format"); break; } GST_LOG_OBJECT (videorate, "upstream duration: %" G_GINT64_FORMAT, duration); /* Shouldn't this be a multiplication if the direction is downstream? */ if (GST_CLOCK_TIME_IS_VALID (duration)) { duration = (gint64) (duration / rate); } GST_LOG_OBJECT (videorate, "our duration: %" G_GINT64_FORMAT, duration); gst_query_set_duration (query, format, duration); break; } case GST_QUERY_POSITION: { GstFormat dst_format; gint64 dst_value; gdouble rate; GST_OBJECT_LOCK (videorate); rate = videorate->rate; GST_OBJECT_UNLOCK (videorate); gst_query_parse_position (query, &dst_format, NULL); if (dst_format != GST_FORMAT_TIME) { GST_DEBUG_OBJECT (videorate, "not TIME format"); break; } /* Shouldn't this be a multiplication if the direction is downstream? */ dst_value = (gint64) (gst_segment_to_stream_time (&videorate->segment, GST_FORMAT_TIME, videorate->last_ts / rate)); GST_LOG_OBJECT (videorate, "our position: %" GST_TIME_FORMAT, GST_TIME_ARGS (dst_value)); gst_query_set_position (query, dst_format, dst_value); res = TRUE; break; } default: res = GST_BASE_TRANSFORM_CLASS (parent_class)->query (trans, direction, query); break; } return res; } static gboolean gst_video_rate_propose_allocation (GstBaseTransform * trans, GstQuery * decide_query, GstQuery * query) { GstBaseTransformClass *klass = GST_BASE_TRANSFORM_CLASS (parent_class); gboolean res; /* We should always be passthrough */ g_return_val_if_fail (decide_query == NULL, FALSE); res = klass->propose_allocation (trans, NULL, query); if (res) { guint i = 0; guint n_allocation; guint down_min = 0; n_allocation = gst_query_get_n_allocation_pools (query); while (i < n_allocation) { GstBufferPool *pool = NULL; guint size, min, max; gst_query_parse_nth_allocation_pool (query, i, &pool, &size, &min, &max); if (min == max) { if (pool) gst_object_unref (pool); gst_query_remove_nth_allocation_pool (query, i); n_allocation--; down_min = MAX (min, down_min); continue; } gst_query_set_nth_allocation_pool (query, i, pool, size, min + 1, max); if (pool) gst_object_unref (pool); i++; } if (n_allocation == 0) { GstCaps *caps; GstVideoInfo info; gst_query_parse_allocation (query, &caps, NULL); gst_video_info_from_caps (&info, caps); gst_query_add_allocation_pool (query, NULL, info.size, down_min + 1, 0); } } return res; } static GstFlowReturn gst_video_rate_trans_ip_max_avg (GstVideoRate * videorate, GstBuffer * buf) { GstClockTime ts = GST_BUFFER_TIMESTAMP (buf); videorate->in++; if (!GST_CLOCK_TIME_IS_VALID (ts) || videorate->wanted_diff == 0) goto push; /* drop frames if they exceed our output rate */ if (GST_CLOCK_TIME_IS_VALID (videorate->last_ts)) { GstClockTimeDiff diff = videorate->segment.rate < 0 ? videorate->last_ts - ts : ts - videorate->last_ts; /* Drop buffer if its early compared to the desired frame rate and * the current average is higher than the desired average */ if (diff < videorate->wanted_diff && videorate->average < videorate->wanted_diff) goto drop; /* Update average */ if (videorate->average) { GstClockTimeDiff wanted_diff; if (G_LIKELY (videorate->average_period > videorate->wanted_diff)) wanted_diff = videorate->wanted_diff; else wanted_diff = videorate->average_period * 10; videorate->average = gst_util_uint64_scale_round (videorate->average, videorate->average_period - wanted_diff, videorate->average_period) + gst_util_uint64_scale_round (diff, wanted_diff, videorate->average_period); } else { videorate->average = diff; } } videorate->last_ts = ts; push: videorate->out++; return GST_FLOW_OK; drop: if (!videorate->silent) gst_video_rate_notify_drop (videorate); return GST_BASE_TRANSFORM_FLOW_DROPPED; } /* Check if downstream forces variable framerate (0/1) and if * it is the case, use variable framerate ourself * Otherwise compute the framerate from the 2 buffers that we * have already received and make use of it as wanted framerate */ static void gst_video_rate_check_variable_rate (GstVideoRate * videorate, GstBuffer * buffer) { GstStructure *st; gint fps_d, fps_n; GstCaps *srcpadcaps, *tmpcaps, *downstream_caps; GstPad *pad = NULL; srcpadcaps = gst_pad_get_current_caps (GST_BASE_TRANSFORM_SRC_PAD (videorate)); gst_video_guess_framerate (GST_BUFFER_PTS (buffer) - GST_BUFFER_PTS (videorate->prevbuf), &fps_n, &fps_d); tmpcaps = gst_caps_copy (srcpadcaps); st = gst_caps_get_structure (tmpcaps, 0); gst_structure_set (st, "framerate", GST_TYPE_FRACTION, fps_n, fps_d, NULL); gst_caps_unref (srcpadcaps); pad = gst_pad_get_peer (GST_BASE_TRANSFORM_SRC_PAD (videorate)); downstream_caps = gst_pad_query_caps (pad, NULL); if (pad && !gst_caps_can_intersect (tmpcaps, downstream_caps)) { videorate->force_variable_rate = TRUE; gst_caps_unref (downstream_caps); GST_DEBUG_OBJECT (videorate, "Downstream forces variable framerate" " respecting it"); goto done; } gst_caps_unref (downstream_caps); videorate->to_rate_numerator = fps_n; videorate->to_rate_denominator = fps_d; GST_INFO_OBJECT (videorate, "Computed framerate to %d/%d", videorate->to_rate_numerator, videorate->to_rate_denominator); videorate->updating_caps = TRUE; gst_base_transform_update_src_caps (GST_BASE_TRANSFORM (videorate), tmpcaps); /* also reconfigure sink so that buffer pool can be updated again */ gst_base_transform_reconfigure_sink (GST_BASE_TRANSFORM (videorate)); done: gst_caps_unref (tmpcaps); if (pad) gst_object_unref (pad); } static gboolean gst_video_rate_switch_mode_if_needed (GstVideoRate * videorate) { gboolean switch_mode; GstClockTime avg_period; gboolean skip = FALSE; GST_OBJECT_LOCK (videorate); avg_period = videorate->average_period_set; GST_OBJECT_UNLOCK (videorate); /* MT-safe switching between modes */ if (G_LIKELY (avg_period == videorate->average_period)) return skip; switch_mode = (avg_period == 0 || videorate->average_period == 0); if (!switch_mode) return skip; videorate->average_period = avg_period; videorate->last_ts = GST_CLOCK_TIME_NONE; if (avg_period) { /* enabling average mode */ videorate->average = 0; /* make sure no cached buffers from regular mode are left */ gst_video_rate_swap_prev (videorate, NULL, 0); } else { /* enable regular mode */ videorate->next_ts = GST_CLOCK_TIME_NONE; skip = TRUE; } /* max averaging mode has no latency, normal mode does */ gst_element_post_message (GST_ELEMENT (videorate), gst_message_new_latency (GST_OBJECT (videorate))); return skip; } static gboolean gst_video_rate_do_max_duplicate (GstVideoRate * videorate, GstBuffer * buffer, GstClockTime intime, GstClockTime prevtime, gint * count) { if (videorate->max_duplication_time <= 0) return TRUE; /* We already know that intime and prevtime are not out of order, based * on the previous condition. Using ABS in case rate < 0, in which case * the order is reversed. */ if (ABS (GST_CLOCK_DIFF (intime, prevtime)) > videorate->max_duplication_time) { GST_DEBUG_OBJECT (videorate, "The new buffer (%" GST_TIME_FORMAT ") is further away from previous buffer (%" GST_TIME_FORMAT ") than max-duplication-time (%" GST_TIME_FORMAT ")", GST_TIME_ARGS (intime), GST_TIME_ARGS (prevtime), GST_TIME_ARGS (videorate->max_duplication_time)); /* First send out enough buffers to actually reach the time of the * previous buffer */ if (videorate->segment.rate < 0.0) { while (videorate->next_ts > prevtime) { gst_video_rate_flush_prev (videorate, *count > 0, GST_CLOCK_TIME_NONE); *count += 1; } } else { while (videorate->next_ts <= prevtime) { gst_video_rate_flush_prev (videorate, *count > 0, GST_CLOCK_TIME_NONE); *count += 1; } } if (*count > 1) { videorate->dup += *count - 1; if (!videorate->silent) gst_video_rate_notify_duplicate (videorate); } /* The gap between the two buffers is too large. Don't fill it, just * let a discont through */ videorate->discont = TRUE; if (videorate->segment.rate < 0.0) { videorate->base_ts -= prevtime - intime; } else { videorate->base_ts += intime - prevtime; } videorate->next_ts = intime; /* Swap in new buffer and get rid of old buffer so that starting with * the next input buffer we output from the new position */ gst_video_rate_swap_prev (videorate, buffer, intime); return FALSE; } return TRUE; } static gboolean gst_video_rate_apply_pending_rate (GstVideoRate * videorate) { gboolean ret = FALSE; GST_OBJECT_LOCK (videorate); if (videorate->pending_rate == videorate->rate) goto done; ret = TRUE; videorate->base_ts += gst_util_uint64_scale (videorate->out_frame_count, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); videorate->rate = videorate->pending_rate; videorate->out_frame_count = 0; done: GST_OBJECT_UNLOCK (videorate); return ret; } static GstFlowReturn gst_video_rate_transform_ip (GstBaseTransform * trans, GstBuffer * buffer) { GstVideoRate *videorate; GstFlowReturn res = GST_BASE_TRANSFORM_FLOW_DROPPED; GstClockTime intime, in_ts, in_dur, last_ts; gboolean skip; videorate = GST_VIDEO_RATE (trans); /* make sure the denominators are not 0 */ if (videorate->from_rate_denominator == 0 || videorate->to_rate_denominator == 0) goto not_negotiated; if (videorate->to_rate_numerator == 0 && videorate->prevbuf && !videorate->force_variable_rate) { gst_video_rate_check_variable_rate (videorate, buffer); } skip = gst_video_rate_switch_mode_if_needed (videorate); if (videorate->average_period > 0) return gst_video_rate_trans_ip_max_avg (videorate, buffer); gst_video_rate_apply_pending_rate (videorate); in_ts = GST_BUFFER_TIMESTAMP (buffer); in_dur = GST_BUFFER_DURATION (buffer); if (G_UNLIKELY (!GST_CLOCK_TIME_IS_VALID (in_ts))) { /* For reverse playback, we need all input timestamps as we can't * guess from the previous buffers timestamp and duration */ if (G_UNLIKELY (videorate->segment.rate < 0.0)) goto invalid_buffer; in_ts = videorate->last_ts; if (G_UNLIKELY (!GST_CLOCK_TIME_IS_VALID (in_ts))) goto invalid_buffer; } /* get the time of the next expected buffer timestamp, we use this when the * next buffer has -1 as a timestamp */ last_ts = videorate->last_ts; videorate->last_ts = in_ts; if (GST_CLOCK_TIME_IS_VALID (in_dur) && videorate->segment.rate > 0.0) videorate->last_ts += in_dur; GST_DEBUG_OBJECT (videorate, "got buffer with timestamp %" GST_TIME_FORMAT, GST_TIME_ARGS (in_ts)); /* the input time is the time in the segment + all previously accumulated * segments */ intime = in_ts + videorate->segment.base; /* we need to have two buffers to compare */ if (videorate->prevbuf == NULL || videorate->drop_only) { /* We can calculate the duration of the buffer here if not given for * reverse playback. We need this later */ if (videorate->segment.rate < 0.0 && !GST_BUFFER_DURATION_IS_VALID (buffer)) { /* As we require valid timestamps all the time for reverse playback, we either * have a valid last_ts or we're at the very first buffer. */ if (!GST_CLOCK_TIME_IS_VALID (last_ts)) GST_BUFFER_DURATION (buffer) = videorate->segment.stop - in_ts; else GST_BUFFER_DURATION (buffer) = last_ts - in_ts; } gst_video_rate_swap_prev (videorate, buffer, intime); videorate->in++; if (!GST_CLOCK_TIME_IS_VALID (videorate->next_ts)) { /* new buffer, we expect to output a buffer that matches the first * timestamp in the segment */ if (videorate->skip_to_first || skip) { videorate->next_ts = intime; if (videorate->segment.rate < 0.0) { videorate->base_ts = videorate->segment.stop - in_ts; } else { videorate->base_ts = in_ts - videorate->segment.start; } videorate->out_frame_count = 0; } else { if (videorate->segment.rate < 0.0) { if (videorate->to_rate_numerator) { GstClockTime frame_duration = gst_util_uint64_scale (1, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); videorate->next_ts = videorate->segment.stop + videorate->segment.base; if (videorate->next_ts > frame_duration) videorate->next_ts = MAX (videorate->segment.start, videorate->next_ts - frame_duration); else videorate->next_ts = videorate->segment.start; } else { /* What else can we do? */ videorate->next_ts = intime; } } else { videorate->next_ts = videorate->segment.start + videorate->segment.base; } } } /* In drop-only mode we can already decide here if we should output the * current frame or drop it because it's coming earlier than our minimum * allowed frame period. This also keeps latency down to 0 frames */ if (videorate->drop_only) { if ((videorate->segment.rate > 0.0 && intime >= videorate->next_ts) || (videorate->segment.rate < 0.0 && intime <= videorate->next_ts)) { GstFlowReturn r; /* The buffer received from basetransform is guaranteed to be writable. * It just needs to be reffed so the buffer won't be consumed once pushed and * GstBaseTransform can get its reference back. */ if ((r = gst_video_rate_push_buffer (videorate, gst_buffer_ref (buffer), FALSE, GST_CLOCK_TIME_NONE)) != GST_FLOW_OK) { res = r; goto done; } } /* No need to keep the buffer around for longer */ gst_buffer_replace (&videorate->prevbuf, NULL); } } else { GstClockTime prevtime; gint count = 0; gint64 diff1 = 0, diff2 = 0; prevtime = videorate->prev_ts; GST_LOG_OBJECT (videorate, "BEGINNING prev buf %" GST_TIME_FORMAT " new buf %" GST_TIME_FORMAT " outgoing ts %" GST_TIME_FORMAT, GST_TIME_ARGS (prevtime), GST_TIME_ARGS (intime), GST_TIME_ARGS (videorate->next_ts)); videorate->in++; /* drop new buffer if it's before previous one */ if ((videorate->segment.rate > 0.0 && intime < prevtime) || (videorate->segment.rate < 0.0 && intime > prevtime)) { GST_DEBUG_OBJECT (videorate, "The new buffer (%" GST_TIME_FORMAT ") is before the previous buffer (%" GST_TIME_FORMAT "). Dropping new buffer.", GST_TIME_ARGS (intime), GST_TIME_ARGS (prevtime)); videorate->drop++; if (!videorate->silent) gst_video_rate_notify_drop (videorate); goto done; } if (!gst_video_rate_do_max_duplicate (videorate, buffer, intime, prevtime, &count)) goto done; /* got 2 buffers, see which one is the best */ do { GstClockTime next_ts; if (gst_video_rate_apply_pending_rate (videorate)) goto done; if (videorate->segment.rate < 0.0) { /* Make sure that we have a duration for this buffer. The previous * buffer already has a duration given by either exactly this code, * or the code above for the very first buffer */ g_assert (GST_BUFFER_DURATION_IS_VALID (videorate->prevbuf)); if (!GST_BUFFER_DURATION_IS_VALID (buffer)) GST_BUFFER_DURATION (buffer) = prevtime > intime ? prevtime - intime : 0; } else { /* Make sure that we have a duration for previous buffer */ if (!GST_BUFFER_DURATION_IS_VALID (videorate->prevbuf)) GST_BUFFER_DURATION (videorate->prevbuf) = intime > prevtime ? intime - prevtime : 0; } #ifndef ABSDIFF #define ABSDIFF(a, b) (((a) > (b)) ? (a) - (b) : (b) - (a)) #endif /* take absolute diffs */ if (videorate->segment.rate < 0.0) { GstClockTime next_end_ts; GstClockTime prev_endtime; GstClockTime in_endtime, base_ts_in_segment; next_ts = videorate->next_ts; if (!GST_CLOCK_TIME_IS_VALID (next_ts)) { GST_DEBUG_OBJECT (videorate, "Already reached segment start," "ignoring buffer"); break; } prev_endtime = prevtime + GST_BUFFER_DURATION (videorate->prevbuf); in_endtime = intime + GST_BUFFER_DURATION (buffer); if (videorate->to_rate_numerator) { GstClockTime frame_duration = gst_util_uint64_scale (1, videorate->to_rate_denominator * GST_SECOND, videorate->to_rate_numerator); next_end_ts = next_ts + frame_duration; } else { next_end_ts = next_ts + GST_BUFFER_DURATION (videorate->prevbuf); } base_ts_in_segment = videorate->segment.stop - videorate->base_ts; next_ts = base_ts_in_segment - ( (base_ts_in_segment - next_ts) * videorate->rate); next_end_ts = base_ts_in_segment - (MAX (0, (base_ts_in_segment - next_end_ts)) * videorate->rate); diff1 = ABSDIFF (prev_endtime, next_end_ts); diff2 = ABSDIFF (in_endtime, next_end_ts); GST_LOG_OBJECT (videorate, "diff with prev %" GST_TIME_FORMAT " diff with new %" GST_TIME_FORMAT " outgoing ts %" GST_TIME_FORMAT, GST_TIME_ARGS (diff1), GST_TIME_ARGS (diff2), GST_TIME_ARGS (next_end_ts)); } else { next_ts = videorate->base_ts + ((videorate->next_ts - videorate->base_ts) * videorate->rate); diff1 = ABSDIFF (prevtime, next_ts); diff2 = ABSDIFF (intime, next_ts); GST_LOG_OBJECT (videorate, "diff with prev %" GST_TIME_FORMAT " diff with new %" GST_TIME_FORMAT " outgoing ts %" GST_TIME_FORMAT, GST_TIME_ARGS (diff1), GST_TIME_ARGS (diff2), GST_TIME_ARGS (next_ts)); } /* output first one when its the best */ if (diff1 <= diff2) { GstFlowReturn r; count++; /* on error the _flush function posted a warning already */ if ((r = gst_video_rate_flush_prev (videorate, count > 1, intime)) != GST_FLOW_OK) { res = r; goto done; } } /* continue while the first one was the best, if they were equal avoid * going into an infinite loop */ } while (diff1 < diff2); /* if we outputted the first buffer more then once, we have dups */ if (count > 1) { videorate->dup += count - 1; if (!videorate->silent) gst_video_rate_notify_duplicate (videorate); } /* if we didn't output the first buffer, we have a drop */ else if (count == 0) { videorate->drop++; if (!videorate->silent) gst_video_rate_notify_drop (videorate); GST_LOG_OBJECT (videorate, "new is best, old never used, drop, outgoing ts %" GST_TIME_FORMAT, GST_TIME_ARGS (videorate->next_ts)); } GST_LOG_OBJECT (videorate, "END, putting new in old, diff1 %" GST_TIME_FORMAT ", diff2 %" GST_TIME_FORMAT ", next_ts %" GST_TIME_FORMAT ", in %" G_GUINT64_FORMAT ", out %" G_GUINT64_FORMAT ", drop %" G_GUINT64_FORMAT ", dup %" G_GUINT64_FORMAT, GST_TIME_ARGS (diff1), GST_TIME_ARGS (diff2), GST_TIME_ARGS (videorate->next_ts), videorate->in, videorate->out, videorate->drop, videorate->dup); /* swap in new one when it's the best */ gst_video_rate_swap_prev (videorate, buffer, intime); } done: return res; /* ERRORS */ not_negotiated: { GST_WARNING_OBJECT (videorate, "no framerate negotiated"); res = GST_FLOW_NOT_NEGOTIATED; goto done; } invalid_buffer: { GST_WARNING_OBJECT (videorate, "Got buffer with GST_CLOCK_TIME_NONE timestamp, discarding it"); res = GST_BASE_TRANSFORM_FLOW_DROPPED; goto done; } } static gboolean gst_video_rate_start (GstBaseTransform * trans) { gst_video_rate_reset (GST_VIDEO_RATE (trans)); return TRUE; } static gboolean gst_video_rate_stop (GstBaseTransform * trans) { gst_video_rate_reset (GST_VIDEO_RATE (trans)); return TRUE; } static void gst_videorate_update_duration (GstVideoRate * videorate) { GstMessage *m; m = gst_message_new_duration_changed (GST_OBJECT (videorate)); gst_element_post_message (GST_ELEMENT (videorate), m); } static void gst_video_rate_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstVideoRate *videorate = GST_VIDEO_RATE (object); gboolean latency_changed = FALSE; GST_OBJECT_LOCK (videorate); switch (prop_id) { case PROP_SILENT: videorate->silent = g_value_get_boolean (value); break; case PROP_NEW_PREF: videorate->new_pref = g_value_get_double (value); break; case PROP_SKIP_TO_FIRST: videorate->skip_to_first = g_value_get_boolean (value); break; case PROP_DROP_ONLY:{ gboolean new_value = g_value_get_boolean (value); /* Latency changes if we switch drop-only mode */ latency_changed = new_value != videorate->drop_only; videorate->drop_only = g_value_get_boolean (value); goto reconfigure; } case PROP_AVERAGE_PERIOD: videorate->average_period_set = g_value_get_uint64 (value); break; case PROP_MAX_RATE: g_atomic_int_set (&videorate->max_rate, g_value_get_int (value)); goto reconfigure; case PROP_RATE: videorate->pending_rate = g_value_get_double (value); GST_OBJECT_UNLOCK (videorate); gst_videorate_update_duration (videorate); return; case PROP_MAX_DUPLICATION_TIME: videorate->max_duplication_time = g_value_get_uint64 (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_OBJECT_UNLOCK (videorate); return; reconfigure: GST_OBJECT_UNLOCK (videorate); gst_base_transform_reconfigure_src (GST_BASE_TRANSFORM (videorate)); if (latency_changed) { gst_element_post_message (GST_ELEMENT (videorate), gst_message_new_latency (GST_OBJECT (videorate))); } } static void gst_video_rate_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVideoRate *videorate = GST_VIDEO_RATE (object); GST_OBJECT_LOCK (videorate); switch (prop_id) { case PROP_IN: g_value_set_uint64 (value, videorate->in); break; case PROP_OUT: g_value_set_uint64 (value, videorate->out); break; case PROP_DUP: g_value_set_uint64 (value, videorate->dup); break; case PROP_DROP: g_value_set_uint64 (value, videorate->drop); break; case PROP_SILENT: g_value_set_boolean (value, videorate->silent); break; case PROP_NEW_PREF: g_value_set_double (value, videorate->new_pref); break; case PROP_SKIP_TO_FIRST: g_value_set_boolean (value, videorate->skip_to_first); break; case PROP_DROP_ONLY: g_value_set_boolean (value, videorate->drop_only); break; case PROP_AVERAGE_PERIOD: g_value_set_uint64 (value, videorate->average_period_set); break; case PROP_MAX_RATE: g_value_set_int (value, g_atomic_int_get (&videorate->max_rate)); break; case PROP_RATE: g_value_set_double (value, videorate->pending_rate); break; case PROP_MAX_DUPLICATION_TIME: g_value_set_uint64 (value, videorate->max_duplication_time); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } GST_OBJECT_UNLOCK (videorate); } static gboolean plugin_init (GstPlugin * plugin) { GST_DEBUG_CATEGORY_INIT (video_rate_debug, "videorate", 0, "VideoRate stream fixer"); return gst_element_register (plugin, "videorate", GST_RANK_NONE, GST_TYPE_VIDEO_RATE); } GST_PLUGIN_DEFINE (GST_VERSION_MAJOR, GST_VERSION_MINOR, videorate, "Adjusts video frames", plugin_init, VERSION, GST_LICENSE, GST_PACKAGE_NAME, GST_PACKAGE_ORIGIN)