/* * GStreamer * Copyright (C) 2009 Carl-Anton Ingmarsson * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-vdpauvideopostprocess * * FIXME:Describe vdpaumpegdec here. * * * Example launch line * |[ * gst-launch -v -m fakesrc ! vdpauvideopostprocess ! fakesink silent=TRUE * ]| * */ /* * TODO: * + add support for more postprocessing options * + mixing videos. (This should perhaps be done in a separate element based on * VdpOutputSurface) */ #ifdef HAVE_CONFIG_H # include #endif #include #include #include "gstvdputils.h" #include "gstvdpvideobuffer.h" #include "gstvdpoutputbuffer.h" #include "gstvdpvideopostprocess.h" GST_DEBUG_CATEGORY_STATIC (gst_vdp_vpp_debug); #define GST_CAT_DEFAULT gst_vdp_vpp_debug /* Filter signals and args */ enum { /* FILL ME */ LAST_SIGNAL }; enum { PROP_0, PROP_FORCE_ASPECT_RATIO, PROP_DEINTERLACE_MODE, PROP_DEINTERLACE_METHOD, PROP_NOISE_REDUCTION, PROP_SHARPENING, PROP_INVERSE_TELECINE }; /* the capabilities of the inputs and outputs. * * describe the real formats here. */ static GstStaticPadTemplate sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VDP_VIDEO_CAPS)); static GstStaticPadTemplate src_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VDP_OUTPUT_CAPS)); #define DEBUG_INIT(bla) \ GST_DEBUG_CATEGORY_INIT (gst_vdp_vpp_debug, "vdpauvideopostprocess", 0, "VDPAU video surface to output surface"); GST_BOILERPLATE_FULL (GstVdpVideoPostProcess, gst_vdp_vpp, GstElement, GST_TYPE_ELEMENT, DEBUG_INIT); static void gst_vdp_vpp_finalize (GObject * object); #define GST_TYPE_VDP_DEINTERLACE_METHODS (gst_vdp_deinterlace_methods_get_type ()) static GType gst_vdp_deinterlace_methods_get_type (void) { static GType deinterlace_methods_type = 0; static const GEnumValue methods_types[] = { {GST_VDP_DEINTERLACE_METHOD_BOB, "Vertically scale a single field to the size of a single frame.", "bob"}, {GST_VDP_DEINTERLACE_METHOD_TEMPORAL, "Motion Adaptive: Simple Detection", "temporal"}, {GST_VDP_DEINTERLACE_METHOD_TEMPORAL_SPATIAL, "Motion Adaptive: Advanced Detection", "temporal-spatial"}, {0, NULL, NULL}, }; if (!deinterlace_methods_type) { deinterlace_methods_type = g_enum_register_static ("GstVdpDeinterlaceMethods", methods_types); } return deinterlace_methods_type; } #define GST_TYPE_VDP_DEINTERLACE_MODES (gst_vdp_deinterlace_modes_get_type ()) static GType gst_vdp_deinterlace_modes_get_type (void) { static GType deinterlace_modes_type = 0; static const GEnumValue modes_types[] = { {GST_VDP_DEINTERLACE_MODE_AUTO, "Auto detection", "auto"}, {GST_VDP_DEINTERLACE_MODE_INTERLACED, "Enfore deinterlacing", "interlaced"}, {GST_VDP_DEINTERLACE_MODE_DISABLED, "Run in passthrough mode", "disabled"}, {0, NULL, NULL}, }; if (!deinterlace_modes_type) { deinterlace_modes_type = g_enum_register_static ("GstVdpDeinterlaceModes", modes_types); } return deinterlace_modes_type; } static void gst_vdp_vpp_set_attribute_float (GstVdpVideoPostProcess * vpp, VdpVideoMixerAttribute attribute, gfloat value) { VdpVideoMixerAttribute attributes[1]; const void *attribute_values[1]; VdpStatus status; attributes[0] = attribute; attribute_values[0] = &value; status = vpp->device->vdp_video_mixer_set_attribute_values (vpp->mixer, 1, attributes, attribute_values); if (status != VDP_STATUS_OK) { GST_WARNING_OBJECT (vpp, "Couldn't set noise reduction level on mixer, " "error returned from vdpau was: %s", vpp->device->vdp_get_error_string (status)); } } static void gst_vdp_vpp_activate_feature (GstVdpVideoPostProcess * vpp, VdpVideoMixerFeature feature, gboolean activate) { VdpVideoMixerFeature features[1]; VdpBool enable[1]; VdpStatus status; features[0] = feature; if (activate) enable[0] = VDP_TRUE; else enable[0] = VDP_FALSE; status = vpp->device->vdp_video_mixer_set_feature_enables (vpp->mixer, 1, features, enable); if (status != VDP_STATUS_OK) { GST_WARNING_OBJECT (vpp, "Couldn't set deinterlace method on mixer, " "error returned from vdpau was: %s", vpp->device->vdp_get_error_string (status)); } } static VdpVideoMixerFeature gst_vdp_feature_from_deinterlace_method (GstVdpDeinterlaceMethods method) { gint i; VdpVideoMixerFeature feature; typedef struct { GstVdpDeinterlaceMethods method; VdpVideoMixerFeature feature; } VdpDeinterlaceMethod; VdpDeinterlaceMethod deinterlace_methods[2] = { {GST_VDP_DEINTERLACE_METHOD_TEMPORAL, VDP_VIDEO_MIXER_FEATURE_DEINTERLACE_TEMPORAL}, {GST_VDP_DEINTERLACE_METHOD_TEMPORAL_SPATIAL, VDP_VIDEO_MIXER_FEATURE_DEINTERLACE_TEMPORAL_SPATIAL} }; for (i = 0; i < G_N_ELEMENTS (deinterlace_methods); i++) { if (deinterlace_methods[i].method == method) { feature = deinterlace_methods[i].feature; break; } } return feature; } static void gst_vdp_vpp_activate_deinterlace_method (GstVdpVideoPostProcess * vpp, GstVdpDeinterlaceMethods method, gboolean activate) { gst_vdp_vpp_activate_feature (vpp, gst_vdp_feature_from_deinterlace_method (method), activate); } static void gst_vdp_picture_clear (GstVdpPicture pic) { if (pic.buf) { gst_buffer_unref (GST_BUFFER (pic.buf)); pic.buf = NULL; } } static gboolean gst_vdp_vpp_is_interlaced (GstVdpVideoPostProcess * vpp) { if (vpp->mode == GST_VDP_DEINTERLACE_MODE_INTERLACED) return TRUE; if (vpp->mode == GST_VDP_DEINTERLACE_MODE_AUTO && vpp->interlaced) return TRUE; return FALSE; } static guint gst_vdp_vpp_get_required_pictures (GstVdpVideoPostProcess * vpp) { guint ret; if (vpp->noise_reduction != 0.0) return 2; if (!gst_vdp_vpp_is_interlaced (vpp)) return 1; switch (vpp->method) { case GST_VDP_DEINTERLACE_METHOD_BOB: ret = 1; break; case GST_VDP_DEINTERLACE_METHOD_TEMPORAL: case GST_VDP_DEINTERLACE_METHOD_TEMPORAL_SPATIAL: ret = 2; break; } return ret; } static gboolean gst_vdp_vpp_get_next_picture (GstVdpVideoPostProcess * vpp, GstVdpPicture * current_pic, guint32 * video_surfaces_past_count, VdpVideoSurface * video_surfaces_past, guint32 * video_surfaces_future_count, VdpVideoSurface * video_surfaces_future) { gint i; gint required_pictures; required_pictures = gst_vdp_vpp_get_required_pictures (vpp); if (vpp->n_future_pictures < required_pictures) return FALSE; *current_pic = vpp->future_pictures[0]; for (i = 0; i < vpp->n_future_pictures - 1; i++) { vpp->future_pictures[i] = vpp->future_pictures[i + 1]; } vpp->future_pictures[vpp->n_future_pictures - 1].buf = NULL; vpp->n_future_pictures--; *video_surfaces_future_count = vpp->n_future_pictures; for (i = 0; i < vpp->n_future_pictures; i++) video_surfaces_future[i] = vpp->future_pictures[i].buf->surface; *video_surfaces_past_count = vpp->n_past_pictures; for (i = 0; i < vpp->n_past_pictures; i++) video_surfaces_past[i] = vpp->past_pictures[i].buf->surface; if (vpp->n_past_pictures == MAX_PICTURES) { gst_vdp_picture_clear (vpp->past_pictures[MAX_PICTURES - 1]); vpp->n_past_pictures--; } /* move every picture upwards one step in the array */ for (i = vpp->n_past_pictures; i > 0; i--) { vpp->past_pictures[i] = vpp->past_pictures[i - 1]; } /* add the picture to the past surfaces */ vpp->past_pictures[0] = *current_pic; vpp->n_past_pictures++; return TRUE; } static void gst_vdp_vpp_add_buffer (GstVdpVideoPostProcess * vpp, GstVdpVideoBuffer * buf) { gboolean repeated; gboolean tff; gboolean onefield; GstVdpPicture pic1, pic2; if (!gst_vdp_vpp_is_interlaced (vpp)) { pic1.buf = buf; pic1.structure = VDP_VIDEO_MIXER_PICTURE_STRUCTURE_FRAME; pic1.timestamp = GST_BUFFER_TIMESTAMP (buf); vpp->future_pictures[vpp->n_future_pictures++] = pic1; return; } repeated = GST_BUFFER_FLAG_IS_SET (buf, GST_VIDEO_BUFFER_RFF); tff = GST_BUFFER_FLAG_IS_SET (buf, GST_VIDEO_BUFFER_TFF); onefield = GST_BUFFER_FLAG_IS_SET (buf, GST_VIDEO_BUFFER_ONEFIELD); pic1.buf = buf; pic2.buf = (GstVdpVideoBuffer *) gst_buffer_ref (GST_BUFFER (buf)); if (tff) { pic1.structure = VDP_VIDEO_MIXER_PICTURE_STRUCTURE_TOP_FIELD; pic2.structure = VDP_VIDEO_MIXER_PICTURE_STRUCTURE_BOTTOM_FIELD; } else { pic1.structure = VDP_VIDEO_MIXER_PICTURE_STRUCTURE_BOTTOM_FIELD; pic1.structure = VDP_VIDEO_MIXER_PICTURE_STRUCTURE_TOP_FIELD; } pic1.timestamp = GST_BUFFER_TIMESTAMP (buf); pic2.timestamp = pic1.timestamp + vpp->field_duration; if (repeated) { vpp->future_pictures[vpp->n_future_pictures++] = pic1; pic1.timestamp = pic2.timestamp; vpp->future_pictures[vpp->n_future_pictures++] = pic1; gst_buffer_ref (GST_BUFFER (pic1.buf)); pic2.timestamp += vpp->field_duration; vpp->future_pictures[vpp->n_future_pictures++] = pic2; } else if (!onefield) { vpp->future_pictures[vpp->n_future_pictures++] = pic1; vpp->future_pictures[vpp->n_future_pictures++] = pic2; } else { vpp->future_pictures[vpp->n_future_pictures++] = pic1; gst_buffer_unref (GST_BUFFER (pic2.buf)); } } static GstFlowReturn gst_vdp_vpp_create_mixer (GstVdpVideoPostProcess * vpp, GstVdpDevice * device) { #define VDP_NUM_MIXER_PARAMETER 3 #define MAX_NUM_FEATURES 5 GstStructure *structure; gint chroma_type; gint width, height; VdpStatus status; VdpVideoMixerFeature features[5]; guint n_features = 0; VdpVideoMixerParameter parameters[VDP_NUM_MIXER_PARAMETER] = { VDP_VIDEO_MIXER_PARAMETER_VIDEO_SURFACE_WIDTH, VDP_VIDEO_MIXER_PARAMETER_VIDEO_SURFACE_HEIGHT, VDP_VIDEO_MIXER_PARAMETER_CHROMA_TYPE }; const void *parameter_values[VDP_NUM_MIXER_PARAMETER]; structure = gst_caps_get_structure (GST_PAD_CAPS (vpp->sinkpad), 0); if (!gst_structure_get_int (structure, "chroma-type", &chroma_type) || !gst_structure_get_int (structure, "width", &width) || !gst_structure_get_int (structure, "height", &height)) return GST_FLOW_ERROR; parameter_values[0] = &width; parameter_values[1] = &height; parameter_values[2] = &chroma_type; if (gst_vdp_vpp_is_interlaced (vpp) && vpp->method != GST_VDP_DEINTERLACE_METHOD_BOB) { features[n_features++] = gst_vdp_feature_from_deinterlace_method (vpp->method); } if (vpp->noise_reduction > 0.0) features[n_features++] = VDP_VIDEO_MIXER_FEATURE_NOISE_REDUCTION; if (vpp->sharpening != 0.0) features[n_features++] = VDP_VIDEO_MIXER_FEATURE_SHARPNESS; if (vpp->inverse_telecine) features[n_features++] = VDP_VIDEO_MIXER_FEATURE_INVERSE_TELECINE; status = device->vdp_video_mixer_create (device->device, n_features, features, VDP_NUM_MIXER_PARAMETER, parameters, parameter_values, &vpp->mixer); if (status != VDP_STATUS_OK) { GST_ELEMENT_ERROR (vpp, RESOURCE, READ, ("Could not create vdpau video mixer"), ("Error returned from vdpau was: %s", device->vdp_get_error_string (status))); return GST_FLOW_ERROR; } vpp->device = g_object_ref (device); if (vpp->noise_reduction > 0.0) { gst_vdp_vpp_set_attribute_float (vpp, VDP_VIDEO_MIXER_ATTRIBUTE_NOISE_REDUCTION_LEVEL, vpp->noise_reduction); } if (vpp->sharpening != 0.0) { gst_vdp_vpp_set_attribute_float (vpp, VDP_VIDEO_MIXER_ATTRIBUTE_SHARPNESS_LEVEL, vpp->sharpening); } return GST_FLOW_OK; } static GstFlowReturn gst_vdp_vpp_alloc_output_buffer (GstVdpVideoPostProcess * vpp, GstCaps * caps, GstVdpOutputBuffer ** outbuf) { GstFlowReturn ret; ret = gst_pad_alloc_buffer_and_set_caps (vpp->srcpad, 0, 0, caps, (GstBuffer **) outbuf); if (ret != GST_FLOW_OK) return ret; if (!vpp->device) { ret = gst_vdp_vpp_create_mixer (vpp, (*outbuf)->device); } if (ret != GST_FLOW_OK) gst_buffer_unref (GST_BUFFER (*outbuf)); return ret; } static gint gst_greatest_common_divisor (gint a, gint b) { while (b != 0) { int temp = a; a = b; b = temp % b; } return ABS (a); } static gboolean gst_fraction_double (gint * n_out, gint * d_out) { gint n, d, gcd; n = *n_out; d = *d_out; if (d == 0) return FALSE; if (n == 0 || (n == G_MAXINT && d == 1)) return TRUE; gcd = gst_greatest_common_divisor (n, d); n /= gcd; d /= gcd; if (G_MAXINT / 2 >= ABS (n)) { n *= 2; } else if (d >= 2) { d /= 2; } else return FALSE; *n_out = n; *d_out = d; return TRUE; } static gboolean gst_vdp_vpp_sink_setcaps (GstPad * pad, GstCaps * caps) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (gst_pad_get_parent (pad)); GstStructure *structure; GstCaps *output_caps, *allowed_caps, *src_caps; gboolean res; /* extract interlaced flag */ structure = gst_caps_get_structure (caps, 0); gst_structure_get_boolean (structure, "interlaced", &vpp->interlaced); allowed_caps = gst_pad_get_allowed_caps (vpp->srcpad); structure = gst_caps_get_structure (allowed_caps, 0); output_caps = gst_vdp_video_to_output_caps (caps); src_caps = gst_caps_intersect (output_caps, allowed_caps); gst_caps_truncate (src_caps); if (gst_caps_is_empty (src_caps)) goto invalid_caps; GST_DEBUG ("output_caps: %" GST_PTR_FORMAT " allowed_caps: %" GST_PTR_FORMAT " src_caps: %" GST_PTR_FORMAT, output_caps, allowed_caps, src_caps); gst_caps_unref (output_caps); gst_caps_unref (allowed_caps); if (gst_vdp_vpp_is_interlaced (vpp)) { gint fps_n, fps_d; structure = gst_caps_get_structure (src_caps, 0); if (!gst_structure_get_fraction (structure, "framerate", &fps_n, &fps_d)) { gst_caps_unref (src_caps); goto invalid_caps; } gst_fraction_double (&fps_n, &fps_d); gst_structure_set (structure, "framerate", GST_TYPE_FRACTION, fps_n, fps_d, NULL); gst_structure_remove_field (structure, "interlaced"); vpp->field_duration = gst_util_uint64_scale (GST_SECOND, fps_d, fps_n); } res = gst_pad_set_caps (vpp->srcpad, src_caps); done: gst_object_unref (vpp); return res; invalid_caps: GST_ERROR_OBJECT (vpp, "invalid caps: %" GST_PTR_FORMAT, caps); res = FALSE; goto done; } static void gst_vdp_vpp_flush (GstVdpVideoPostProcess * vpp) { gint i; for (i = 0; i < vpp->n_future_pictures; i++) { gst_vdp_picture_clear (vpp->future_pictures[i]); } vpp->n_future_pictures = 0; for (i = 0; i < vpp->n_past_pictures; i++) { gst_vdp_picture_clear (vpp->past_pictures[i]); } vpp->n_past_pictures = 0; } static void gst_vdp_vpp_start (GstVdpVideoPostProcess * vpp) { gint i; vpp->interlaced = FALSE; vpp->field_duration = GST_CLOCK_TIME_NONE; vpp->mixer = VDP_INVALID_HANDLE; vpp->device = NULL; for (i = 0; i < MAX_PICTURES; i++) { vpp->future_pictures[i].buf = NULL; vpp->past_pictures[i].buf = NULL; } vpp->n_future_pictures = 0; vpp->n_past_pictures = 0; } static void gst_vdp_vpp_stop (GstVdpVideoPostProcess * vpp) { if (vpp->mixer != VDP_INVALID_HANDLE) vpp->device->vdp_video_mixer_destroy (vpp->mixer); if (vpp->device) g_object_unref (vpp->device); gst_vdp_vpp_flush (vpp); } static GstFlowReturn gst_vdp_vpp_chain (GstPad * pad, GstBuffer * buffer) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (gst_pad_get_parent (pad)); GstFlowReturn ret = GST_FLOW_OK; GstVdpPicture current_pic; guint32 video_surfaces_past_count; VdpVideoSurface video_surfaces_past[MAX_PICTURES]; guint32 video_surfaces_future_count; VdpVideoSurface video_surfaces_future[MAX_PICTURES]; if (G_UNLIKELY (GST_BUFFER_FLAG_IS_SET (buffer, GST_BUFFER_FLAG_DISCONT))) { GST_DEBUG_OBJECT (vpp, "Received discont buffer"); gst_vdp_vpp_flush (vpp); } gst_vdp_vpp_add_buffer (vpp, GST_VDP_VIDEO_BUFFER (buffer)); while (gst_vdp_vpp_get_next_picture (vpp, ¤t_pic, &video_surfaces_past_count, video_surfaces_past, &video_surfaces_future_count, video_surfaces_future)) { GstVdpOutputBuffer *outbuf; GstStructure *structure; GstVideoRectangle src_r = { 0, } , dest_r = { 0,}; gint par_n, par_d; VdpRect rect; GstVdpDevice *device; VdpStatus status; ret = gst_vdp_vpp_alloc_output_buffer (vpp, GST_PAD_CAPS (vpp->srcpad), &outbuf); if (ret != GST_FLOW_OK) break; structure = gst_caps_get_structure (GST_BUFFER_CAPS (buffer), 0); if (!gst_structure_get_int (structure, "width", &src_r.w) || !gst_structure_get_int (structure, "height", &src_r.h)) goto invalid_caps; if (gst_structure_get_fraction (structure, "pixel-aspect-ratio", &par_n, &par_d)) { gint new_width; new_width = gst_util_uint64_scale_int (src_r.w, par_n, par_d); src_r.x += (src_r.w - new_width) / 2; src_r.w = new_width; } structure = gst_caps_get_structure (GST_BUFFER_CAPS (outbuf), 0); if (!gst_structure_get_int (structure, "width", &dest_r.w) || !gst_structure_get_int (structure, "height", &dest_r.h)) goto invalid_caps; if (vpp->force_aspect_ratio) { GstVideoRectangle res_r; gst_video_sink_center_rect (src_r, dest_r, &res_r, TRUE); rect.x0 = res_r.x; rect.x1 = res_r.w + res_r.x; rect.y0 = res_r.y; rect.y1 = res_r.h + res_r.y; } else { rect.x0 = 0; rect.x1 = dest_r.w; rect.y0 = 0; rect.y1 = dest_r.h; } device = vpp->device; status = device->vdp_video_mixer_render (vpp->mixer, VDP_INVALID_HANDLE, NULL, current_pic.structure, video_surfaces_past_count, video_surfaces_past, current_pic.buf->surface, video_surfaces_future_count, video_surfaces_future, NULL, outbuf->surface, NULL, &rect, 0, NULL); if (status != VDP_STATUS_OK) { GST_ELEMENT_ERROR (vpp, RESOURCE, READ, ("Could not post process frame"), ("Error returned from vdpau was: %s", device->vdp_get_error_string (status))); ret = GST_FLOW_ERROR; goto done; } GST_BUFFER_TIMESTAMP (outbuf) = current_pic.timestamp; if (gst_vdp_vpp_is_interlaced (vpp)) GST_BUFFER_DURATION (outbuf) = vpp->field_duration; else GST_BUFFER_DURATION (outbuf) = GST_BUFFER_DURATION (current_pic.buf); if (GST_BUFFER_FLAG_IS_SET (current_pic.buf, GST_BUFFER_FLAG_DISCONT)) GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_DISCONT); if (GST_BUFFER_FLAG_IS_SET (current_pic.buf, GST_BUFFER_FLAG_PREROLL)) GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_PREROLL); if (GST_BUFFER_FLAG_IS_SET (current_pic.buf, GST_BUFFER_FLAG_GAP)) GST_BUFFER_FLAG_SET (outbuf, GST_BUFFER_FLAG_GAP); ret = gst_pad_push (vpp->srcpad, GST_BUFFER (outbuf)); if (ret != GST_FLOW_OK) break; continue; invalid_caps: gst_buffer_unref (GST_BUFFER (outbuf)); ret = GST_FLOW_ERROR; break; } done: gst_object_unref (vpp); return ret; } static GstCaps * gst_vdp_vpp_sink_getcaps (GstPad * pad) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (gst_pad_get_parent (pad)); GstCaps *caps; if (vpp->device) caps = gst_vdp_video_buffer_get_allowed_video_caps (vpp->device); else caps = gst_static_pad_template_get_caps (&sink_template); gst_object_unref (vpp); return caps; } static GstFlowReturn gst_vdp_vpp_sink_bufferalloc (GstPad * pad, guint64 offset, guint size, GstCaps * caps, GstBuffer ** buf) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (gst_pad_get_parent (pad)); GstVdpOutputBuffer *outbuf; GstFlowReturn ret = GST_FLOW_ERROR; GstVdpDevice *device = NULL; GstStructure *structure; gint width, height; gint chroma_type; if (!vpp->device) { /* if we haven't got a device yet we must alloc a buffer downstream to get it */ GstCaps *src_caps = gst_pad_get_allowed_caps (vpp->srcpad); gst_pad_fixate_caps (vpp->srcpad, src_caps); ret = gst_pad_alloc_buffer (vpp->srcpad, 0, 0, src_caps, (GstBuffer **) & outbuf); gst_caps_unref (src_caps); if (ret != GST_FLOW_OK) goto error; device = outbuf->device; gst_buffer_unref (GST_BUFFER (outbuf)); } else device = vpp->device; structure = gst_caps_get_structure (caps, 0); if (!gst_structure_get_int (structure, "width", &width) || !gst_structure_get_int (structure, "height", &height) || !gst_structure_get_int (structure, "chroma-type", &chroma_type)) goto error; *buf = GST_BUFFER (gst_vdp_video_buffer_new (device, chroma_type, width, height)); if (*buf == NULL) goto error; GST_BUFFER_SIZE (*buf) = size; GST_BUFFER_OFFSET (*buf) = offset; gst_buffer_set_caps (*buf, caps); ret = GST_FLOW_OK; error: gst_object_unref (vpp); return ret; } static gboolean gst_vdp_vpp_sink_event (GstPad * pad, GstEvent * event) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (gst_pad_get_parent (pad)); gboolean res; switch (GST_EVENT_TYPE (event)) { case GST_EVENT_FLUSH_STOP: { GST_DEBUG_OBJECT (vpp, "flush stop"); gst_vdp_vpp_flush (vpp); res = gst_pad_push_event (vpp->srcpad, event); break; } default: res = gst_pad_event_default (pad, event); } gst_object_unref (vpp); return res; } static GstStateChangeReturn gst_vdp_vpp_change_state (GstElement * element, GstStateChange transition) { GstVdpVideoPostProcess *vpp; GstStateChangeReturn ret; vpp = GST_VDP_VIDEO_POST_PROCESS (element); switch (transition) { case GST_STATE_CHANGE_READY_TO_PAUSED: gst_vdp_vpp_start (vpp); break; default: break; } ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY: gst_vdp_vpp_stop (vpp); break; default: break; } return ret; } /* GObject vmethod implementations */ static void gst_vdp_vpp_get_property (GObject * object, guint property_id, GValue * value, GParamSpec * pspec) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (object); switch (property_id) { case PROP_FORCE_ASPECT_RATIO: g_value_set_boolean (value, vpp->force_aspect_ratio); break; case PROP_DEINTERLACE_MODE: g_value_set_enum (value, vpp->mode); break; case PROP_DEINTERLACE_METHOD: g_value_set_enum (value, vpp->method); break; case PROP_NOISE_REDUCTION: g_value_set_float (value, vpp->noise_reduction); break; case PROP_SHARPENING: g_value_set_float (value, vpp->sharpening); break; case PROP_INVERSE_TELECINE: g_value_set_boolean (value, vpp->inverse_telecine); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); break; } } /* GObject vmethod implementations */ static void gst_vdp_vpp_set_property (GObject * object, guint property_id, const GValue * value, GParamSpec * pspec) { GstVdpVideoPostProcess *vpp = GST_VDP_VIDEO_POST_PROCESS (object); switch (property_id) { case PROP_FORCE_ASPECT_RATIO: vpp->force_aspect_ratio = g_value_get_boolean (value); break; case PROP_DEINTERLACE_MODE: vpp->mode = g_value_get_enum (value); break; case PROP_DEINTERLACE_METHOD: { GstVdpDeinterlaceMethods oldvalue; oldvalue = vpp->method; vpp->method = g_value_get_enum (value); if (oldvalue == vpp->method) break; if (vpp->device) { if (oldvalue != GST_VDP_DEINTERLACE_METHOD_BOB) gst_vdp_vpp_activate_deinterlace_method (vpp, oldvalue, FALSE); if (vpp->method != GST_VDP_DEINTERLACE_METHOD_BOB && vpp->device) gst_vdp_vpp_activate_deinterlace_method (vpp, oldvalue, TRUE); } break; } case PROP_NOISE_REDUCTION: { gfloat old_value; old_value = vpp->noise_reduction; vpp->noise_reduction = g_value_get_float (value); if (vpp->noise_reduction == old_value) break; if (vpp->device) { if (vpp->noise_reduction == 0.0) gst_vdp_vpp_activate_feature (vpp, VDP_VIDEO_MIXER_FEATURE_NOISE_REDUCTION, FALSE); if (old_value == 0.0) gst_vdp_vpp_activate_feature (vpp, VDP_VIDEO_MIXER_FEATURE_NOISE_REDUCTION, TRUE); gst_vdp_vpp_set_attribute_float (vpp, VDP_VIDEO_MIXER_ATTRIBUTE_NOISE_REDUCTION_LEVEL, vpp->noise_reduction); } break; } case PROP_SHARPENING: { gfloat old_value; old_value = vpp->sharpening; vpp->sharpening = g_value_get_float (value); if (vpp->sharpening == old_value) break; if (vpp->device) { if (vpp->sharpening == 0.0) gst_vdp_vpp_activate_feature (vpp, VDP_VIDEO_MIXER_FEATURE_SHARPNESS, FALSE); if (old_value == 0.0) gst_vdp_vpp_activate_feature (vpp, VDP_VIDEO_MIXER_FEATURE_SHARPNESS, TRUE); gst_vdp_vpp_set_attribute_float (vpp, VDP_VIDEO_MIXER_ATTRIBUTE_SHARPNESS_LEVEL, vpp->sharpening); } break; } case PROP_INVERSE_TELECINE: { vpp->inverse_telecine = g_value_get_boolean (value); if (vpp->device) { gst_vdp_vpp_activate_feature (vpp, VDP_VIDEO_MIXER_FEATURE_INVERSE_TELECINE, vpp->inverse_telecine); } break; } default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, property_id, pspec); break; } } /* GType vmethod implementations */ static void gst_vdp_vpp_base_init (gpointer gclass) { GstElementClass *element_class = GST_ELEMENT_CLASS (gclass); gst_element_class_set_details_simple (element_class, "VDPAU Mpeg Decoder", "Filter/Converter/Decoder/Video", "Post process GstVdpVideoBuffers and output GstVdpOutputBuffers", "Carl-Anton Ingmarsson "); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&sink_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&src_template)); } /* initialize the vdpaumpegdecoder's class */ static void gst_vdp_vpp_class_init (GstVdpVideoPostProcessClass * klass) { GObjectClass *gobject_class; GstElementClass *gstelement_class; gobject_class = (GObjectClass *) klass; gstelement_class = (GstElementClass *) klass; gobject_class->get_property = gst_vdp_vpp_get_property; gobject_class->set_property = gst_vdp_vpp_set_property; gobject_class->finalize = gst_vdp_vpp_finalize; g_object_class_install_property (gobject_class, PROP_FORCE_ASPECT_RATIO, g_param_spec_boolean ("force-aspect-ratio", "Force aspect ratio", "When enabled, the plugin will only scale up the input surface to the" "maximum size where the aspect ratio can be preserved", FALSE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_DEINTERLACE_MODE, g_param_spec_enum ("mode", "Deinterlace mode", "Specifies if the element should deinterlace or not", GST_TYPE_VDP_DEINTERLACE_MODES, GST_VDP_DEINTERLACE_MODE_AUTO, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_DEINTERLACE_METHOD, g_param_spec_enum ("method", "Deinterlace method", "Specifies which deinterlace method to use", GST_TYPE_VDP_DEINTERLACE_METHODS, GST_VDP_DEINTERLACE_METHOD_BOB, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_NOISE_REDUCTION, g_param_spec_float ("noise-reduction", "Noise reduction", "The amount of noise reduction that should be done", 0.0, 1.0, 0.0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_SHARPENING, g_param_spec_float ("sharpening", "Sharpening", "The amount of sharpening or blurring to be applied", -1.0, 1.0, 0.0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); g_object_class_install_property (gobject_class, PROP_INVERSE_TELECINE, g_param_spec_boolean ("inverse-telecine", "Inverse telecine", "Whether inverse telecine should be used", FALSE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); gstelement_class->change_state = gst_vdp_vpp_change_state; } static void gst_vdp_vpp_init (GstVdpVideoPostProcess * vpp, GstVdpVideoPostProcessClass * gclass) { vpp->device = NULL; vpp->force_aspect_ratio = FALSE; vpp->mode = GST_VDP_DEINTERLACE_MODE_AUTO; vpp->method = GST_VDP_DEINTERLACE_METHOD_BOB; vpp->noise_reduction = 0.0; vpp->sharpening = 0.0; /* SRC PAD */ vpp->srcpad = gst_pad_new_from_static_template (&src_template, "src"); gst_element_add_pad (GST_ELEMENT (vpp), vpp->srcpad); /* SINK PAD */ vpp->sinkpad = gst_pad_new_from_static_template (&sink_template, "sink"); gst_element_add_pad (GST_ELEMENT (vpp), vpp->sinkpad); gst_pad_set_getcaps_function (vpp->sinkpad, gst_vdp_vpp_sink_getcaps); gst_pad_set_setcaps_function (vpp->sinkpad, gst_vdp_vpp_sink_setcaps); gst_pad_set_chain_function (vpp->sinkpad, GST_DEBUG_FUNCPTR (gst_vdp_vpp_chain)); gst_pad_set_event_function (vpp->sinkpad, GST_DEBUG_FUNCPTR (gst_vdp_vpp_sink_event)); gst_pad_set_bufferalloc_function (vpp->sinkpad, gst_vdp_vpp_sink_bufferalloc); } static void gst_vdp_vpp_finalize (GObject * object) { }