/* * gstvaapipostproc.c - VA-API video postprocessing * * Copyright (C) 2012-2013 Intel Corporation * Author: Gwenole Beauchesne * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public License * as published by the Free Software Foundation; either version 2.1 * of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free * Software Foundation, Inc., 51 Franklin Street, Fifth Floor, * Boston, MA 02110-1301 USA */ /** * SECTION:gstvaapipostproc * @short_description: A video postprocessing filter * * vaapipostproc consists in various postprocessing algorithms to be * applied to VA surfaces. So far, only basic bob deinterlacing is * implemented. */ #include "gst/vaapi/sysdeps.h" #include #include "gstvaapipostproc.h" #include "gstvaapipluginutil.h" #include "gstvaapivideocontext.h" #include "gstvaapivideobuffer.h" #if GST_CHECK_VERSION(1,0,0) #include "gstvaapivideobufferpool.h" #include "gstvaapivideomemory.h" #endif #define GST_PLUGIN_NAME "vaapipostproc" #define GST_PLUGIN_DESC "A video postprocessing filter" GST_DEBUG_CATEGORY_STATIC(gst_debug_vaapipostproc); #define GST_CAT_DEFAULT gst_debug_vaapipostproc /* Default templates */ static const char gst_vaapipostproc_sink_caps_str[] = #if GST_CHECK_VERSION(1,1,0) GST_VIDEO_CAPS_MAKE_WITH_FEATURES( GST_CAPS_FEATURE_MEMORY_VAAPI_SURFACE, "{ ENCODED, NV12, I420, YV12 }") ", " #else GST_VAAPI_SURFACE_CAPS ", " #endif GST_CAPS_INTERLACED_MODES "; " #if GST_CHECK_VERSION(1,0,0) GST_VIDEO_CAPS_MAKE(GST_VIDEO_FORMATS_ALL) ", " #else "video/x-raw-yuv, " "width = " GST_VIDEO_SIZE_RANGE ", " "height = " GST_VIDEO_SIZE_RANGE ", " "framerate = " GST_VIDEO_FPS_RANGE ", " #endif GST_CAPS_INTERLACED_MODES; static const char gst_vaapipostproc_src_caps_str[] = #if GST_CHECK_VERSION(1,1,0) GST_VIDEO_CAPS_MAKE_WITH_FEATURES( GST_CAPS_FEATURE_MEMORY_VAAPI_SURFACE, "{ ENCODED, NV12, I420, YV12 }") ", " #else GST_VAAPI_SURFACE_CAPS ", " #endif GST_CAPS_INTERLACED_FALSE; static GstStaticPadTemplate gst_vaapipostproc_sink_factory = GST_STATIC_PAD_TEMPLATE( "sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS(gst_vaapipostproc_sink_caps_str)); static GstStaticPadTemplate gst_vaapipostproc_src_factory = GST_STATIC_PAD_TEMPLATE( "src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS(gst_vaapipostproc_src_caps_str)); /* GstImplementsInterface interface */ #if !GST_CHECK_VERSION(1,0,0) static gboolean gst_vaapipostproc_implements_interface_supported( GstImplementsInterface *iface, GType type ) { return (type == GST_TYPE_VIDEO_CONTEXT); } static void gst_vaapipostproc_implements_iface_init(GstImplementsInterfaceClass *iface) { iface->supported = gst_vaapipostproc_implements_interface_supported; } #endif /* GstVideoContext interface */ #if !GST_CHECK_VERSION(1,1,0) static void gst_vaapipostproc_set_video_context( GstVideoContext *context, const gchar *type, const GValue *value ) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(context); gst_vaapi_set_display(type, value, &postproc->display); if (postproc->uploader) gst_vaapi_uploader_ensure_display(postproc->uploader, postproc->display); } static void gst_video_context_interface_init(GstVideoContextInterface *iface) { iface->set_context = gst_vaapipostproc_set_video_context; } #define GstVideoContextClass GstVideoContextInterface #endif G_DEFINE_TYPE_WITH_CODE( GstVaapiPostproc, gst_vaapipostproc, GST_TYPE_BASE_TRANSFORM, #if !GST_CHECK_VERSION(1,0,0) G_IMPLEMENT_INTERFACE(GST_TYPE_IMPLEMENTS_INTERFACE, gst_vaapipostproc_implements_iface_init); #endif #if !GST_CHECK_VERSION(1,1,0) G_IMPLEMENT_INTERFACE(GST_TYPE_VIDEO_CONTEXT, gst_video_context_interface_init) #endif ) enum { PROP_0, PROP_FORMAT, PROP_WIDTH, PROP_HEIGHT, PROP_FORCE_ASPECT_RATIO, PROP_DEINTERLACE_MODE, PROP_DEINTERLACE_METHOD, PROP_DENOISE, PROP_SHARPEN, }; #define DEFAULT_FORMAT GST_VIDEO_FORMAT_ENCODED #define DEFAULT_DEINTERLACE_MODE GST_VAAPI_DEINTERLACE_MODE_AUTO #define DEFAULT_DEINTERLACE_METHOD GST_VAAPI_DEINTERLACE_METHOD_BOB #define GST_VAAPI_TYPE_DEINTERLACE_MODE \ gst_vaapi_deinterlace_mode_get_type() static GType gst_vaapi_deinterlace_mode_get_type(void) { static GType deinterlace_mode_type = 0; static const GEnumValue mode_types[] = { { GST_VAAPI_DEINTERLACE_MODE_AUTO, "Auto detection", "auto" }, { GST_VAAPI_DEINTERLACE_MODE_INTERLACED, "Force deinterlacing", "interlaced" }, { GST_VAAPI_DEINTERLACE_MODE_DISABLED, "Never deinterlace", "disabled" }, { 0, NULL, NULL }, }; if (!deinterlace_mode_type) { deinterlace_mode_type = g_enum_register_static("GstVaapiDeinterlaceMode", mode_types); } return deinterlace_mode_type; } static void ds_reset(GstVaapiDeinterlaceState *ds) { guint i; for (i = 0; i < G_N_ELEMENTS(ds->buffers); i++) gst_buffer_replace(&ds->buffers[i], NULL); ds->buffers_index = 0; ds->num_surfaces = 0; ds->deint = FALSE; ds->tff = FALSE; } static void ds_add_buffer(GstVaapiDeinterlaceState *ds, GstBuffer *buf) { gst_buffer_replace(&ds->buffers[ds->buffers_index], buf); ds->buffers_index = (ds->buffers_index + 1) % G_N_ELEMENTS(ds->buffers); } static inline GstBuffer * ds_get_buffer(GstVaapiDeinterlaceState *ds, guint index) { /* Note: the index increases towards older buffers. i.e. buffer at index 0 means the immediately preceding buffer in the history, buffer at index 1 means the one preceding the surface at index 0, etc. */ const guint n = ds->buffers_index + G_N_ELEMENTS(ds->buffers) - index - 1; return ds->buffers[n % G_N_ELEMENTS(ds->buffers)]; } static void ds_set_surfaces(GstVaapiDeinterlaceState *ds) { GstVaapiVideoMeta *meta; guint i; ds->num_surfaces = 0; for (i = 0; i < G_N_ELEMENTS(ds->buffers); i++) { GstBuffer * const buf = ds_get_buffer(ds, i); if (!buf) break; meta = gst_buffer_get_vaapi_video_meta(buf); ds->surfaces[ds->num_surfaces++] = gst_vaapi_video_meta_get_surface(meta); } } static GstVaapiFilterOpInfo * find_filter_op(GPtrArray *filter_ops, GstVaapiFilterOp op) { guint i; if (filter_ops) { for (i = 0; i < filter_ops->len; i++) { GstVaapiFilterOpInfo * const filter_op = g_ptr_array_index(filter_ops, i); if (filter_op->op == op) return filter_op; } } return NULL; } #if GST_CHECK_VERSION(1,1,0) static void gst_vaapipostproc_set_context(GstElement *element, GstContext *context) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(element); GstVaapiDisplay *display = NULL; if (gst_vaapi_video_context_get_display(context, &display)) { GST_INFO_OBJECT(element, "set display %p", display); gst_vaapi_display_replace(&postproc->display, display); gst_vaapi_display_unref(display); } } #endif static inline gboolean gst_vaapipostproc_ensure_display(GstVaapiPostproc *postproc) { return gst_vaapi_ensure_display(postproc, GST_VAAPI_DISPLAY_TYPE_ANY, &postproc->display); } static gboolean gst_vaapipostproc_ensure_uploader(GstVaapiPostproc *postproc) { if (!gst_vaapipostproc_ensure_display(postproc)) return FALSE; if (!postproc->uploader) { postproc->uploader = gst_vaapi_uploader_new(postproc->display); if (!postproc->uploader) return FALSE; } if (!gst_vaapi_uploader_ensure_display(postproc->uploader, postproc->display)) return FALSE; return TRUE; } static gboolean gst_vaapipostproc_ensure_uploader_caps(GstVaapiPostproc *postproc) { #if !GST_CHECK_VERSION(1,0,0) if (postproc->is_raw_yuv && !gst_vaapi_uploader_ensure_caps( postproc->uploader, postproc->sinkpad_caps, NULL)) return FALSE; #endif return TRUE; } static gboolean gst_vaapipostproc_ensure_filter(GstVaapiPostproc *postproc) { if (postproc->filter) return TRUE; if (!gst_vaapipostproc_ensure_display(postproc)) return FALSE; postproc->filter = gst_vaapi_filter_new(postproc->display); if (!postproc->filter) return FALSE; return TRUE; } static gboolean gst_vaapipostproc_ensure_filter_caps(GstVaapiPostproc *postproc) { if (!gst_vaapipostproc_ensure_filter(postproc)) return FALSE; postproc->filter_ops = gst_vaapi_filter_get_operations(postproc->filter); if (!postproc->filter_ops) return FALSE; postproc->filter_formats = gst_vaapi_filter_get_formats(postproc->filter); if (!postproc->filter_formats) return FALSE; return TRUE; } static gboolean gst_vaapipostproc_create(GstVaapiPostproc *postproc) { if (!gst_vaapipostproc_ensure_display(postproc)) return FALSE; if (!gst_vaapipostproc_ensure_uploader(postproc)) return FALSE; if (!gst_vaapipostproc_ensure_uploader_caps(postproc)) return FALSE; if (gst_vaapipostproc_ensure_filter(postproc)) postproc->use_vpp = TRUE; return TRUE; } static void gst_vaapipostproc_destroy_filter(GstVaapiPostproc *postproc) { if (postproc->filter_formats) { g_array_unref(postproc->filter_formats); postproc->filter_formats = NULL; } if (postproc->filter_ops) { g_ptr_array_unref(postproc->filter_ops); postproc->filter_ops = NULL; } gst_vaapi_filter_replace(&postproc->filter, NULL); gst_vaapi_video_pool_replace(&postproc->filter_pool, NULL); } static void gst_vaapipostproc_destroy(GstVaapiPostproc *postproc) { ds_reset(&postproc->deinterlace_state); #if GST_CHECK_VERSION(1,0,0) g_clear_object(&postproc->sinkpad_buffer_pool); #endif g_clear_object(&postproc->uploader); gst_vaapipostproc_destroy_filter(postproc); gst_vaapi_display_replace(&postproc->display, NULL); gst_caps_replace(&postproc->allowed_sinkpad_caps, NULL); gst_caps_replace(&postproc->sinkpad_caps, NULL); gst_caps_replace(&postproc->allowed_srcpad_caps, NULL); gst_caps_replace(&postproc->srcpad_caps, NULL); } static gboolean gst_vaapipostproc_start(GstBaseTransform *trans) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); ds_reset(&postproc->deinterlace_state); if (!gst_vaapipostproc_ensure_display(postproc)) return FALSE; return TRUE; } static gboolean gst_vaapipostproc_stop(GstBaseTransform *trans) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); ds_reset(&postproc->deinterlace_state); gst_vaapi_display_replace(&postproc->display, NULL); return TRUE; } static gboolean is_interlaced_buffer(GstVaapiPostproc *postproc, GstBuffer *buf) { if (!(postproc->flags & GST_VAAPI_POSTPROC_FLAG_DEINTERLACE)) return FALSE; switch (GST_VIDEO_INFO_INTERLACE_MODE(&postproc->sinkpad_info)) { case GST_VIDEO_INTERLACE_MODE_MIXED: #if GST_CHECK_VERSION(1,0,0) if (!GST_BUFFER_FLAG_IS_SET(buf, GST_VIDEO_BUFFER_FLAG_INTERLACED)) return FALSE; #else if (GST_BUFFER_FLAG_IS_SET(buf, GST_VIDEO_BUFFER_PROGRESSIVE)) return FALSE; #endif break; default: break; } return TRUE; } static GstBuffer * create_output_buffer(GstVaapiPostproc *postproc) { GstBuffer *outbuf; /* Create a raw VA video buffer without GstVaapiVideoMeta attached to it yet, as this will be done next in the transform() hook */ outbuf = gst_vaapi_video_buffer_new_empty(); if (!outbuf) goto error_create_buffer; #if !GST_CHECK_VERSION(1,0,0) gst_buffer_set_caps(outbuf, postproc->srcpad_caps); #endif return outbuf; /* ERRORS */ error_create_buffer: { GST_ERROR("failed to create output video buffer"); return NULL; } } static inline void append_output_buffer_metadata(GstBuffer *outbuf, GstBuffer *inbuf, guint flags) { gst_buffer_copy_into(outbuf, inbuf, flags | GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_META | GST_BUFFER_COPY_MEMORY, 0, -1); } static gboolean deint_method_is_advanced(GstVaapiDeinterlaceMethod deint_method) { gboolean is_advanced; switch (deint_method) { case GST_VAAPI_DEINTERLACE_METHOD_MOTION_ADAPTIVE: case GST_VAAPI_DEINTERLACE_METHOD_MOTION_COMPENSATED: is_advanced = TRUE; break; default: is_advanced = FALSE; break; } return is_advanced; } static GstVaapiDeinterlaceMethod get_next_deint_method(GstVaapiDeinterlaceMethod deint_method) { switch (deint_method) { case GST_VAAPI_DEINTERLACE_METHOD_MOTION_COMPENSATED: deint_method = GST_VAAPI_DEINTERLACE_METHOD_MOTION_ADAPTIVE; break; default: /* Default to basic "bob" for all others */ deint_method = GST_VAAPI_DEINTERLACE_METHOD_BOB; break; } return deint_method; } static gboolean set_best_deint_method(GstVaapiPostproc *postproc, guint flags, GstVaapiDeinterlaceMethod *deint_method_ptr) { GstVaapiDeinterlaceMethod deint_method = postproc->deinterlace_method; gboolean success; for (;;) { success = gst_vaapi_filter_set_deinterlacing(postproc->filter, deint_method, flags); if (success || deint_method == GST_VAAPI_DEINTERLACE_METHOD_BOB) break; deint_method = get_next_deint_method(deint_method); } *deint_method_ptr = deint_method; return success; } static GstFlowReturn gst_vaapipostproc_process_vpp(GstBaseTransform *trans, GstBuffer *inbuf, GstBuffer *outbuf) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); GstVaapiDeinterlaceState * const ds = &postproc->deinterlace_state; GstVaapiVideoMeta *inbuf_meta, *outbuf_meta; GstVaapiSurface *inbuf_surface, *outbuf_surface; GstVaapiFilterStatus status; GstClockTime timestamp; GstFlowReturn ret; GstBuffer *fieldbuf; GstVaapiDeinterlaceMethod deint_method; guint flags, deint_flags; gboolean tff, deint, deint_refs, deint_changed; /* Validate filters */ if ((postproc->flags & GST_VAAPI_POSTPROC_FLAG_FORMAT) && !gst_vaapi_filter_set_format(postproc->filter, postproc->format)) return GST_FLOW_NOT_SUPPORTED; if ((postproc->flags & GST_VAAPI_POSTPROC_FLAG_DENOISE) && !gst_vaapi_filter_set_denoising_level(postproc->filter, postproc->denoise_level)) return GST_FLOW_NOT_SUPPORTED; if ((postproc->flags & GST_VAAPI_POSTPROC_FLAG_SHARPEN) && !gst_vaapi_filter_set_denoising_level(postproc->filter, postproc->sharpen_level)) return GST_FLOW_NOT_SUPPORTED; inbuf_meta = gst_buffer_get_vaapi_video_meta(inbuf); if (!inbuf_meta) goto error_invalid_buffer; inbuf_surface = gst_vaapi_video_meta_get_surface(inbuf_meta); timestamp = GST_BUFFER_TIMESTAMP(inbuf); tff = GST_BUFFER_FLAG_IS_SET(inbuf, GST_VIDEO_BUFFER_FLAG_TFF); deint = is_interlaced_buffer(postproc, inbuf); /* Drop references if deinterlacing conditions changed */ deint_changed = deint != ds->deint; if (deint_changed || (ds->num_surfaces > 0 && tff != ds->tff)) ds_reset(ds); ds->deint = deint; ds->tff = tff; flags = gst_vaapi_video_meta_get_render_flags(inbuf_meta) & ~(GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD| GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD); /* First field */ deint_method = postproc->deinterlace_method; deint_refs = deint_method_is_advanced(deint_method); if (postproc->flags & GST_VAAPI_POSTPROC_FLAG_DEINTERLACE) { fieldbuf = create_output_buffer(postproc); if (!fieldbuf) goto error_create_buffer; outbuf_meta = gst_vaapi_video_meta_new_from_pool(postproc->filter_pool); if (!outbuf_meta) goto error_create_meta; outbuf_surface = gst_vaapi_video_meta_get_surface(outbuf_meta); if (deint) { deint_flags = (tff ? GST_VAAPI_DEINTERLACE_FLAG_TOPFIELD : 0); if (tff) deint_flags |= GST_VAAPI_DEINTERLACE_FLAG_TFF; if (!set_best_deint_method(postproc, deint_flags, &deint_method)) goto error_op_deinterlace; if (deint_method != postproc->deinterlace_method) { GST_DEBUG("unsupported deinterlace-method %u. Using %u instead", postproc->deinterlace_method, deint_method); postproc->deinterlace_method = deint_method; deint_refs = deint_method_is_advanced(deint_method); } if (deint_refs) { ds_set_surfaces(ds); if (!gst_vaapi_filter_set_deinterlacing_references( postproc->filter, ds->surfaces, ds->num_surfaces, NULL, 0)) goto error_op_deinterlace; } } else if (deint_changed) { // Reset internal filter to non-deinterlacing mode deint_method = GST_VAAPI_DEINTERLACE_METHOD_NONE; if (!gst_vaapi_filter_set_deinterlacing(postproc->filter, deint_method, 0)) goto error_op_deinterlace; } status = gst_vaapi_filter_process(postproc->filter, inbuf_surface, outbuf_surface, flags); if (status != GST_VAAPI_FILTER_STATUS_SUCCESS) goto error_process_vpp; gst_buffer_set_vaapi_video_meta(fieldbuf, outbuf_meta); gst_vaapi_video_meta_unref(outbuf_meta); GST_BUFFER_TIMESTAMP(fieldbuf) = timestamp; GST_BUFFER_DURATION(fieldbuf) = postproc->field_duration; ret = gst_pad_push(trans->srcpad, fieldbuf); if (ret != GST_FLOW_OK) goto error_push_buffer; } fieldbuf = NULL; /* Second field */ outbuf_meta = gst_vaapi_video_meta_new_from_pool(postproc->filter_pool); if (!outbuf_meta) goto error_create_meta; outbuf_surface = gst_vaapi_video_meta_get_surface(outbuf_meta); if (deint) { deint_flags = (tff ? 0 : GST_VAAPI_DEINTERLACE_FLAG_TOPFIELD); if (tff) deint_flags |= GST_VAAPI_DEINTERLACE_FLAG_TFF; if (!gst_vaapi_filter_set_deinterlacing(postproc->filter, deint_method, deint_flags)) goto error_op_deinterlace; if (deint_refs && !gst_vaapi_filter_set_deinterlacing_references( postproc->filter, ds->surfaces, ds->num_surfaces, NULL, 0)) goto error_op_deinterlace; } else if (deint_changed && !gst_vaapi_filter_set_deinterlacing( postproc->filter, deint_method, 0)) goto error_op_deinterlace; status = gst_vaapi_filter_process(postproc->filter, inbuf_surface, outbuf_surface, flags); if (status != GST_VAAPI_FILTER_STATUS_SUCCESS) goto error_process_vpp; if (!(postproc->flags & GST_VAAPI_POSTPROC_FLAG_DEINTERLACE)) gst_buffer_copy_into(outbuf, inbuf, GST_BUFFER_COPY_TIMESTAMPS, 0, -1); else { GST_BUFFER_TIMESTAMP(outbuf) = timestamp + postproc->field_duration; GST_BUFFER_DURATION(outbuf) = postproc->field_duration; } gst_buffer_set_vaapi_video_meta(outbuf, outbuf_meta); gst_vaapi_video_meta_unref(outbuf_meta); if (deint && deint_refs) ds_add_buffer(ds, inbuf); return GST_FLOW_OK; /* ERRORS */ error_invalid_buffer: { GST_ERROR("failed to validate source buffer"); return GST_FLOW_ERROR; } error_create_buffer: { GST_ERROR("failed to create output buffer"); return GST_FLOW_ERROR; } error_create_meta: { GST_ERROR("failed to create new output buffer meta"); gst_buffer_replace(&fieldbuf, NULL); gst_vaapi_video_meta_unref(outbuf_meta); return GST_FLOW_ERROR; } error_op_deinterlace: { GST_ERROR("failed to apply deinterlacing filter"); gst_buffer_replace(&fieldbuf, NULL); gst_vaapi_video_meta_unref(outbuf_meta); return GST_FLOW_NOT_SUPPORTED; } error_process_vpp: { GST_ERROR("failed to apply VPP filters (error %d)", status); gst_buffer_replace(&fieldbuf, NULL); gst_vaapi_video_meta_unref(outbuf_meta); return GST_FLOW_ERROR; } error_push_buffer: { if (ret != GST_FLOW_FLUSHING) GST_ERROR("failed to push output buffer to video sink"); return GST_FLOW_ERROR; } } static GstFlowReturn gst_vaapipostproc_process(GstBaseTransform *trans, GstBuffer *inbuf, GstBuffer *outbuf) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); GstVaapiVideoMeta *meta; GstClockTime timestamp; GstFlowReturn ret; GstBuffer *fieldbuf; guint fieldbuf_flags, outbuf_flags, flags; gboolean tff, deint; meta = gst_buffer_get_vaapi_video_meta(inbuf); if (!meta) goto error_invalid_buffer; timestamp = GST_BUFFER_TIMESTAMP(inbuf); tff = GST_BUFFER_FLAG_IS_SET(inbuf, GST_VIDEO_BUFFER_FLAG_TFF); deint = is_interlaced_buffer(postproc, inbuf); flags = gst_vaapi_video_meta_get_render_flags(meta) & ~(GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD| GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD); /* First field */ fieldbuf = create_output_buffer(postproc); if (!fieldbuf) goto error_create_buffer; append_output_buffer_metadata(fieldbuf, inbuf, 0); meta = gst_buffer_get_vaapi_video_meta(fieldbuf); fieldbuf_flags = flags; fieldbuf_flags |= deint ? ( tff ? GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD : GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD) : GST_VAAPI_PICTURE_STRUCTURE_FRAME; gst_vaapi_video_meta_set_render_flags(meta, fieldbuf_flags); GST_BUFFER_TIMESTAMP(fieldbuf) = timestamp; GST_BUFFER_DURATION(fieldbuf) = postproc->field_duration; ret = gst_pad_push(trans->srcpad, fieldbuf); if (ret != GST_FLOW_OK) goto error_push_buffer; /* Second field */ append_output_buffer_metadata(outbuf, inbuf, 0); meta = gst_buffer_get_vaapi_video_meta(outbuf); outbuf_flags = flags; outbuf_flags |= deint ? ( tff ? GST_VAAPI_PICTURE_STRUCTURE_BOTTOM_FIELD : GST_VAAPI_PICTURE_STRUCTURE_TOP_FIELD) : GST_VAAPI_PICTURE_STRUCTURE_FRAME; gst_vaapi_video_meta_set_render_flags(meta, outbuf_flags); GST_BUFFER_TIMESTAMP(outbuf) = timestamp + postproc->field_duration; GST_BUFFER_DURATION(outbuf) = postproc->field_duration; return GST_FLOW_OK; /* ERRORS */ error_invalid_buffer: { GST_ERROR("failed to validate source buffer"); return GST_FLOW_ERROR; } error_create_buffer: { GST_ERROR("failed to create output buffer"); return GST_FLOW_EOS; } error_push_buffer: { if (ret != GST_FLOW_FLUSHING) GST_ERROR("failed to push output buffer to video sink"); return GST_FLOW_EOS; } } static GstFlowReturn gst_vaapipostproc_passthrough(GstBaseTransform *trans, GstBuffer *inbuf, GstBuffer *outbuf) { GstVaapiVideoMeta *meta; /* No video processing needed, simply copy buffer metadata */ meta = gst_buffer_get_vaapi_video_meta(inbuf); if (!meta) goto error_invalid_buffer; append_output_buffer_metadata(outbuf, inbuf, GST_BUFFER_COPY_TIMESTAMPS); return GST_FLOW_OK; /* ERRORS */ error_invalid_buffer: { GST_ERROR("failed to validate source buffer"); return GST_FLOW_ERROR; } } static gboolean is_deinterlace_enabled(GstVaapiPostproc *postproc, GstVideoInfo *vip) { gboolean deinterlace; switch (postproc->deinterlace_mode) { case GST_VAAPI_DEINTERLACE_MODE_AUTO: deinterlace = GST_VIDEO_INFO_IS_INTERLACED(vip); break; case GST_VAAPI_DEINTERLACE_MODE_INTERLACED: deinterlace = TRUE; break; default: deinterlace = FALSE; break; } return deinterlace; } static gboolean video_info_changed(GstVideoInfo *old_vip, GstVideoInfo *new_vip) { if (GST_VIDEO_INFO_FORMAT(old_vip) != GST_VIDEO_INFO_FORMAT(new_vip)) return TRUE; if (GST_VIDEO_INFO_INTERLACE_MODE(old_vip) != GST_VIDEO_INFO_INTERLACE_MODE(new_vip)) return TRUE; if (GST_VIDEO_INFO_WIDTH(old_vip) != GST_VIDEO_INFO_WIDTH(new_vip)) return TRUE; if (GST_VIDEO_INFO_HEIGHT(old_vip) != GST_VIDEO_INFO_HEIGHT(new_vip)) return TRUE; return FALSE; } static gboolean gst_vaapipostproc_update_sink_caps(GstVaapiPostproc *postproc, GstCaps *caps, gboolean *caps_changed_ptr) { GstVideoInfo vi; gboolean deinterlace; if (!gst_video_info_from_caps(&vi, caps)) return FALSE; if (video_info_changed(&vi, &postproc->sinkpad_info)) postproc->sinkpad_info = vi, *caps_changed_ptr = TRUE; deinterlace = is_deinterlace_enabled(postproc, &vi); if (deinterlace) postproc->flags |= GST_VAAPI_POSTPROC_FLAG_DEINTERLACE; postproc->field_duration = gst_util_uint64_scale( GST_SECOND, GST_VIDEO_INFO_FPS_D(&vi), (1 + deinterlace) * GST_VIDEO_INFO_FPS_N(&vi)); postproc->is_raw_yuv = GST_VIDEO_INFO_IS_YUV(&vi); #if !GST_CHECK_VERSION(1,0,0) if (postproc->is_raw_yuv) { /* Ensure the uploader is set up for upstream allocated buffers */ GstVaapiUploader * const uploader = postproc->uploader; if (!gst_vaapi_uploader_ensure_display(uploader, postproc->display)) return FALSE; if (!gst_vaapi_uploader_ensure_caps(uploader, caps, NULL)) return FALSE; } #endif return TRUE; } static gboolean gst_vaapipostproc_update_src_caps(GstVaapiPostproc *postproc, GstCaps *caps, gboolean *caps_changed_ptr) { GstVideoInfo vi; if (!gst_video_info_from_caps(&vi, caps)) return FALSE; if (video_info_changed(&vi, &postproc->srcpad_info)) postproc->srcpad_info = vi, *caps_changed_ptr = TRUE; if (postproc->format != GST_VIDEO_INFO_FORMAT(&postproc->sinkpad_info)) postproc->flags |= GST_VAAPI_POSTPROC_FLAG_FORMAT; if ((postproc->width || postproc->height) && postproc->width != GST_VIDEO_INFO_WIDTH(&postproc->sinkpad_info) && postproc->height != GST_VIDEO_INFO_HEIGHT(&postproc->sinkpad_info)) postproc->flags |= GST_VAAPI_POSTPROC_FLAG_SIZE; return TRUE; } static gboolean ensure_allowed_sinkpad_caps(GstVaapiPostproc *postproc) { GstCaps *out_caps, *yuv_caps; if (postproc->allowed_sinkpad_caps) return TRUE; /* Create VA caps */ #if GST_CHECK_VERSION(1,1,0) out_caps = gst_static_pad_template_get_caps( &gst_vaapipostproc_sink_factory); #else out_caps = gst_caps_from_string(GST_VAAPI_SURFACE_CAPS ", " GST_CAPS_INTERLACED_MODES); #endif if (!out_caps) { GST_ERROR("failed to create VA sink caps"); return FALSE; } /* Append YUV caps */ if (gst_vaapipostproc_ensure_uploader(postproc)) { yuv_caps = gst_vaapi_uploader_get_caps(postproc->uploader); if (yuv_caps) { out_caps = gst_caps_make_writable(out_caps); gst_caps_append(out_caps, gst_caps_copy(yuv_caps)); } else GST_WARNING("failed to create YUV sink caps"); } postproc->allowed_sinkpad_caps = out_caps; /* XXX: append VA/VPP filters */ return TRUE; } /* Build list of supported formats */ static gboolean build_format_list_value(GArray *formats, GValue *out_value) { GValue v_format = { 0, }; guint i; #if GST_CHECK_VERSION(1,0,0) const gchar *str; g_value_init(out_value, GST_TYPE_LIST); g_value_init(&v_format, G_TYPE_STRING); g_value_set_string(&v_format, "encoded"); gst_value_list_append_value(out_value, &v_format); for (i = 0; i < formats->len; i++) { GstVideoFormat const format = g_array_index(formats, GstVideoFormat, i); str = gst_vaapi_video_format_to_string(format); if (!str) continue; g_value_set_string(&v_format, str); gst_value_list_append_value(out_value, &v_format); } #else guint32 fourcc; g_value_init(out_value, GST_TYPE_LIST); g_value_init(&v_format, GST_TYPE_FOURCC); for (i = 0; i < formats->len; i++) { GstVideoFormat const format = g_array_index(formats, GstVideoFormat, i); fourcc = gst_video_format_to_fourcc(format); if (!fourcc) continue; gst_value_set_fourcc(&v_format, fourcc); gst_value_list_append_value(out_value, &v_format); } #endif g_value_unset(&v_format); return TRUE; } /* Fixup output caps so that to reflect the supported set of pixel formats */ static GstCaps * expand_allowed_srcpad_caps(GstVaapiPostproc *postproc, GstCaps *caps) { GValue value = { 0, }; guint i, num_structures; gboolean had_filter; had_filter = postproc->filter != NULL; if (!had_filter && !gst_vaapipostproc_ensure_filter(postproc)) goto cleanup; if (!gst_vaapipostproc_ensure_filter_caps(postproc)) goto cleanup; /* Reset "format" field for each structure */ if (!build_format_list_value(postproc->filter_formats, &value)) goto cleanup; num_structures = gst_caps_get_size(caps); for (i = 0; i < num_structures; i++) { GstStructure * const structure = gst_caps_get_structure(caps, i); if (!structure) continue; gst_structure_set_value(structure, "format", &value); } g_value_unset(&value); cleanup: if (!had_filter) gst_vaapipostproc_destroy_filter(postproc); return caps; } static gboolean ensure_allowed_srcpad_caps(GstVaapiPostproc *postproc) { GstCaps *out_caps; if (postproc->allowed_srcpad_caps) return TRUE; /* Create initial caps from pad template */ out_caps = gst_caps_from_string(gst_vaapipostproc_src_caps_str); if (!out_caps) { GST_ERROR("failed to create VA src caps"); return FALSE; } postproc->allowed_srcpad_caps = expand_allowed_srcpad_caps(postproc, out_caps); return postproc->allowed_srcpad_caps != NULL; } static void find_best_size(GstVaapiPostproc *postproc, GstVideoInfo *vip, guint *width_ptr, guint *height_ptr) { guint width, height; width = GST_VIDEO_INFO_WIDTH(vip); height = GST_VIDEO_INFO_HEIGHT(vip); if (postproc->width && postproc->height) { width = postproc->width; height = postproc->height; } else if (postproc->keep_aspect) { const gdouble ratio = (gdouble)width / height; if (postproc->width) { width = postproc->width; height = postproc->width / ratio; } else if (postproc->height) { height = postproc->height; width = postproc->height * ratio; } } else if (postproc->width) width = postproc->width; else if (postproc->height) height = postproc->height; *width_ptr = width; *height_ptr = height; } static GstCaps * gst_vaapipostproc_transform_caps_impl(GstBaseTransform *trans, GstPadDirection direction, GstCaps *caps) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); GstVideoInfo vi; GstVideoFormat format; GstCaps *out_caps; guint width, height; /* Generate the sink pad caps, that could be fixated afterwards */ if (direction == GST_PAD_SRC) { if (!ensure_allowed_sinkpad_caps(postproc)) return NULL; return gst_caps_ref(postproc->allowed_sinkpad_caps); } /* Generate complete set of src pad caps if non-fixated sink pad caps are provided */ if (!gst_caps_is_fixed(caps)) { if (!ensure_allowed_srcpad_caps(postproc)) return NULL; return gst_caps_ref(postproc->allowed_srcpad_caps); } /* Generate the expected src pad caps, from the current fixated sink pad caps */ if (!gst_video_info_from_caps(&vi, caps)) return NULL; // Set double framerate in interlaced mode if (is_deinterlace_enabled(postproc, &vi)) { gint fps_n = GST_VIDEO_INFO_FPS_N(&vi); gint fps_d = GST_VIDEO_INFO_FPS_D(&vi); if (!gst_util_fraction_multiply(fps_n, fps_d, 2, 1, &fps_n, &fps_d)) return NULL; GST_VIDEO_INFO_FPS_N(&vi) = fps_n; GST_VIDEO_INFO_FPS_D(&vi) = fps_d; } // Signal the other pad that we only generate progressive frames GST_VIDEO_INFO_INTERLACE_MODE(&vi) = GST_VIDEO_INTERLACE_MODE_PROGRESSIVE; // Update size from user-specified parameters #if GST_CHECK_VERSION(1,1,0) format = postproc->format; #else format = GST_VIDEO_FORMAT_ENCODED; #endif find_best_size(postproc, &vi, &width, &height); gst_video_info_set_format(&vi, format, width, height); #if GST_CHECK_VERSION(1,1,0) out_caps = gst_video_info_to_caps(&vi); if (!out_caps) return NULL; gst_caps_set_features(out_caps, 0, gst_caps_features_new(GST_CAPS_FEATURE_MEMORY_VAAPI_SURFACE, NULL)); #else /* XXX: gst_video_info_to_caps() from GStreamer 0.10 does not reconstruct suitable caps for "encoded" video formats */ out_caps = gst_caps_from_string(GST_VAAPI_SURFACE_CAPS_NAME); if (!out_caps) return NULL; gst_caps_set_simple(out_caps, "type", G_TYPE_STRING, "vaapi", "opengl", G_TYPE_BOOLEAN, USE_GLX, "width", G_TYPE_INT, GST_VIDEO_INFO_WIDTH(&vi), "height", G_TYPE_INT, GST_VIDEO_INFO_HEIGHT(&vi), "framerate", GST_TYPE_FRACTION, GST_VIDEO_INFO_FPS_N(&vi), GST_VIDEO_INFO_FPS_D(&vi), "pixel-aspect-ratio", GST_TYPE_FRACTION, GST_VIDEO_INFO_PAR_N(&vi), GST_VIDEO_INFO_PAR_D(&vi), NULL); gst_caps_set_interlaced(out_caps, &vi); #endif return out_caps; } #if GST_CHECK_VERSION(1,0,0) static GstCaps * gst_vaapipostproc_transform_caps(GstBaseTransform *trans, GstPadDirection direction, GstCaps *caps, GstCaps *filter) { GstCaps *out_caps; caps = gst_vaapipostproc_transform_caps_impl(trans, direction, caps); if (caps && filter) { out_caps = gst_caps_intersect_full(caps, filter, GST_CAPS_INTERSECT_FIRST); gst_caps_unref(caps); return out_caps; } return caps; } #else #define gst_vaapipostproc_transform_caps \ gst_vaapipostproc_transform_caps_impl #endif #if GST_CHECK_VERSION(1,0,0) typedef gsize GstBaseTransformSizeType; #else typedef guint GstBaseTransformSizeType; #endif static gboolean gst_vaapipostproc_transform_size(GstBaseTransform *trans, GstPadDirection direction, GstCaps *caps, GstBaseTransformSizeType size, GstCaps *othercaps, GstBaseTransformSizeType *othersize) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); if (direction == GST_PAD_SINK || !postproc->is_raw_yuv) *othersize = 0; else *othersize = size; return TRUE; } static GstBuffer * get_source_buffer(GstVaapiPostproc *postproc, GstBuffer *inbuf) { GstVaapiVideoMeta *meta; GstBuffer *outbuf; #if GST_CHECK_VERSION(1,0,0) GstVideoFrame src_frame, out_frame; #endif meta = gst_buffer_get_vaapi_video_meta(inbuf); if (meta) return gst_buffer_ref(inbuf); #if GST_CHECK_VERSION(1,0,0) if (!postproc->is_raw_yuv) goto error_invalid_buffer; if (!postproc->sinkpad_buffer_pool) goto error_no_pool; if (!gst_buffer_pool_set_active(postproc->sinkpad_buffer_pool, TRUE)) goto error_active_pool; outbuf = NULL; if (gst_buffer_pool_acquire_buffer(postproc->sinkpad_buffer_pool, &outbuf, NULL) != GST_FLOW_OK) goto error_create_buffer; if (!gst_video_frame_map(&src_frame, &postproc->sinkpad_info, inbuf, GST_MAP_READ)) goto error_map_src_buffer; if (!gst_video_frame_map(&out_frame, &postproc->sinkpad_info, outbuf, GST_MAP_WRITE)) goto error_map_dst_buffer; if (!gst_video_frame_copy(&out_frame, &src_frame)) goto error_copy_buffer; gst_video_frame_unmap(&out_frame); gst_video_frame_unmap(&src_frame); gst_buffer_copy_into(outbuf, inbuf, GST_BUFFER_COPY_TIMESTAMPS, 0, -1); return outbuf; /* ERRORS */ error_invalid_buffer: { GST_ERROR("failed to validate source buffer"); return NULL; } error_no_pool: { GST_ERROR("no buffer pool was negotiated"); return NULL; } error_active_pool: { GST_ERROR("failed to activate buffer pool"); return NULL; } error_map_dst_buffer: { gst_video_frame_unmap(&src_frame); // fall-through } error_map_src_buffer: { GST_ERROR("failed to map buffer"); gst_buffer_unref(outbuf); return NULL; } #else outbuf = gst_vaapi_uploader_get_buffer(postproc->uploader); if (!outbuf) goto error_create_buffer; if (!gst_vaapi_uploader_process(postproc->uploader, inbuf, outbuf)) goto error_copy_buffer; gst_buffer_copy_metadata(outbuf, inbuf, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS); return outbuf; #endif /* ERRORS */ error_create_buffer: { GST_ERROR("failed to create buffer"); return NULL; } error_copy_buffer: { GST_ERROR("failed to upload buffer to VA surface"); #if GST_CHECK_VERSION(1,0,0) gst_video_frame_unmap(&out_frame); gst_video_frame_unmap(&src_frame); #endif gst_buffer_unref(outbuf); return NULL; } } static GstFlowReturn gst_vaapipostproc_transform(GstBaseTransform *trans, GstBuffer *inbuf, GstBuffer *outbuf) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); GstBuffer *buf; GstFlowReturn ret; buf = get_source_buffer(postproc, inbuf); if (!buf) return GST_FLOW_ERROR; ret = GST_FLOW_NOT_SUPPORTED; if (postproc->flags) { /* Use VA/VPP extensions to process this frame */ if (postproc->use_vpp && postproc->flags != GST_VAAPI_POSTPROC_FLAG_DEINTERLACE) { ret = gst_vaapipostproc_process_vpp(trans, buf, outbuf); if (ret != GST_FLOW_NOT_SUPPORTED) goto done; GST_WARNING("unsupported VPP filters. Disabling"); postproc->use_vpp = FALSE; } /* Only append picture structure meta data (top/bottom field) */ if (postproc->flags & GST_VAAPI_POSTPROC_FLAG_DEINTERLACE) { ret = gst_vaapipostproc_process(trans, buf, outbuf); if (ret != GST_FLOW_NOT_SUPPORTED) goto done; } } /* Fallback: passthrough to the downstream element as is */ ret = gst_vaapipostproc_passthrough(trans, buf, outbuf); done: gst_buffer_unref(buf); return ret; } static GstFlowReturn gst_vaapipostproc_prepare_output_buffer(GstBaseTransform *trans, GstBuffer *inbuf, #if !GST_CHECK_VERSION(1,0,0) gint size, GstCaps *caps, #endif GstBuffer **outbuf_ptr) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); *outbuf_ptr = create_output_buffer(postproc); return *outbuf_ptr ? GST_FLOW_OK : GST_FLOW_ERROR; } static gboolean ensure_sinkpad_buffer_pool(GstVaapiPostproc *postproc, GstCaps *caps) { #if GST_CHECK_VERSION(1,0,0) GstBufferPool *pool; GstCaps *pool_caps; GstStructure *config; GstVideoInfo vi; gboolean need_pool; if (!gst_vaapipostproc_ensure_display(postproc)) return FALSE; if (postproc->sinkpad_buffer_pool) { config = gst_buffer_pool_get_config(postproc->sinkpad_buffer_pool); gst_buffer_pool_config_get_params(config, &pool_caps, NULL, NULL, NULL); need_pool = !gst_caps_is_equal(caps, pool_caps); gst_structure_free(config); if (!need_pool) return TRUE; g_clear_object(&postproc->sinkpad_buffer_pool); postproc->sinkpad_buffer_size = 0; } pool = gst_vaapi_video_buffer_pool_new(postproc->display); if (!pool) goto error_create_pool; gst_video_info_init(&vi); gst_video_info_from_caps(&vi, caps); if (GST_VIDEO_INFO_FORMAT(&vi) == GST_VIDEO_FORMAT_ENCODED) { GST_DEBUG("assume sink pad buffer pool format is NV12"); gst_video_info_set_format(&vi, GST_VIDEO_FORMAT_NV12, GST_VIDEO_INFO_WIDTH(&vi), GST_VIDEO_INFO_HEIGHT(&vi)); } postproc->sinkpad_buffer_size = vi.size; config = gst_buffer_pool_get_config(pool); gst_buffer_pool_config_set_params(config, caps, postproc->sinkpad_buffer_size, 0, 0); gst_buffer_pool_config_add_option(config, GST_BUFFER_POOL_OPTION_VAAPI_VIDEO_META); gst_buffer_pool_config_add_option(config, GST_BUFFER_POOL_OPTION_VIDEO_META); if (!gst_buffer_pool_set_config(pool, config)) goto error_pool_config; postproc->sinkpad_buffer_pool = pool; return TRUE; /* ERRORS */ error_create_pool: { GST_ERROR("failed to create buffer pool"); return FALSE; } error_pool_config: { GST_ERROR("failed to reset buffer pool config"); gst_object_unref(pool); return FALSE; } #else return TRUE; #endif } static gboolean ensure_srcpad_buffer_pool(GstVaapiPostproc *postproc, GstCaps *caps) { GstVideoInfo vi; GstVaapiVideoPool *pool; gst_video_info_init(&vi); gst_video_info_from_caps(&vi, caps); gst_video_info_set_format(&vi, postproc->format, GST_VIDEO_INFO_WIDTH(&vi), GST_VIDEO_INFO_HEIGHT(&vi)); if (!video_info_changed(&vi, &postproc->filter_pool_info)) return TRUE; postproc->filter_pool_info = vi; pool = gst_vaapi_surface_pool_new(postproc->display, &postproc->filter_pool_info); if (!pool) return FALSE; gst_vaapi_video_pool_replace(&postproc->filter_pool, pool); gst_vaapi_video_pool_unref(pool); return TRUE; } static gboolean gst_vaapipostproc_set_caps(GstBaseTransform *trans, GstCaps *caps, GstCaps *out_caps) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); gboolean caps_changed = FALSE; if (!gst_vaapipostproc_update_sink_caps(postproc, caps, &caps_changed)) return FALSE; if (!gst_vaapipostproc_update_src_caps(postproc, out_caps, &caps_changed)) return FALSE; if (caps_changed) { gst_vaapipostproc_destroy(postproc); gst_caps_replace(&postproc->sinkpad_caps, caps); gst_caps_replace(&postproc->srcpad_caps, out_caps); if (!gst_vaapipostproc_create(postproc)) return FALSE; } if (!ensure_sinkpad_buffer_pool(postproc, caps)) return FALSE; if (!ensure_srcpad_buffer_pool(postproc, out_caps)) return FALSE; return TRUE; } static gboolean gst_vaapipostproc_query(GstBaseTransform *trans, GstPadDirection direction, GstQuery *query) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); GST_INFO_OBJECT(trans, "query type `%s'", GST_QUERY_TYPE_NAME(query)); if (gst_vaapi_reply_to_query(query, postproc->display)) { GST_DEBUG("sharing display %p", postproc->display); return TRUE; } return GST_BASE_TRANSFORM_CLASS(gst_vaapipostproc_parent_class)->query( trans, direction, query); } #if GST_CHECK_VERSION(1,0,0) static gboolean gst_vaapipostproc_propose_allocation(GstBaseTransform *trans, GstQuery *decide_query, GstQuery *query) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(trans); GstCaps *caps = NULL; gboolean need_pool; /* Let vaapidecode allocate the video buffers */ if (!postproc->is_raw_yuv) return FALSE; gst_query_parse_allocation(query, &caps, &need_pool); if (need_pool) { if (!caps) goto error_no_caps; if (!ensure_sinkpad_buffer_pool(postproc, caps)) return FALSE; gst_query_add_allocation_pool(query, postproc->sinkpad_buffer_pool, postproc->sinkpad_buffer_size, 0, 0); } gst_query_add_allocation_meta(query, GST_VAAPI_VIDEO_META_API_TYPE, NULL); gst_query_add_allocation_meta(query, GST_VIDEO_META_API_TYPE, NULL); gst_query_add_allocation_meta(query, GST_VIDEO_CROP_META_API_TYPE, NULL); gst_query_add_allocation_meta(query, GST_VIDEO_OVERLAY_COMPOSITION_META_API_TYPE, NULL); return TRUE; /* ERRORS */ error_no_caps: { GST_ERROR("no caps specified"); return FALSE; } } #endif static void gst_vaapipostproc_finalize(GObject *object) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(object); gst_vaapipostproc_destroy(postproc); G_OBJECT_CLASS(gst_vaapipostproc_parent_class)->finalize(object); } static void gst_vaapipostproc_set_property( GObject *object, guint prop_id, const GValue *value, GParamSpec *pspec ) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(object); switch (prop_id) { case PROP_FORMAT: postproc->format = g_value_get_enum(value); break; case PROP_WIDTH: postproc->width = g_value_get_uint(value); break; case PROP_HEIGHT: postproc->height = g_value_get_uint(value); break; case PROP_FORCE_ASPECT_RATIO: postproc->keep_aspect = g_value_get_boolean(value); break; case PROP_DEINTERLACE_MODE: postproc->deinterlace_mode = g_value_get_enum(value); break; case PROP_DEINTERLACE_METHOD: postproc->deinterlace_method = g_value_get_enum(value); break; case PROP_DENOISE: postproc->denoise_level = g_value_get_float(value); postproc->flags |= GST_VAAPI_POSTPROC_FLAG_DENOISE; break; case PROP_SHARPEN: postproc->sharpen_level = g_value_get_float(value); postproc->flags |= GST_VAAPI_POSTPROC_FLAG_SHARPEN; break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID(object, prop_id, pspec); break; } } static void gst_vaapipostproc_get_property( GObject *object, guint prop_id, GValue *value, GParamSpec *pspec ) { GstVaapiPostproc * const postproc = GST_VAAPIPOSTPROC(object); switch (prop_id) { case PROP_FORMAT: g_value_set_enum(value, postproc->format); break; case PROP_WIDTH: g_value_set_uint(value, postproc->width); break; case PROP_HEIGHT: g_value_set_uint(value, postproc->height); break; case PROP_FORCE_ASPECT_RATIO: g_value_set_boolean(value, postproc->keep_aspect); break; case PROP_DEINTERLACE_MODE: g_value_set_enum(value, postproc->deinterlace_mode); break; case PROP_DEINTERLACE_METHOD: g_value_set_enum(value, postproc->deinterlace_method); break; case PROP_DENOISE: g_value_set_float(value, postproc->denoise_level); break; case PROP_SHARPEN: g_value_set_float(value, postproc->sharpen_level); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID(object, prop_id, pspec); break; } } static void gst_vaapipostproc_class_init(GstVaapiPostprocClass *klass) { GObjectClass * const object_class = G_OBJECT_CLASS(klass); GstElementClass * const element_class = GST_ELEMENT_CLASS(klass); GstBaseTransformClass * const trans_class = GST_BASE_TRANSFORM_CLASS(klass); GstPadTemplate *pad_template; GPtrArray *filter_ops; GstVaapiFilterOpInfo *filter_op; GST_DEBUG_CATEGORY_INIT(gst_debug_vaapipostproc, GST_PLUGIN_NAME, 0, GST_PLUGIN_DESC); object_class->finalize = gst_vaapipostproc_finalize; object_class->set_property = gst_vaapipostproc_set_property; object_class->get_property = gst_vaapipostproc_get_property; trans_class->start = gst_vaapipostproc_start; trans_class->stop = gst_vaapipostproc_stop; trans_class->transform_caps = gst_vaapipostproc_transform_caps; trans_class->transform_size = gst_vaapipostproc_transform_size; trans_class->transform = gst_vaapipostproc_transform; trans_class->set_caps = gst_vaapipostproc_set_caps; trans_class->query = gst_vaapipostproc_query; #if GST_CHECK_VERSION(1,0,0) trans_class->propose_allocation = gst_vaapipostproc_propose_allocation; #endif trans_class->prepare_output_buffer = gst_vaapipostproc_prepare_output_buffer; #if GST_CHECK_VERSION(1,1,0) element_class->set_context = gst_vaapipostproc_set_context; #endif gst_element_class_set_static_metadata(element_class, "VA-API video postprocessing", "Filter/Converter/Video", GST_PLUGIN_DESC, "Gwenole Beauchesne "); /* sink pad */ pad_template = gst_static_pad_template_get(&gst_vaapipostproc_sink_factory); gst_element_class_add_pad_template(element_class, pad_template); /* src pad */ pad_template = gst_static_pad_template_get(&gst_vaapipostproc_src_factory); gst_element_class_add_pad_template(element_class, pad_template); /** * GstVaapiPostproc:deinterlace-mode: * * This selects whether the deinterlacing should always be applied or if * they should only be applied on content that has the "interlaced" flag * on the caps. */ g_object_class_install_property (object_class, PROP_DEINTERLACE_MODE, g_param_spec_enum("deinterlace-mode", "Deinterlace mode", "Deinterlace mode to use", GST_VAAPI_TYPE_DEINTERLACE_MODE, DEFAULT_DEINTERLACE_MODE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVaapiPostproc:deinterlace-method: * * This selects the deinterlacing method to apply. */ g_object_class_install_property (object_class, PROP_DEINTERLACE_METHOD, g_param_spec_enum("deinterlace-method", "Deinterlace method", "Deinterlace method to use", GST_VAAPI_TYPE_DEINTERLACE_METHOD, DEFAULT_DEINTERLACE_METHOD, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); filter_ops = gst_vaapi_filter_get_operations(NULL); if (!filter_ops) return; /** * GstVaapiPostproc:format: * * The forced output pixel format, expressed as a #GstVideoFormat. */ filter_op = find_filter_op(filter_ops, GST_VAAPI_FILTER_OP_FORMAT); if (filter_op) g_object_class_install_property(object_class, PROP_FORMAT, filter_op->pspec); /** * GstVaapiPostproc:width: * * The forced output width in pixels. If set to zero, the width is * calculated from the height if aspect ration is preserved, or * inherited from the sink caps width */ g_object_class_install_property (object_class, PROP_WIDTH, g_param_spec_uint("width", "Width", "Forced output width", 0, G_MAXINT, 0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVaapiPostproc:height: * * The forced output height in pixels. If set to zero, the height * is calculated from the width if aspect ration is preserved, or * inherited from the sink caps height */ g_object_class_install_property (object_class, PROP_HEIGHT, g_param_spec_uint("height", "Height", "Forced output height", 0, G_MAXINT, 0, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVaapiPostproc:force-aspect-ratio: * * When enabled, scaling respects video aspect ratio; when * disabled, the video is distorted to fit the width and height * properties. */ g_object_class_install_property (object_class, PROP_FORCE_ASPECT_RATIO, g_param_spec_boolean("force-aspect-ratio", "Force aspect ratio", "When enabled, scaling will respect original aspect ratio", TRUE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); /** * GstVaapiPostproc:denoise: * * The level of noise reduction to apply. */ filter_op = find_filter_op(filter_ops, GST_VAAPI_FILTER_OP_DENOISE); if (filter_op) g_object_class_install_property(object_class, PROP_DENOISE, filter_op->pspec); /** * GstVaapiPostproc:sharpen: * * The level of sharpening to apply for positive values, or the * level of blurring for negative values. */ filter_op = find_filter_op(filter_ops, GST_VAAPI_FILTER_OP_SHARPEN); if (filter_op) g_object_class_install_property(object_class, PROP_SHARPEN, filter_op->pspec); g_ptr_array_unref(filter_ops); } static void gst_vaapipostproc_init(GstVaapiPostproc *postproc) { postproc->format = DEFAULT_FORMAT; postproc->deinterlace_mode = DEFAULT_DEINTERLACE_MODE; postproc->deinterlace_method = DEFAULT_DEINTERLACE_METHOD; postproc->field_duration = GST_CLOCK_TIME_NONE; postproc->keep_aspect = TRUE; gst_video_info_init(&postproc->sinkpad_info); gst_video_info_init(&postproc->srcpad_info); gst_video_info_init(&postproc->filter_pool_info); }