/* GStreamer * Copyright (C) 2022 Intel Corporation * Author: He Junyan * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the0 * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ #include "gstvabaseenc.h" #include #include #include "vacompat.h" #include "gstvacaps.h" #define GST_CAT_DEFAULT gst_va_base_enc_debug GST_DEBUG_CATEGORY_STATIC (GST_CAT_DEFAULT); struct _GstVaBaseEncPrivate { GstVideoInfo sinkpad_info; GstBufferPool *raw_pool; }; enum { PROP_DEVICE_PATH = 1, N_PROPERTIES }; static GParamSpec *properties[N_PROPERTIES]; /** * GstVaBaseEnc: * * A base class implementation for VA-API Encoders. * * Since: 1.22 */ /* *INDENT-OFF* */ #define gst_va_base_enc_parent_class parent_class G_DEFINE_ABSTRACT_TYPE_WITH_CODE (GstVaBaseEnc, gst_va_base_enc, GST_TYPE_VIDEO_ENCODER, G_ADD_PRIVATE (GstVaBaseEnc) GST_DEBUG_CATEGORY_INIT (gst_va_base_enc_debug, "vabaseenc", 0, "vabaseenc element");); /* *INDENT-ON* */ static void gst_va_base_enc_reset_state_default (GstVaBaseEnc * base) { base->frame_duration = GST_CLOCK_TIME_NONE; base->width = 0; base->height = 0; base->profile = VAProfileNone; base->rt_format = 0; base->codedbuf_size = 0; g_atomic_int_set (&base->reconf, FALSE); } static void _flush_all_frames (GstVaBaseEnc * base) { g_queue_clear_full (&base->reorder_list, (GDestroyNotify) gst_video_codec_frame_unref); g_queue_clear_full (&base->output_list, (GDestroyNotify) gst_video_codec_frame_unref); g_queue_clear_full (&base->ref_list, (GDestroyNotify) gst_video_codec_frame_unref); } static gboolean gst_va_base_enc_open (GstVideoEncoder * venc) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); GstVaBaseEncClass *klass = GST_VA_BASE_ENC_GET_CLASS (venc); gboolean ret = FALSE; if (!gst_va_ensure_element_data (venc, klass->render_device_path, &base->display)) return FALSE; g_object_notify (G_OBJECT (base), "device-path"); if (!g_atomic_pointer_get (&base->encoder)) { GstVaEncoder *va_encoder; va_encoder = gst_va_encoder_new (base->display, klass->codec, klass->entrypoint); if (va_encoder) ret = TRUE; gst_object_replace ((GstObject **) (&base->encoder), (GstObject *) va_encoder); gst_clear_object (&va_encoder); } else { ret = TRUE; } return ret; } static gboolean gst_va_base_enc_start (GstVideoEncoder * venc) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); gst_va_base_enc_reset_state (base); base->input_frame_count = 0; base->output_frame_count = 0; base->input_state = NULL; /* Set the minimum pts to some huge value (1000 hours). This keeps * the dts at the start of the stream from needing to be * negative. */ base->start_pts = GST_SECOND * 60 * 60 * 1000; gst_video_encoder_set_min_pts (venc, base->start_pts); return TRUE; } static gboolean gst_va_base_enc_close (GstVideoEncoder * venc) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); gst_clear_object (&base->encoder); gst_clear_object (&base->display); return TRUE; } static gboolean gst_va_base_enc_stop (GstVideoEncoder * venc) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); _flush_all_frames (base); if (!gst_va_encoder_close (base->encoder)) { GST_ERROR_OBJECT (base, "Failed to close the VA encoder"); return FALSE; } if (base->priv->raw_pool) gst_buffer_pool_set_active (base->priv->raw_pool, FALSE); gst_clear_object (&base->priv->raw_pool); if (base->input_state) gst_video_codec_state_unref (base->input_state); return TRUE; } static GstCaps * gst_va_base_enc_get_caps (GstVideoEncoder * venc, GstCaps * filter) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); GstCaps *caps = NULL, *tmp; if (base->encoder) caps = gst_va_encoder_get_sinkpad_caps (base->encoder); if (caps) { if (filter) { tmp = gst_caps_intersect_full (filter, caps, GST_CAPS_INTERSECT_FIRST); gst_caps_unref (caps); caps = tmp; } } else { caps = gst_video_encoder_proxy_getcaps (venc, NULL, filter); } GST_LOG_OBJECT (base, "Returning caps %" GST_PTR_FORMAT, caps); return caps; } static GstBufferPool * _get_sinkpad_pool (GstVaBaseEnc * base) { GstAllocator *allocator; GstAllocationParams params = { 0, }; guint size, usage_hint = 0; GArray *surface_formats = NULL; GstCaps *caps = NULL; if (base->priv->raw_pool) return base->priv->raw_pool; g_assert (base->input_state); caps = gst_caps_copy (base->input_state->caps); gst_caps_set_features_simple (caps, gst_caps_features_from_string (GST_CAPS_FEATURE_MEMORY_VA)); gst_allocation_params_init (¶ms); size = GST_VIDEO_INFO_SIZE (&base->input_state->info); surface_formats = gst_va_encoder_get_surface_formats (base->encoder); allocator = gst_va_allocator_new (base->display, surface_formats); base->priv->raw_pool = gst_va_pool_new_with_config (caps, size, 1, 0, usage_hint, GST_VA_FEATURE_AUTO, allocator, ¶ms); gst_clear_caps (&caps); if (!base->priv->raw_pool) { gst_object_unref (allocator); return NULL; } gst_va_allocator_get_format (allocator, &base->priv->sinkpad_info, NULL, NULL); gst_object_unref (allocator); if (!gst_buffer_pool_set_active (base->priv->raw_pool, TRUE)) { GST_WARNING_OBJECT (base, "Failed to activate sinkpad pool"); return NULL; } return base->priv->raw_pool; } static gboolean _try_import_buffer (GstVaBaseEnc * base, GstBuffer * inbuf) { VASurfaceID surface; /* The VA buffer. */ surface = gst_va_buffer_get_surface (inbuf); if (surface != VA_INVALID_ID && (gst_va_buffer_peek_display (inbuf) == base->display)) return TRUE; /* TODO: DMA buffer. */ return FALSE; } static GstFlowReturn gst_va_base_enc_import_input_buffer (GstVaBaseEnc * base, GstBuffer * inbuf, GstBuffer ** buf) { GstBuffer *buffer = NULL; GstBufferPool *pool; GstFlowReturn ret; GstVideoFrame in_frame, out_frame; gboolean imported, copied; imported = _try_import_buffer (base, inbuf); if (imported) { *buf = gst_buffer_ref (inbuf); return GST_FLOW_OK; } /* input buffer doesn't come from a vapool, thus it is required to * have a pool, grab from it a new buffer and copy the input * buffer to the new one */ if (!(pool = _get_sinkpad_pool (base))) return GST_FLOW_ERROR; ret = gst_buffer_pool_acquire_buffer (pool, &buffer, NULL); if (ret != GST_FLOW_OK) return ret; GST_LOG_OBJECT (base, "copying input frame"); if (!gst_video_frame_map (&in_frame, &base->input_state->info, inbuf, GST_MAP_READ)) goto invalid_buffer; if (!gst_video_frame_map (&out_frame, &base->priv->sinkpad_info, buffer, GST_MAP_WRITE)) { gst_video_frame_unmap (&in_frame); goto invalid_buffer; } copied = gst_video_frame_copy (&out_frame, &in_frame); gst_video_frame_unmap (&out_frame); gst_video_frame_unmap (&in_frame); if (!copied) goto invalid_buffer; /* strictly speaking this is not needed but let's play safe */ if (!gst_buffer_copy_into (buffer, inbuf, GST_BUFFER_COPY_FLAGS | GST_BUFFER_COPY_TIMESTAMPS, 0, -1)) return GST_FLOW_ERROR; *buf = buffer; return GST_FLOW_OK; invalid_buffer: { GST_ELEMENT_WARNING (base, CORE, NOT_IMPLEMENTED, (NULL), ("invalid video buffer received")); if (buffer) gst_buffer_unref (buffer); return GST_FLOW_ERROR; } } static GstBuffer * gst_va_base_enc_create_output_buffer (GstVaBaseEnc * base, GstVaEncodePicture * picture) { guint coded_size; goffset offset; GstBuffer *buf = NULL; VASurfaceID surface; VACodedBufferSegment *seg, *seg_list; /* Wait for encoding to finish */ surface = gst_va_encode_picture_get_raw_surface (picture); if (!va_sync_surface (base->display, surface)) goto error; seg_list = NULL; if (!va_map_buffer (base->display, picture->coded_buffer, (gpointer *) & seg_list)) goto error; if (!seg_list) { va_unmap_buffer (base->display, picture->coded_buffer); GST_WARNING_OBJECT (base, "coded buffer has no segment list"); goto error; } coded_size = 0; for (seg = seg_list; seg; seg = seg->next) coded_size += seg->size; buf = gst_video_encoder_allocate_output_buffer (GST_VIDEO_ENCODER_CAST (base), coded_size); if (!buf) { va_unmap_buffer (base->display, picture->coded_buffer); GST_ERROR_OBJECT (base, "Failed to allocate output buffer, size %d", coded_size); goto error; } offset = 0; for (seg = seg_list; seg; seg = seg->next) { gsize write_size; write_size = gst_buffer_fill (buf, offset, seg->buf, seg->size); if (write_size != seg->size) { GST_WARNING_OBJECT (base, "Segment size is %d, but copied %" G_GSIZE_FORMAT, seg->size, write_size); break; } offset += seg->size; } va_unmap_buffer (base->display, picture->coded_buffer); return buf; error: gst_clear_buffer (&buf); return NULL; } static GstAllocator * _allocator_from_caps (GstVaBaseEnc * base, GstCaps * caps) { GstAllocator *allocator = NULL; if (gst_caps_is_dmabuf (caps)) { allocator = gst_va_dmabuf_allocator_new (base->display); } else { GArray *surface_formats = gst_va_encoder_get_surface_formats (base->encoder); allocator = gst_va_allocator_new (base->display, surface_formats); } return allocator; } static gboolean gst_va_base_enc_propose_allocation (GstVideoEncoder * venc, GstQuery * query) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); GstAllocator *allocator = NULL; GstAllocationParams params = { 0, }; GstBufferPool *pool; GstCaps *caps; GstVideoInfo info; gboolean need_pool = FALSE; guint size, usage_hint = 0; gst_query_parse_allocation (query, &caps, &need_pool); if (!caps) return FALSE; if (!gst_video_info_from_caps (&info, caps)) { GST_ERROR_OBJECT (base, "Cannot parse caps %" GST_PTR_FORMAT, caps); return FALSE; } size = GST_VIDEO_INFO_SIZE (&info); gst_allocation_params_init (¶ms); if (!(allocator = _allocator_from_caps (base, caps))) return FALSE; pool = gst_va_pool_new_with_config (caps, size, 1, 0, usage_hint, GST_VA_FEATURE_AUTO, allocator, ¶ms); if (!pool) { gst_object_unref (allocator); goto config_failed; } gst_query_add_allocation_param (query, allocator, ¶ms); gst_query_add_allocation_pool (query, pool, size, 0, 0); GST_DEBUG_OBJECT (base, "proposing %" GST_PTR_FORMAT " with allocator %" GST_PTR_FORMAT, pool, allocator); gst_object_unref (allocator); gst_object_unref (pool); gst_query_add_allocation_meta (query, GST_VIDEO_META_API_TYPE, NULL); return TRUE; /* ERRORS */ config_failed: { GST_ERROR_OBJECT (base, "failed to set config"); return FALSE; } } static GstFlowReturn _push_buffer_to_downstream (GstVaBaseEnc * base, GstVideoCodecFrame * frame) { GstVaEncodePicture *enc_picture; GstVaBaseEncClass *base_class = GST_VA_BASE_ENC_GET_CLASS (base); GstBuffer *buf; if (base_class->prepare_output) base_class->prepare_output (base, frame); enc_picture = *((GstVaEncodePicture **) gst_video_codec_frame_get_user_data (frame)); buf = gst_va_base_enc_create_output_buffer (base, enc_picture); if (!buf) { GST_ERROR_OBJECT (base, "Failed to create output buffer"); goto error; } gst_buffer_replace (&frame->output_buffer, buf); gst_clear_buffer (&buf); GST_LOG_OBJECT (base, "Push to downstream: frame system_frame_number: %d," " pts: %" GST_TIME_FORMAT ", dts: %" GST_TIME_FORMAT " duration: %" GST_TIME_FORMAT ", buffer size: %" G_GSIZE_FORMAT, frame->system_frame_number, GST_TIME_ARGS (frame->pts), GST_TIME_ARGS (frame->dts), GST_TIME_ARGS (frame->duration), gst_buffer_get_size (frame->output_buffer)); return gst_video_encoder_finish_frame (GST_VIDEO_ENCODER (base), frame); error: gst_clear_buffer (&frame->output_buffer); gst_clear_buffer (&buf); gst_video_encoder_finish_frame (GST_VIDEO_ENCODER (base), frame); return GST_FLOW_ERROR; } static GstFlowReturn _push_out_one_buffer (GstVaBaseEnc * base) { GstVideoCodecFrame *frame_out; GstFlowReturn ret; guint32 system_frame_number; frame_out = g_queue_pop_head (&base->output_list); gst_video_codec_frame_unref (frame_out); system_frame_number = frame_out->system_frame_number; ret = _push_buffer_to_downstream (base, frame_out); if (ret != GST_FLOW_OK) { GST_DEBUG_OBJECT (base, "fails to push one buffer, system_frame_number " "%d: %s", system_frame_number, gst_flow_get_name (ret)); } return ret; } static GstFlowReturn gst_va_base_enc_drain (GstVideoEncoder * venc) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); GstVaBaseEncClass *base_class = GST_VA_BASE_ENC_GET_CLASS (base); GstFlowReturn ret = GST_FLOW_OK; GstVideoCodecFrame *frame_enc = NULL; gboolean is_last = FALSE; GST_DEBUG_OBJECT (base, "Encoder is draining"); /* Kickout all cached frames */ if (!base_class->reorder_frame (base, NULL, TRUE, &frame_enc)) { ret = GST_FLOW_ERROR; goto error_and_purge_all; } while (frame_enc) { is_last = FALSE; if (g_queue_is_empty (&base->reorder_list)) is_last = TRUE; ret = base_class->encode_frame (base, frame_enc, is_last); if (ret != GST_FLOW_OK) goto error_and_purge_all; frame_enc = NULL; ret = _push_out_one_buffer (base); if (ret != GST_FLOW_OK) goto error_and_purge_all; if (!base_class->reorder_frame (base, NULL, TRUE, &frame_enc)) { ret = GST_FLOW_ERROR; goto error_and_purge_all; } } g_assert (g_queue_is_empty (&base->reorder_list)); /* Output all frames. */ while (!g_queue_is_empty (&base->output_list)) { ret = _push_out_one_buffer (base); if (ret != GST_FLOW_OK) goto error_and_purge_all; } /* Also clear the reference list. */ g_queue_clear_full (&base->ref_list, (GDestroyNotify) gst_video_codec_frame_unref); return GST_FLOW_OK; error_and_purge_all: if (frame_enc) { gst_clear_buffer (&frame_enc->output_buffer); gst_video_encoder_finish_frame (venc, frame_enc); } if (!g_queue_is_empty (&base->output_list)) { GST_WARNING_OBJECT (base, "Still %d frame in the output list" " after drain", g_queue_get_length (&base->output_list)); while (!g_queue_is_empty (&base->output_list)) { frame_enc = g_queue_pop_head (&base->output_list); gst_video_codec_frame_unref (frame_enc); gst_clear_buffer (&frame_enc->output_buffer); gst_video_encoder_finish_frame (venc, frame_enc); } } if (!g_queue_is_empty (&base->reorder_list)) { GST_WARNING_OBJECT (base, "Still %d frame in the reorder list" " after drain", g_queue_get_length (&base->reorder_list)); while (!g_queue_is_empty (&base->reorder_list)) { frame_enc = g_queue_pop_head (&base->reorder_list); gst_video_codec_frame_unref (frame_enc); gst_clear_buffer (&frame_enc->output_buffer); gst_video_encoder_finish_frame (venc, frame_enc); } } /* Also clear the reference list. */ g_queue_clear_full (&base->ref_list, (GDestroyNotify) gst_video_codec_frame_unref); return ret; } static gboolean gst_va_base_enc_reset (GstVaBaseEnc * base) { GstVaBaseEncClass *base_class = GST_VA_BASE_ENC_GET_CLASS (base); GST_DEBUG_OBJECT (base, "Reconfiguration"); if (gst_va_base_enc_drain (GST_VIDEO_ENCODER (base)) != GST_FLOW_OK) return FALSE; if (!base_class->reconfig (base)) { GST_ERROR_OBJECT (base, "Error at reconfiguration error"); return FALSE; } return TRUE; } static GstFlowReturn gst_va_base_enc_handle_frame (GstVideoEncoder * venc, GstVideoCodecFrame * frame) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); GstVaBaseEncClass *base_class = GST_VA_BASE_ENC_GET_CLASS (base); GstFlowReturn ret; GstBuffer *in_buf = NULL; GstVideoCodecFrame *frame_encode = NULL; GST_LOG_OBJECT (venc, "handle frame id %d, dts %" GST_TIME_FORMAT ", pts %" GST_TIME_FORMAT, frame->system_frame_number, GST_TIME_ARGS (GST_BUFFER_DTS (frame->input_buffer)), GST_TIME_ARGS (GST_BUFFER_PTS (frame->input_buffer))); if (g_atomic_int_compare_and_exchange (&base->reconf, TRUE, FALSE)) { if (!gst_va_base_enc_reset (base)) return GST_FLOW_ERROR; } ret = gst_va_base_enc_import_input_buffer (base, frame->input_buffer, &in_buf); if (ret != GST_FLOW_OK) goto error_buffer_invalid; gst_buffer_replace (&frame->input_buffer, in_buf); gst_clear_buffer (&in_buf); if (!base_class->new_frame (base, frame)) goto error_new_frame; if (!base_class->reorder_frame (base, frame, FALSE, &frame_encode)) goto error_reorder; /* pass it to reorder list and we should not use it again. */ frame = NULL; while (frame_encode) { ret = base_class->encode_frame (base, frame_encode, FALSE); if (ret != GST_FLOW_OK) goto error_encode; while (g_queue_get_length (&base->output_list) > 0) ret = _push_out_one_buffer (base); frame_encode = NULL; if (!base_class->reorder_frame (base, NULL, FALSE, &frame_encode)) goto error_reorder; } return ret; error_buffer_invalid: { GST_ELEMENT_ERROR (venc, STREAM, ENCODE, ("Failed to import the input frame: %s.", gst_flow_get_name (ret)), (NULL)); gst_clear_buffer (&in_buf); gst_clear_buffer (&frame->output_buffer); gst_video_encoder_finish_frame (venc, frame); return ret; } error_new_frame: { GST_ELEMENT_ERROR (venc, STREAM, ENCODE, ("Failed to create the input frame."), (NULL)); gst_clear_buffer (&frame->output_buffer); gst_video_encoder_finish_frame (venc, frame); return GST_FLOW_ERROR; } error_reorder: { GST_ELEMENT_ERROR (venc, STREAM, ENCODE, ("Failed to reorder the input frame."), (NULL)); if (frame) { gst_clear_buffer (&frame->output_buffer); gst_video_encoder_finish_frame (venc, frame); } return GST_FLOW_ERROR; } error_encode: { GST_ELEMENT_ERROR (venc, STREAM, ENCODE, ("Failed to encode the frame %s.", gst_flow_get_name (ret)), (NULL)); gst_clear_buffer (&frame_encode->output_buffer); gst_video_encoder_finish_frame (venc, frame_encode); return ret; } } static GstFlowReturn gst_va_base_enc_finish (GstVideoEncoder * venc) { return gst_va_base_enc_drain (venc); } static gboolean gst_va_base_enc_set_format (GstVideoEncoder * venc, GstVideoCodecState * state) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); g_return_val_if_fail (state->caps != NULL, FALSE); if (base->input_state) gst_video_codec_state_unref (base->input_state); base->input_state = gst_video_codec_state_ref (state); if (!gst_va_base_enc_reset (base)) return FALSE; /* Sub class should open the encoder if reconfig succeeds. */ return gst_va_encoder_is_open (base->encoder); } static gboolean gst_va_base_enc_flush (GstVideoEncoder * venc) { _flush_all_frames (GST_VA_BASE_ENC (venc)); return TRUE; } static gboolean _query_context (GstVaBaseEnc * base, GstQuery * query) { GstVaDisplay *display = NULL; gboolean ret; gst_object_replace ((GstObject **) & display, (GstObject *) base->display); ret = gst_va_handle_context_query (GST_ELEMENT_CAST (base), query, display); gst_clear_object (&display); return ret; } static gboolean gst_va_base_enc_src_query (GstVideoEncoder * venc, GstQuery * query) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); gboolean ret = FALSE; switch (GST_QUERY_TYPE (query)) { case GST_QUERY_CONTEXT:{ ret = _query_context (base, query); break; } case GST_QUERY_CAPS:{ GstCaps *caps = NULL, *tmp, *filter = NULL; GstVaEncoder *va_encoder = NULL; gboolean fixed_caps; gst_object_replace ((GstObject **) & va_encoder, (GstObject *) base->encoder); gst_query_parse_caps (query, &filter); fixed_caps = GST_PAD_IS_FIXED_CAPS (GST_VIDEO_ENCODER_SRC_PAD (venc)); if (!fixed_caps && va_encoder) caps = gst_va_encoder_get_srcpad_caps (va_encoder); gst_clear_object (&va_encoder); if (caps) { if (filter) { tmp = gst_caps_intersect_full (filter, caps, GST_CAPS_INTERSECT_FIRST); gst_caps_unref (caps); caps = tmp; } GST_LOG_OBJECT (base, "Returning caps %" GST_PTR_FORMAT, caps); gst_query_set_caps_result (query, caps); gst_caps_unref (caps); ret = TRUE; break; } /* else jump to default */ } default: ret = GST_VIDEO_ENCODER_CLASS (parent_class)->src_query (venc, query); break; } return ret; } static gboolean gst_va_base_enc_sink_query (GstVideoEncoder * venc, GstQuery * query) { GstVaBaseEnc *base = GST_VA_BASE_ENC (venc); if (GST_QUERY_TYPE (query) == GST_QUERY_CONTEXT) return _query_context (base, query); return GST_VIDEO_ENCODER_CLASS (parent_class)->sink_query (venc, query); } static void gst_va_base_enc_set_context (GstElement * element, GstContext * context) { GstVaDisplay *old_display, *new_display; GstVaBaseEnc *base = GST_VA_BASE_ENC (element); GstVaBaseEncClass *klass = GST_VA_BASE_ENC_GET_CLASS (base); gboolean ret; old_display = base->display ? gst_object_ref (base->display) : NULL; ret = gst_va_handle_set_context (element, context, klass->render_device_path, &base->display); new_display = base->display ? gst_object_ref (base->display) : NULL; if (!ret || (old_display && new_display && old_display != new_display && base->encoder)) { GST_ELEMENT_WARNING (element, RESOURCE, BUSY, ("Can't replace VA display while operating"), (NULL)); } gst_clear_object (&old_display); gst_clear_object (&new_display); GST_ELEMENT_CLASS (parent_class)->set_context (element, context); } static void gst_va_base_enc_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVaBaseEnc *base = GST_VA_BASE_ENC (object); switch (prop_id) { case PROP_DEVICE_PATH:{ if (!(base->display && GST_IS_VA_DISPLAY_DRM (base->display))) { g_value_set_string (value, NULL); return; } g_object_get_property (G_OBJECT (base->display), "path", value); break; } default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); } } static void gst_va_base_enc_init (GstVaBaseEnc * self) { g_queue_init (&self->reorder_list); g_queue_init (&self->ref_list); g_queue_init (&self->output_list); self->priv = gst_va_base_enc_get_instance_private (self); } static void gst_va_base_enc_dispose (GObject * object) { _flush_all_frames (GST_VA_BASE_ENC (object)); gst_va_base_enc_close (GST_VIDEO_ENCODER (object)); G_OBJECT_CLASS (parent_class)->dispose (object); } void gst_va_base_enc_class_init (GstVaBaseEncClass * klass) { GObjectClass *gobject_class = G_OBJECT_CLASS (klass); GstElementClass *element_class = GST_ELEMENT_CLASS (klass); GstVideoEncoderClass *encoder_class = GST_VIDEO_ENCODER_CLASS (klass); gobject_class->get_property = gst_va_base_enc_get_property; gobject_class->dispose = gst_va_base_enc_dispose; element_class->set_context = GST_DEBUG_FUNCPTR (gst_va_base_enc_set_context); encoder_class->open = GST_DEBUG_FUNCPTR (gst_va_base_enc_open); encoder_class->close = GST_DEBUG_FUNCPTR (gst_va_base_enc_close); encoder_class->start = GST_DEBUG_FUNCPTR (gst_va_base_enc_start); encoder_class->stop = GST_DEBUG_FUNCPTR (gst_va_base_enc_stop); encoder_class->getcaps = GST_DEBUG_FUNCPTR (gst_va_base_enc_get_caps); encoder_class->src_query = GST_DEBUG_FUNCPTR (gst_va_base_enc_src_query); encoder_class->sink_query = GST_DEBUG_FUNCPTR (gst_va_base_enc_sink_query); encoder_class->propose_allocation = GST_DEBUG_FUNCPTR (gst_va_base_enc_propose_allocation); encoder_class->handle_frame = GST_DEBUG_FUNCPTR (gst_va_base_enc_handle_frame); encoder_class->set_format = GST_DEBUG_FUNCPTR (gst_va_base_enc_set_format); encoder_class->finish = GST_DEBUG_FUNCPTR (gst_va_base_enc_finish); encoder_class->flush = GST_DEBUG_FUNCPTR (gst_va_base_enc_flush); klass->reset_state = GST_DEBUG_FUNCPTR (gst_va_base_enc_reset_state_default); /** * GstVaBaseEnc:device-path: * * It shows the DRM device path used for the VA operation, if any. */ properties[PROP_DEVICE_PATH] = g_param_spec_string ("device-path", "Device Path", "DRM device path", NULL, G_PARAM_READABLE | G_PARAM_STATIC_STRINGS); g_object_class_install_properties (gobject_class, N_PROPERTIES, properties); gst_type_mark_as_plugin_api (GST_TYPE_VA_BASE_ENC, 0); } /*********************** Helper Functions ****************************/ gboolean gst_va_base_enc_add_rate_control_parameter (GstVaBaseEnc * base, GstVaEncodePicture * picture, guint32 rc_mode, guint max_bitrate_bits, guint target_percentage, guint32 qp_i, guint32 min_qp, guint32 max_qp, guint32 mbbrc) { uint32_t window_size; struct VAEncMiscParameterRateControlWrap { VAEncMiscParameterType type; VAEncMiscParameterRateControl rate_control; } rate_control; if (rc_mode == VA_RC_NONE || rc_mode == VA_RC_CQP) return TRUE; window_size = rc_mode == VA_RC_VBR ? max_bitrate_bits / 2 : max_bitrate_bits; /* *INDENT-OFF* */ rate_control = (struct VAEncMiscParameterRateControlWrap) { .type = VAEncMiscParameterTypeRateControl, .rate_control = { .bits_per_second = max_bitrate_bits, .target_percentage = target_percentage, .window_size = window_size, .initial_qp = qp_i, .min_qp = min_qp, .max_qp = max_qp, .rc_flags.bits.mb_rate_control = mbbrc, .quality_factor = 0, }, }; /* *INDENT-ON* */ if (!gst_va_encoder_add_param (base->encoder, picture, VAEncMiscParameterBufferType, &rate_control, sizeof (rate_control))) { GST_ERROR_OBJECT (base, "Failed to create the race control parameter"); return FALSE; } return TRUE; } gboolean gst_va_base_enc_add_quality_level_parameter (GstVaBaseEnc * base, GstVaEncodePicture * picture, guint target_usage) { /* *INDENT-OFF* */ struct { VAEncMiscParameterType type; VAEncMiscParameterBufferQualityLevel ql; } quality_level = { .type = VAEncMiscParameterTypeQualityLevel, .ql.quality_level = target_usage, }; /* *INDENT-ON* */ if (target_usage == 0) return TRUE; if (!gst_va_encoder_add_param (base->encoder, picture, VAEncMiscParameterBufferType, &quality_level, sizeof (quality_level))) { GST_ERROR_OBJECT (base, "Failed to create the quality level parameter"); return FALSE; } return TRUE; } gboolean gst_va_base_enc_add_frame_rate_parameter (GstVaBaseEnc * base, GstVaEncodePicture * picture) { /* *INDENT-OFF* */ struct { VAEncMiscParameterType type; VAEncMiscParameterFrameRate fr; } framerate = { .type = VAEncMiscParameterTypeFrameRate, /* denominator = framerate >> 16 & 0xffff; * numerator = framerate & 0xffff; */ .fr.framerate = (GST_VIDEO_INFO_FPS_N (&base->input_state->info) & 0xffff) | ((GST_VIDEO_INFO_FPS_D (&base->input_state->info) & 0xffff) << 16) }; /* *INDENT-ON* */ if (!gst_va_encoder_add_param (base->encoder, picture, VAEncMiscParameterBufferType, &framerate, sizeof (framerate))) { GST_ERROR_OBJECT (base, "Failed to create the frame rate parameter"); return FALSE; } return TRUE; } gboolean gst_va_base_enc_add_hrd_parameter (GstVaBaseEnc * base, GstVaEncodePicture * picture, guint32 rc_mode, guint cpb_length_bits) { /* *INDENT-OFF* */ struct { VAEncMiscParameterType type; VAEncMiscParameterHRD hrd; } hrd = { .type = VAEncMiscParameterTypeHRD, .hrd = { .buffer_size = cpb_length_bits, .initial_buffer_fullness = cpb_length_bits / 2, }, }; /* *INDENT-ON* */ if (rc_mode == VA_RC_NONE || rc_mode == VA_RC_CQP || rc_mode == VA_RC_VCM) return TRUE; if (!gst_va_encoder_add_param (base->encoder, picture, VAEncMiscParameterBufferType, &hrd, sizeof (hrd))) { GST_ERROR_OBJECT (base, "Failed to create the HRD parameter"); return FALSE; } return TRUE; } gboolean gst_va_base_enc_add_trellis_parameter (GstVaBaseEnc * base, GstVaEncodePicture * picture, gboolean use_trellis) { /* *INDENT-OFF* */ struct { VAEncMiscParameterType type; VAEncMiscParameterQuantization tr; } trellis = { .type = VAEncMiscParameterTypeQuantization, .tr.quantization_flags.bits = { .disable_trellis = 0, .enable_trellis_I = 1, .enable_trellis_B = 1, .enable_trellis_P = 1, }, }; /* *INDENT-ON* */ if (!use_trellis) return TRUE; if (!gst_va_encoder_add_param (base->encoder, picture, VAEncMiscParameterBufferType, &trellis, sizeof (trellis))) { GST_ERROR_OBJECT (base, "Failed to create the trellis parameter"); return FALSE; } return TRUE; } void gst_va_base_enc_add_codec_tag (GstVaBaseEnc * base, const gchar * codec_name) { GstVideoEncoder *venc = GST_VIDEO_ENCODER (base); GstTagList *tags = gst_tag_list_new_empty (); const gchar *encoder_name; guint bitrate = 0; g_object_get (venc, "bitrate", &bitrate, NULL); if (bitrate > 0) gst_tag_list_add (tags, GST_TAG_MERGE_REPLACE, GST_TAG_NOMINAL_BITRATE, bitrate, NULL); if ((encoder_name = gst_element_class_get_metadata (GST_ELEMENT_GET_CLASS (venc), GST_ELEMENT_METADATA_LONGNAME))) gst_tag_list_add (tags, GST_TAG_MERGE_REPLACE, GST_TAG_ENCODER, encoder_name, NULL); gst_tag_list_add (tags, GST_TAG_MERGE_REPLACE, GST_TAG_CODEC, codec_name, NULL); gst_video_encoder_merge_tags (venc, tags, GST_TAG_MERGE_REPLACE); gst_tag_list_unref (tags); } void gst_va_base_enc_reset_state (GstVaBaseEnc * base) { GstVaBaseEncClass *klass = GST_VA_BASE_ENC_GET_CLASS (base); g_assert (klass->reset_state); klass->reset_state (base); } /* *INDENT-OFF* */ #define UPDATE_PROPERTY \ GST_OBJECT_LOCK (base); \ if (*old_val == new_val) { \ GST_OBJECT_UNLOCK (base); \ return; \ } \ *old_val = new_val; \ GST_OBJECT_UNLOCK (base); \ if (pspec) \ g_object_notify_by_pspec (G_OBJECT (base), pspec); \ void gst_va_base_enc_update_property_uint (GstVaBaseEnc * base, guint32 * old_val, guint32 new_val, GParamSpec * pspec) { UPDATE_PROPERTY } void gst_va_base_enc_update_property_bool (GstVaBaseEnc * base, gboolean * old_val, gboolean new_val, GParamSpec * pspec) { UPDATE_PROPERTY } #undef UPDATE_PROPERTY /* *INDENT-ON* */