mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-19 14:56:36 +00:00
9c21923f04
Introduce GstH264Decoder based Nvidia H.264 decoder element. Similar the element factory name of to v4l2 stateless codec, this element can be configured with factory name "gstnvh264sldec". Note that "sl" in the name stands for "stateless" For now, existing nvh264dec covers more profile and formats (e.g., interlaced stream) than this implementation. However, this implementation allows us to control lower level parameters such as decoded picture buffer management and therefore we can get a chance to improve performance in terms of latency. Part-of: <https://gitlab.freedesktop.org/gstreamer/gst-plugins-bad/-/merge_requests/1198>
1102 lines
32 KiB
C
1102 lines
32 KiB
C
/* GStreamer
|
|
* Copyright (C) 2017 Ericsson AB. All rights reserved.
|
|
* Copyright (C) 2020 Seungha Yang <seungha@centricular.com>
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer
|
|
* in the documentation and/or other materials provided with the
|
|
* distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#ifdef HAVE_NVCODEC_GST_GL
|
|
#include <gst/gl/gl.h>
|
|
#include <gst/gl/gstglfuncs.h>
|
|
#endif
|
|
|
|
#include "gstnvdecoder.h"
|
|
#include <string.h>
|
|
|
|
|
|
GST_DEBUG_CATEGORY_EXTERN (gst_nv_decoder_debug);
|
|
#define GST_CAT_DEFAULT gst_nv_decoder_debug
|
|
|
|
typedef struct _GstNvDecoderFrameInfo
|
|
{
|
|
gboolean available;
|
|
} GstNvDecoderFrameInfo;
|
|
|
|
struct _GstNvDecoder
|
|
{
|
|
GstObject parent;
|
|
GstCudaContext *context;
|
|
CUvideodecoder decoder_handle;
|
|
|
|
GstNvDecoderFrameInfo *frame_pool;
|
|
guint pool_size;
|
|
|
|
GstVideoInfo info;
|
|
};
|
|
|
|
static void gst_nv_decoder_dispose (GObject * object);
|
|
static void gst_nv_decoder_finalize (GObject * object);
|
|
|
|
#define parent_class gst_nv_decoder_parent_class
|
|
G_DEFINE_TYPE (GstNvDecoder, gst_nv_decoder, GST_TYPE_OBJECT);
|
|
|
|
static void
|
|
gst_nv_decoder_class_init (GstNvDecoderClass * klass)
|
|
{
|
|
GObjectClass *gobject_class = G_OBJECT_CLASS (klass);
|
|
|
|
gobject_class->dispose = gst_nv_decoder_dispose;
|
|
gobject_class->finalize = gst_nv_decoder_finalize;
|
|
}
|
|
|
|
static void
|
|
gst_nv_decoder_init (GstNvDecoder * self)
|
|
{
|
|
}
|
|
|
|
static void
|
|
gst_nv_decoder_dispose (GObject * object)
|
|
{
|
|
GstNvDecoder *self = GST_NV_DECODER (object);
|
|
|
|
gst_clear_object (&self->context);
|
|
|
|
G_OBJECT_CLASS (parent_class)->dispose (object);
|
|
}
|
|
|
|
static void
|
|
gst_nv_decoder_finalize (GObject * object)
|
|
{
|
|
GstNvDecoder *self = GST_NV_DECODER (object);
|
|
|
|
g_free (self->frame_pool);
|
|
if (self->decoder_handle)
|
|
gst_cuda_result (CuvidDestroyDecoder (self->decoder_handle));
|
|
|
|
G_OBJECT_CLASS (parent_class)->finalize (object);
|
|
}
|
|
|
|
static cudaVideoChromaFormat
|
|
chroma_format_from_video_format (GstVideoFormat format)
|
|
{
|
|
switch (format) {
|
|
case GST_VIDEO_FORMAT_NV12:
|
|
case GST_VIDEO_FORMAT_P010_10LE:
|
|
case GST_VIDEO_FORMAT_P010_10BE:
|
|
case GST_VIDEO_FORMAT_P016_LE:
|
|
case GST_VIDEO_FORMAT_P016_BE:
|
|
return cudaVideoChromaFormat_420;
|
|
case GST_VIDEO_FORMAT_Y444:
|
|
case GST_VIDEO_FORMAT_Y444_16LE:
|
|
case GST_VIDEO_FORMAT_Y444_16BE:
|
|
return cudaVideoChromaFormat_444;
|
|
default:
|
|
g_assert_not_reached ();
|
|
break;
|
|
}
|
|
|
|
return cudaVideoChromaFormat_420;
|
|
}
|
|
|
|
static guint
|
|
bitdepth_minus8_from_video_format (GstVideoFormat format)
|
|
{
|
|
switch (format) {
|
|
case GST_VIDEO_FORMAT_NV12:
|
|
case GST_VIDEO_FORMAT_Y444:
|
|
return 0;
|
|
case GST_VIDEO_FORMAT_P010_10LE:
|
|
case GST_VIDEO_FORMAT_P010_10BE:
|
|
return 2;
|
|
case GST_VIDEO_FORMAT_P016_LE:
|
|
case GST_VIDEO_FORMAT_P016_BE:
|
|
case GST_VIDEO_FORMAT_Y444_16LE:
|
|
case GST_VIDEO_FORMAT_Y444_16BE:
|
|
return 8;
|
|
default:
|
|
g_assert_not_reached ();
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static cudaVideoSurfaceFormat
|
|
output_format_from_video_format (GstVideoFormat format)
|
|
{
|
|
switch (format) {
|
|
case GST_VIDEO_FORMAT_NV12:
|
|
return cudaVideoSurfaceFormat_NV12;
|
|
case GST_VIDEO_FORMAT_P010_10LE:
|
|
case GST_VIDEO_FORMAT_P010_10BE:
|
|
case GST_VIDEO_FORMAT_P016_LE:
|
|
case GST_VIDEO_FORMAT_P016_BE:
|
|
return cudaVideoSurfaceFormat_P016;
|
|
case GST_VIDEO_FORMAT_Y444:
|
|
return cudaVideoSurfaceFormat_YUV444;
|
|
case GST_VIDEO_FORMAT_Y444_16LE:
|
|
case GST_VIDEO_FORMAT_Y444_16BE:
|
|
return cudaVideoSurfaceFormat_YUV444_16Bit;
|
|
default:
|
|
g_assert_not_reached ();
|
|
break;
|
|
}
|
|
|
|
return cudaVideoSurfaceFormat_NV12;
|
|
}
|
|
|
|
static gboolean
|
|
gst_nv_decoder_prepare_frame_pool (GstNvDecoder * self, guint pool_size)
|
|
{
|
|
gint i;
|
|
|
|
self->frame_pool = g_new (GstNvDecoderFrameInfo, pool_size);
|
|
|
|
for (i = 0; i < pool_size; i++)
|
|
self->frame_pool[i].available = TRUE;
|
|
|
|
self->pool_size = pool_size;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
GstNvDecoder *
|
|
gst_nv_decoder_new (GstCudaContext * context, cudaVideoCodec codec,
|
|
GstVideoInfo * info, guint pool_size)
|
|
{
|
|
GstNvDecoder *decoder;
|
|
CUVIDDECODECREATEINFO create_info = { 0, };
|
|
GstVideoFormat format;
|
|
|
|
g_return_val_if_fail (GST_IS_CUDA_CONTEXT (context), NULL);
|
|
g_return_val_if_fail (codec < cudaVideoCodec_NumCodecs, NULL);
|
|
g_return_val_if_fail (info != NULL, NULL);
|
|
g_return_val_if_fail (pool_size > 0, NULL);
|
|
|
|
decoder = g_object_new (GST_TYPE_NV_DECODER, NULL);
|
|
decoder->context = gst_object_ref (context);
|
|
gst_object_ref_sink (decoder);
|
|
|
|
format = GST_VIDEO_INFO_FORMAT (info);
|
|
|
|
/* FIXME: check aligned resolution or actaul coded resolution */
|
|
create_info.ulWidth = GST_VIDEO_INFO_WIDTH (info);;
|
|
create_info.ulHeight = GST_VIDEO_INFO_HEIGHT (info);;
|
|
create_info.ulNumDecodeSurfaces = pool_size;
|
|
create_info.CodecType = codec;
|
|
create_info.ChromaFormat = chroma_format_from_video_format (format);
|
|
create_info.ulCreationFlags = cudaVideoCreate_Default;
|
|
create_info.bitDepthMinus8 = bitdepth_minus8_from_video_format (format);
|
|
create_info.ulIntraDecodeOnly = 0;
|
|
|
|
create_info.display_area.left = 0;
|
|
create_info.display_area.top = 0;
|
|
create_info.display_area.right = GST_VIDEO_INFO_WIDTH (info);
|
|
create_info.display_area.bottom = GST_VIDEO_INFO_HEIGHT (info);
|
|
create_info.OutputFormat = output_format_from_video_format (format);
|
|
create_info.DeinterlaceMode = cudaVideoDeinterlaceMode_Weave;
|
|
|
|
create_info.ulTargetWidth = GST_VIDEO_INFO_WIDTH (info);;
|
|
create_info.ulTargetHeight = GST_VIDEO_INFO_HEIGHT (info);
|
|
/* we always copy decoded picture to output buffer */
|
|
create_info.ulNumOutputSurfaces = 1;
|
|
|
|
create_info.target_rect.left = 0;
|
|
create_info.target_rect.top = 0;
|
|
create_info.target_rect.right = GST_VIDEO_INFO_WIDTH (info);
|
|
create_info.target_rect.bottom = GST_VIDEO_INFO_HEIGHT (info);
|
|
|
|
if (!gst_cuda_context_push (context)) {
|
|
GST_ERROR_OBJECT (decoder, "Failed to lock CUDA context");
|
|
goto error;
|
|
}
|
|
|
|
if (!gst_cuda_result (CuvidCreateDecoder (&decoder->decoder_handle,
|
|
&create_info))) {
|
|
GST_ERROR_OBJECT (decoder, "Cannot create decoder instance");
|
|
goto error;
|
|
}
|
|
|
|
if (!gst_cuda_context_pop (NULL)) {
|
|
GST_ERROR_OBJECT (decoder, "Failed to unlock CUDA context");
|
|
goto error;
|
|
}
|
|
|
|
if (!gst_nv_decoder_prepare_frame_pool (decoder, pool_size)) {
|
|
GST_ERROR_OBJECT (decoder, "Cannot prepare internal surface buffer pool");
|
|
goto error;
|
|
}
|
|
|
|
decoder->info = *info;
|
|
|
|
return decoder;
|
|
|
|
error:
|
|
gst_clear_object (&decoder);
|
|
return NULL;
|
|
}
|
|
|
|
GstNvDecoderFrame *
|
|
gst_nv_decoder_new_frame (GstNvDecoder * decoder)
|
|
{
|
|
GstNvDecoderFrame *frame;
|
|
gint i;
|
|
gint index_to_use = -1;
|
|
|
|
g_return_val_if_fail (GST_IS_NV_DECODER (decoder), NULL);
|
|
|
|
for (i = 0; i < decoder->pool_size; i++) {
|
|
if (decoder->frame_pool[i].available) {
|
|
decoder->frame_pool[i].available = FALSE;
|
|
index_to_use = i;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (index_to_use < 0) {
|
|
GST_ERROR_OBJECT (decoder, "No available frame");
|
|
return NULL;
|
|
}
|
|
|
|
frame = g_new0 (GstNvDecoderFrame, 1);
|
|
frame->index = index_to_use;
|
|
frame->decoder = gst_object_ref (decoder);
|
|
|
|
GST_LOG_OBJECT (decoder, "New frame %p (index %d)", frame, frame->index);
|
|
|
|
return frame;
|
|
}
|
|
|
|
/* must be called with gst_cuda_context_push */
|
|
static gboolean
|
|
gst_nv_decoder_frame_map (GstNvDecoderFrame * frame)
|
|
{
|
|
GstNvDecoder *self;
|
|
CUVIDPROCPARAMS params = { 0 };
|
|
|
|
g_return_val_if_fail (frame != NULL, FALSE);
|
|
g_return_val_if_fail (frame->index >= 0, FALSE);
|
|
g_return_val_if_fail (GST_IS_NV_DECODER (frame->decoder), FALSE);
|
|
|
|
self = frame->decoder;
|
|
|
|
/* TODO: check interlaced */
|
|
params.progressive_frame = 1;
|
|
|
|
if (frame->mapped) {
|
|
GST_WARNING_OBJECT (self, "Frame %p is mapped already", frame);
|
|
return TRUE;
|
|
}
|
|
|
|
if (!gst_cuda_result (CuvidMapVideoFrame (self->decoder_handle,
|
|
frame->index, &frame->devptr, &frame->pitch, ¶ms))) {
|
|
GST_ERROR_OBJECT (self, "Cannot map picture");
|
|
return FALSE;
|
|
}
|
|
|
|
frame->mapped = TRUE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/* must be called with gst_cuda_context_push */
|
|
static void
|
|
gst_nv_decoder_frame_unmap (GstNvDecoderFrame * frame)
|
|
{
|
|
GstNvDecoder *self;
|
|
|
|
g_return_if_fail (frame != NULL);
|
|
g_return_if_fail (frame->index >= 0);
|
|
g_return_if_fail (GST_IS_NV_DECODER (frame->decoder));
|
|
|
|
self = frame->decoder;
|
|
|
|
if (!frame->mapped) {
|
|
GST_WARNING_OBJECT (self, "Frame %p is not mapped", frame);
|
|
return;
|
|
}
|
|
|
|
if (!gst_cuda_result (CuvidUnmapVideoFrame (self->decoder_handle,
|
|
frame->devptr))) {
|
|
GST_ERROR_OBJECT (self, "Cannot unmap picture");
|
|
}
|
|
|
|
frame->mapped = FALSE;
|
|
}
|
|
|
|
void
|
|
gst_nv_decoder_frame_free (GstNvDecoderFrame * frame)
|
|
{
|
|
GstNvDecoder *self;
|
|
|
|
g_assert (frame != NULL);
|
|
|
|
GST_LOG ("Free frame %p (index %d)", frame, frame->index);
|
|
|
|
if (frame->decoder) {
|
|
self = frame->decoder;
|
|
if (frame->mapped && gst_cuda_context_push (self->context)) {
|
|
gst_nv_decoder_frame_unmap (frame);
|
|
gst_cuda_context_pop (NULL);
|
|
}
|
|
|
|
if (frame->index < self->pool_size) {
|
|
self->frame_pool[frame->index].available = TRUE;
|
|
} else {
|
|
GST_WARNING_OBJECT (self,
|
|
"Frame %p has invalid index %d", frame, frame->index);
|
|
}
|
|
|
|
gst_object_unref (self);
|
|
}
|
|
|
|
g_free (frame);
|
|
}
|
|
|
|
gboolean
|
|
gst_nv_decoder_decode_picture (GstNvDecoder * decoder, CUVIDPICPARAMS * params)
|
|
{
|
|
GstCudaContext *ctx = decoder->context;
|
|
gboolean ret = TRUE;
|
|
|
|
GST_LOG_OBJECT (decoder, "picture index: %u", params->CurrPicIdx);
|
|
|
|
if (!gst_cuda_context_push (ctx)) {
|
|
GST_ERROR_OBJECT (decoder, "Failed to push CUDA context");
|
|
return FALSE;
|
|
}
|
|
|
|
if (!gst_cuda_result (CuvidDecodePicture (decoder->decoder_handle, params))) {
|
|
GST_ERROR_OBJECT (decoder, "Failed to decode picture");
|
|
ret = FALSE;
|
|
}
|
|
|
|
if (!gst_cuda_context_pop (NULL)) {
|
|
GST_WARNING_OBJECT (decoder, "Failed to pop CUDA context");
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
#ifdef HAVE_NVCODEC_GST_GL
|
|
static gboolean
|
|
gst_nv_decoder_register_cuda_resource (GstNvDecoder * self, GstMemory * mem,
|
|
GstCudaGraphicsResource * resource)
|
|
{
|
|
GstMapInfo info;
|
|
gboolean ret = FALSE;
|
|
|
|
if (!gst_cuda_context_push (self->context)) {
|
|
GST_ERROR_OBJECT (self, "Failed to push CUDA context");
|
|
return FALSE;
|
|
}
|
|
|
|
if (gst_memory_map (mem, &info, GST_MAP_READ | GST_MAP_GL)) {
|
|
GstGLMemoryPBO *gl_mem = (GstGLMemoryPBO *) mem;
|
|
GstGLBuffer *gl_buffer = gl_mem->pbo;
|
|
|
|
GST_LOG_OBJECT (self,
|
|
"Register glbuffer %d to CUDA resource", gl_buffer->id);
|
|
|
|
/* register resource without read/write only flags, since
|
|
* downstream CUDA elements (e.g., nvenc) might want to access
|
|
* this resource later. Instead, use map flags during map/unmap */
|
|
if (gst_cuda_graphics_resource_register_gl_buffer (resource,
|
|
gl_buffer->id, CU_GRAPHICS_REGISTER_FLAGS_NONE)) {
|
|
ret = TRUE;
|
|
} else {
|
|
GST_WARNING_OBJECT (self, "Failed to register memory");
|
|
}
|
|
|
|
gst_memory_unmap (mem, &info);
|
|
} else {
|
|
GST_WARNING_OBJECT (self, "Failed to map memory");
|
|
}
|
|
|
|
if (!gst_cuda_context_pop (NULL))
|
|
GST_WARNING_OBJECT (self, "Failed to pop CUDA context");
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstCudaGraphicsResource *
|
|
gst_nv_decoder_ensure_cuda_graphics_resource (GstNvDecoder * self,
|
|
GstMemory * mem)
|
|
{
|
|
GQuark quark;
|
|
GstCudaGraphicsResource *resource;
|
|
|
|
if (!gst_is_gl_memory_pbo (mem)) {
|
|
GST_WARNING_OBJECT (self, "memory is not GL PBO memory, %s",
|
|
mem->allocator->mem_type);
|
|
return NULL;
|
|
}
|
|
|
|
quark = gst_cuda_quark_from_id (GST_CUDA_QUARK_GRAPHICS_RESOURCE);
|
|
|
|
resource = (GstCudaGraphicsResource *)
|
|
gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), quark);
|
|
|
|
if (!resource) {
|
|
gboolean ret;
|
|
|
|
resource = gst_cuda_graphics_resource_new (self->context,
|
|
GST_OBJECT (GST_GL_BASE_MEMORY_CAST (mem)->context),
|
|
GST_CUDA_GRAPHICS_RESOURCE_GL_BUFFER);
|
|
|
|
ret = gst_nv_decoder_register_cuda_resource (self, mem, resource);
|
|
if (!ret) {
|
|
GST_WARNING_OBJECT (self, "Couldn't register resource");
|
|
gst_cuda_graphics_resource_free (resource);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
gst_mini_object_set_qdata (GST_MINI_OBJECT (mem), quark, resource,
|
|
(GDestroyNotify) gst_cuda_graphics_resource_free);
|
|
}
|
|
|
|
return resource;
|
|
}
|
|
|
|
typedef struct
|
|
{
|
|
GstNvDecoder *self;
|
|
gboolean ret;
|
|
GstNvDecoderFrame *frame;
|
|
GstBuffer *output_buffer;
|
|
} GstNvDecoderCopyToGLData;
|
|
|
|
static void
|
|
gst_nv_decoder_copy_frame_to_gl_internal (GstGLContext * context,
|
|
GstNvDecoderCopyToGLData * data)
|
|
{
|
|
GstNvDecoder *self = data->self;
|
|
GstNvDecoderFrame *frame = data->frame;
|
|
GstCudaGraphicsResource **resources;
|
|
guint num_resources;
|
|
guint i;
|
|
CUDA_MEMCPY2D copy_params = { 0, };
|
|
GstVideoInfo *info = &self->info;
|
|
|
|
data->ret = TRUE;
|
|
|
|
num_resources = gst_buffer_n_memory (data->output_buffer);
|
|
resources = g_newa (GstCudaGraphicsResource *, num_resources);
|
|
|
|
for (i = 0; i < num_resources; i++) {
|
|
GstMemory *mem;
|
|
|
|
mem = gst_buffer_peek_memory (data->output_buffer, i);
|
|
resources[i] = gst_nv_decoder_ensure_cuda_graphics_resource (self, mem);
|
|
if (!resources[i]) {
|
|
GST_WARNING_OBJECT (self, "could not register %dth memory", i);
|
|
data->ret = FALSE;
|
|
|
|
return;
|
|
}
|
|
|
|
/* Need PBO -> texture */
|
|
GST_MINI_OBJECT_FLAG_SET (mem, GST_GL_BASE_MEMORY_TRANSFER_NEED_UPLOAD);
|
|
}
|
|
|
|
if (!gst_cuda_context_push (self->context)) {
|
|
GST_WARNING_OBJECT (self, "Failed to push CUDA context");
|
|
data->ret = FALSE;
|
|
return;
|
|
}
|
|
|
|
copy_params.srcMemoryType = CU_MEMORYTYPE_DEVICE;
|
|
copy_params.srcPitch = frame->pitch;
|
|
copy_params.dstMemoryType = CU_MEMORYTYPE_DEVICE;
|
|
|
|
for (i = 0; i < num_resources; i++) {
|
|
CUdeviceptr dst_ptr;
|
|
gsize size;
|
|
CUgraphicsResource cuda_resource =
|
|
gst_cuda_graphics_resource_map (resources[i], NULL,
|
|
CU_GRAPHICS_MAP_RESOURCE_FLAGS_WRITE_DISCARD);
|
|
|
|
if (!cuda_resource) {
|
|
GST_WARNING_OBJECT (self, "failed to map CUDA resources");
|
|
data->ret = FALSE;
|
|
goto unmap_video_frame;
|
|
}
|
|
|
|
if (!gst_cuda_result (CuGraphicsResourceGetMappedPointer (&dst_ptr, &size,
|
|
cuda_resource))) {
|
|
GST_WARNING_OBJECT (self, "failed to map CUDA resource");
|
|
data->ret = FALSE;
|
|
break;
|
|
}
|
|
|
|
copy_params.dstPitch = GST_VIDEO_INFO_PLANE_STRIDE (info, i);
|
|
copy_params.WidthInBytes = GST_VIDEO_INFO_COMP_WIDTH (info, i)
|
|
* GST_VIDEO_INFO_COMP_PSTRIDE (info, i);
|
|
|
|
copy_params.srcDevice = frame->devptr +
|
|
(i * frame->pitch * GST_VIDEO_INFO_HEIGHT (info));
|
|
copy_params.dstDevice = dst_ptr;
|
|
copy_params.Height = GST_VIDEO_INFO_COMP_HEIGHT (info, i);
|
|
|
|
if (!gst_cuda_result (CuMemcpy2DAsync (©_params, NULL))) {
|
|
GST_WARNING_OBJECT (self, "memcpy to mapped array failed");
|
|
data->ret = FALSE;
|
|
}
|
|
}
|
|
|
|
gst_cuda_result (CuStreamSynchronize (NULL));
|
|
|
|
unmap_video_frame:
|
|
for (i = 0; i < num_resources; i++) {
|
|
gst_cuda_graphics_resource_unmap (resources[i], NULL);
|
|
}
|
|
|
|
if (!gst_cuda_context_pop (NULL))
|
|
GST_WARNING_OBJECT (self, "Failed to pop CUDA context");
|
|
}
|
|
|
|
static gboolean
|
|
gst_nv_decoder_copy_frame_to_gl (GstNvDecoder * decoder,
|
|
GstGLContext * context, GstNvDecoderFrame * frame, GstBuffer * buffer)
|
|
{
|
|
GstNvDecoderCopyToGLData data;
|
|
|
|
data.self = decoder;
|
|
data.frame = frame;
|
|
data.output_buffer = buffer;
|
|
|
|
gst_gl_context_thread_add (context,
|
|
(GstGLContextThreadFunc) gst_nv_decoder_copy_frame_to_gl_internal, &data);
|
|
|
|
GST_LOG_OBJECT (decoder, "Copy frame to GL ret %d", data.ret);
|
|
|
|
return data.ret;
|
|
}
|
|
#endif
|
|
|
|
static gboolean
|
|
gst_nv_decoder_copy_frame_to_system (GstNvDecoder * decoder,
|
|
GstNvDecoderFrame * frame, GstBuffer * buffer)
|
|
{
|
|
GstVideoFrame video_frame;
|
|
CUDA_MEMCPY2D copy_params = { 0, };
|
|
gint i;
|
|
gboolean ret = FALSE;
|
|
|
|
if (!gst_video_frame_map (&video_frame, &decoder->info, buffer,
|
|
GST_MAP_WRITE)) {
|
|
GST_ERROR_OBJECT (decoder, "Couldn't map video frame");
|
|
return FALSE;
|
|
}
|
|
|
|
if (!gst_cuda_context_push (decoder->context)) {
|
|
GST_ERROR_OBJECT (decoder, "Failed to pust CUDA context");
|
|
gst_video_frame_unmap (&video_frame);
|
|
return FALSE;
|
|
}
|
|
|
|
copy_params.srcMemoryType = CU_MEMORYTYPE_DEVICE;
|
|
copy_params.srcPitch = frame->pitch;
|
|
copy_params.dstMemoryType = CU_MEMORYTYPE_HOST;
|
|
copy_params.WidthInBytes = GST_VIDEO_INFO_COMP_WIDTH (&decoder->info, 0)
|
|
* GST_VIDEO_INFO_COMP_PSTRIDE (&decoder->info, 0);
|
|
|
|
for (i = 0; i < GST_VIDEO_FRAME_N_PLANES (&video_frame); i++) {
|
|
copy_params.srcDevice = frame->devptr +
|
|
(i * frame->pitch * GST_VIDEO_INFO_HEIGHT (&decoder->info));
|
|
copy_params.dstHost = GST_VIDEO_FRAME_PLANE_DATA (&video_frame, i);
|
|
copy_params.dstPitch = GST_VIDEO_FRAME_PLANE_STRIDE (&video_frame, i);
|
|
copy_params.Height = GST_VIDEO_FRAME_COMP_HEIGHT (&video_frame, i);
|
|
|
|
if (!gst_cuda_result (CuMemcpy2DAsync (©_params, NULL))) {
|
|
GST_ERROR_OBJECT (decoder, "failed to copy %dth plane", i);
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
gst_cuda_result (CuStreamSynchronize (NULL));
|
|
|
|
ret = TRUE;
|
|
|
|
done:
|
|
gst_cuda_context_pop (NULL);
|
|
|
|
gst_video_frame_unmap (&video_frame);
|
|
|
|
GST_LOG_OBJECT (decoder, "Copy frame to system ret %d", ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
gboolean
|
|
gst_nv_decoder_finish_frame (GstNvDecoder * decoder,
|
|
GstNvDecoderOutputType output_type, GstObject * graphics_context,
|
|
GstNvDecoderFrame * frame, GstBuffer * buffer)
|
|
{
|
|
gboolean ret;
|
|
|
|
g_return_val_if_fail (GST_IS_NV_DECODER (decoder), FALSE);
|
|
g_return_val_if_fail (frame != NULL, FALSE);
|
|
g_return_val_if_fail (GST_IS_BUFFER (buffer), FALSE);
|
|
|
|
#ifdef HAVE_NVCODEC_GST_GL
|
|
if (output_type == GST_NV_DECOCER_OUTPUT_TYPE_GL && !graphics_context) {
|
|
if (!GST_IS_GL_CONTEXT (graphics_context)) {
|
|
GST_ERROR_OBJECT (decoder, "Invalid GL Context");
|
|
return FALSE;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
if (!gst_cuda_context_push (decoder->context)) {
|
|
GST_ERROR_OBJECT (decoder, "Failed to pust CUDA context");
|
|
return FALSE;
|
|
}
|
|
|
|
if (!gst_nv_decoder_frame_map (frame)) {
|
|
GST_ERROR_OBJECT (decoder, "Couldn't map frame");
|
|
gst_cuda_context_pop (NULL);
|
|
return FALSE;
|
|
}
|
|
|
|
gst_cuda_context_pop (NULL);
|
|
|
|
#ifdef HAVE_NVCODEC_GST_GL
|
|
if (output_type == GST_NV_DECOCER_OUTPUT_TYPE_GL) {
|
|
ret = gst_nv_decoder_copy_frame_to_gl (decoder,
|
|
GST_GL_CONTEXT (graphics_context), frame, buffer);
|
|
} else
|
|
#endif
|
|
{
|
|
ret = gst_nv_decoder_copy_frame_to_system (decoder, frame, buffer);
|
|
}
|
|
|
|
gst_cuda_context_push (decoder->context);
|
|
gst_nv_decoder_frame_unmap (frame);
|
|
gst_cuda_context_pop (NULL);
|
|
|
|
return ret;
|
|
}
|
|
|
|
typedef enum
|
|
{
|
|
GST_NV_DECODER_FORMAT_FLAG_NONE = (1 << 0),
|
|
GST_NV_DECODER_FORMAT_FLAG_420_8BITS = (1 << 1),
|
|
GST_NV_DECODER_FORMAT_FLAG_420_10BITS = (1 << 2),
|
|
GST_NV_DECODER_FORMAT_FLAG_420_12BITS = (1 << 3),
|
|
GST_NV_DECODER_FORMAT_FLAG_444_8BITS = (1 << 4),
|
|
GST_NV_DECODER_FORMAT_FLAG_444_10BITS = (1 << 5),
|
|
GST_NV_DECODER_FORMAT_FLAG_444_12BITS = (1 << 6),
|
|
} GstNvDecoderFormatFlags;
|
|
|
|
static gboolean
|
|
gst_nv_decocer_get_supported_codec_profiles (GValue * profiles,
|
|
cudaVideoCodec codec, GstNvDecoderFormatFlags flags)
|
|
{
|
|
GValue val = G_VALUE_INIT;
|
|
gboolean ret = FALSE;
|
|
|
|
g_value_init (&val, G_TYPE_STRING);
|
|
|
|
switch (codec) {
|
|
case cudaVideoCodec_H264:
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_420_8BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_420_8BITS) {
|
|
g_value_set_static_string (&val, "constrained-baseline");
|
|
gst_value_list_append_value (profiles, &val);
|
|
|
|
g_value_set_static_string (&val, "baseline");
|
|
gst_value_list_append_value (profiles, &val);
|
|
|
|
g_value_set_static_string (&val, "main");
|
|
gst_value_list_append_value (profiles, &val);
|
|
|
|
g_value_set_static_string (&val, "high");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
/* NVDEC supports only 4:2:0 8bits h264 decoding.
|
|
* following conditions are for the future enhancement */
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_420_10BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_420_10BITS) {
|
|
g_value_set_static_string (&val, "high-10");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_420_12BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_420_12BITS ||
|
|
(flags & GST_NV_DECODER_FORMAT_FLAG_444_8BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_444_8BITS ||
|
|
(flags & GST_NV_DECODER_FORMAT_FLAG_444_10BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_444_10BITS ||
|
|
(flags & GST_NV_DECODER_FORMAT_FLAG_444_12BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_444_12BITS) {
|
|
g_value_set_static_string (&val, "high-4:4:4");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
ret = TRUE;
|
|
break;
|
|
case cudaVideoCodec_HEVC:
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_420_8BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_420_8BITS) {
|
|
g_value_set_static_string (&val, "main");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_420_10BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_420_10BITS) {
|
|
g_value_set_static_string (&val, "main-10");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_420_12BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_420_12BITS) {
|
|
g_value_set_static_string (&val, "main-12");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_444_8BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_444_8BITS) {
|
|
g_value_set_static_string (&val, "main-444");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_444_10BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_444_10BITS) {
|
|
g_value_set_static_string (&val, "main-444-10");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
if ((flags & GST_NV_DECODER_FORMAT_FLAG_444_12BITS) ==
|
|
GST_NV_DECODER_FORMAT_FLAG_444_12BITS) {
|
|
g_value_set_static_string (&val, "main-444-12");
|
|
gst_value_list_append_value (profiles, &val);
|
|
}
|
|
|
|
ret = TRUE;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
g_value_unset (&val);
|
|
|
|
return ret;
|
|
}
|
|
|
|
typedef struct
|
|
{
|
|
cudaVideoCodec codec;
|
|
const gchar *codec_name;
|
|
const gchar *sink_caps_string;
|
|
} GstNvdecoderCodecMap;
|
|
|
|
const GstNvdecoderCodecMap codec_map_list[] = {
|
|
{cudaVideoCodec_MPEG1, "mpegvideo",
|
|
"video/mpeg, mpegversion = (int) 1, systemstream = (boolean) false"},
|
|
{cudaVideoCodec_MPEG2, "mpeg2video",
|
|
"video/mpeg, mpegversion = (int) 2, systemstream = (boolean) false"},
|
|
{cudaVideoCodec_MPEG4, "mpeg4video",
|
|
"video/mpeg, mpegversion = (int) 4, systemstream = (boolean) false"},
|
|
#if 0
|
|
/* FIXME: need verification */
|
|
{cudaVideoCodec_VC1, "vc1"},
|
|
#endif
|
|
/* NOTE: common supported h264 profiles for all GPU architecture
|
|
* 4:2:0, baseline, main, and high profiles
|
|
*/
|
|
{cudaVideoCodec_H264, "h264",
|
|
"video/x-h264, stream-format = (string) byte-stream"
|
|
", alignment = (string) au"
|
|
", profile = (string) { constrained-baseline, baseline, main, high }"},
|
|
{cudaVideoCodec_JPEG, "jpeg", "image/jpeg"},
|
|
#if 0
|
|
/* FIXME: need verification */
|
|
{cudaVideoCodec_H264_SVC, "h264svc"},
|
|
{cudaVideoCodec_H264_MVC, "h264mvc"},
|
|
#endif
|
|
{cudaVideoCodec_HEVC, "h265",
|
|
"video/x-h265, stream-format = (string) byte-stream"
|
|
", alignment = (string) au, profile = (string) { main }"},
|
|
{cudaVideoCodec_VP8, "vp8", "video/x-vp8"},
|
|
{cudaVideoCodec_VP9, "vp9", "video/x-vp9"}
|
|
};
|
|
|
|
gboolean
|
|
gst_nv_decoder_check_device_caps (CUcontext cuda_ctx, cudaVideoCodec codec,
|
|
GstCaps ** sink_template, GstCaps ** src_template)
|
|
{
|
|
CUresult cuda_ret;
|
|
gint max_width = 0, min_width = G_MAXINT;
|
|
gint max_height = 0, min_height = G_MAXINT;
|
|
GstCaps *sink_templ = NULL;
|
|
GstCaps *src_templ = NULL;
|
|
/* FIXME: support 12bits format */
|
|
guint bitdepth_minus8[3] = { 0, 2, 4 };
|
|
GstNvDecoderFormatFlags format_flags = 0;
|
|
gint c_idx, b_idx;
|
|
guint num_support = 0;
|
|
cudaVideoChromaFormat chroma_list[] = {
|
|
#if 0
|
|
/* FIXME: support monochrome */
|
|
cudaVideoChromaFormat_Monochrome,
|
|
/* FIXME: Can our OpenGL support NV16 and its 10/12bits variant?? */
|
|
cudaVideoChromaFormat_422,
|
|
#endif
|
|
cudaVideoChromaFormat_420,
|
|
cudaVideoChromaFormat_444,
|
|
};
|
|
GValue format_list = G_VALUE_INIT;
|
|
GValue format = G_VALUE_INIT;
|
|
GValue profile_list = G_VALUE_INIT;
|
|
const GstNvdecoderCodecMap *codec_map = NULL;
|
|
guint i;
|
|
gboolean ret = FALSE;
|
|
|
|
for (i = 0; i < G_N_ELEMENTS (codec_map_list); i++) {
|
|
if (codec_map_list[i].codec == codec) {
|
|
codec_map = &codec_map_list[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (!codec_map) {
|
|
GST_INFO ("No codec map corresponding to codec %d", codec);
|
|
return FALSE;
|
|
}
|
|
|
|
if (!gst_cuvid_can_get_decoder_caps ()) {
|
|
GST_INFO ("Too old nvidia driver to query decoder capability");
|
|
|
|
src_templ = gst_caps_from_string (GST_VIDEO_CAPS_MAKE ("NV12"));
|
|
|
|
#if HAVE_NVCODEC_GST_GL
|
|
{
|
|
GstCaps *gl_caps = gst_caps_copy (src_templ);
|
|
gst_caps_set_features_simple (gl_caps,
|
|
gst_caps_features_from_string (GST_CAPS_FEATURE_MEMORY_GL_MEMORY));
|
|
gst_caps_append (src_templ, gl_caps);
|
|
}
|
|
#endif
|
|
|
|
sink_templ = gst_caps_from_string (codec_map->sink_caps_string);
|
|
|
|
*src_template = src_templ;
|
|
*sink_template = sink_templ;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
g_value_init (&format_list, GST_TYPE_LIST);
|
|
g_value_init (&format, G_TYPE_STRING);
|
|
g_value_init (&profile_list, GST_TYPE_LIST);
|
|
|
|
if (CuCtxPushCurrent (cuda_ctx) != CUDA_SUCCESS)
|
|
goto done;
|
|
|
|
for (c_idx = 0; c_idx < G_N_ELEMENTS (chroma_list); c_idx++) {
|
|
for (b_idx = 0; b_idx < G_N_ELEMENTS (bitdepth_minus8); b_idx++) {
|
|
CUVIDDECODECAPS decoder_caps = { 0, };
|
|
GstNvDecoderFormatFlags cur_flag = 0;
|
|
|
|
decoder_caps.eCodecType = codec;
|
|
decoder_caps.eChromaFormat = chroma_list[c_idx];
|
|
decoder_caps.nBitDepthMinus8 = bitdepth_minus8[b_idx];
|
|
|
|
cuda_ret = CuvidGetDecoderCaps (&decoder_caps);
|
|
if (cuda_ret != CUDA_SUCCESS) {
|
|
GST_INFO ("could not query %s decoder capability, ret %d",
|
|
codec_map->codec_name, cuda_ret);
|
|
continue;
|
|
} else if (!decoder_caps.bIsSupported) {
|
|
GST_LOG ("%s bit-depth %d with chroma format %d is not supported",
|
|
codec_map->codec_name, bitdepth_minus8[b_idx] + 8, c_idx);
|
|
continue;
|
|
}
|
|
|
|
if (min_width > decoder_caps.nMinWidth)
|
|
min_width = decoder_caps.nMinWidth;
|
|
if (min_height > decoder_caps.nMinHeight)
|
|
min_height = decoder_caps.nMinHeight;
|
|
if (max_width < decoder_caps.nMaxWidth)
|
|
max_width = decoder_caps.nMaxWidth;
|
|
if (max_height < decoder_caps.nMaxHeight)
|
|
max_height = decoder_caps.nMaxHeight;
|
|
|
|
if (chroma_list[c_idx] == cudaVideoChromaFormat_420)
|
|
cur_flag = GST_NV_DECODER_FORMAT_FLAG_420_8BITS;
|
|
else
|
|
cur_flag = GST_NV_DECODER_FORMAT_FLAG_444_8BITS;
|
|
|
|
format_flags |= (cur_flag << (bitdepth_minus8[b_idx] / 2));
|
|
|
|
GST_INFO ("%s bit-depth %d with chroma format %d [%d - %d] x [%d - %d]",
|
|
codec_map->codec_name, bitdepth_minus8[b_idx] + 8, c_idx, min_width,
|
|
max_width, min_height, max_height);
|
|
|
|
switch (chroma_list[c_idx]) {
|
|
case cudaVideoChromaFormat_420:
|
|
if (bitdepth_minus8[b_idx] == 0) {
|
|
g_value_set_string (&format, "NV12");
|
|
} else if (bitdepth_minus8[b_idx] == 2) {
|
|
#if G_BYTE_ORDER == G_LITTLE_ENDIAN
|
|
g_value_set_string (&format, "P010_10LE");
|
|
#else
|
|
g_value_set_string (&format, "P010_10BE");
|
|
#endif
|
|
} else if (bitdepth_minus8[b_idx] == 4) {
|
|
#if G_BYTE_ORDER == G_LITTLE_ENDIAN
|
|
g_value_set_string (&format, "P016_LE");
|
|
#else
|
|
g_value_set_string (&format, "P016_BE");
|
|
#endif
|
|
} else {
|
|
GST_WARNING ("unhandled bitdepth %d", bitdepth_minus8[b_idx] + 8);
|
|
break;
|
|
}
|
|
num_support++;
|
|
gst_value_list_append_value (&format_list, &format);
|
|
break;
|
|
case cudaVideoChromaFormat_444:
|
|
if (cudaVideoCodec_JPEG == codec) {
|
|
/* NVDEC jpeg decoder can decode 4:4:4 format
|
|
* but it produces 4:2:0 frame */
|
|
break;
|
|
}
|
|
|
|
if (bitdepth_minus8[b_idx] == 0) {
|
|
g_value_set_string (&format, "Y444");
|
|
} else if (bitdepth_minus8[b_idx] == 2 || bitdepth_minus8[b_idx] == 4) {
|
|
#if G_BYTE_ORDER == G_LITTLE_ENDIAN
|
|
g_value_set_string (&format, "Y444_16LE");
|
|
#else
|
|
g_value_set_string (&format, "Y444_16BE");
|
|
#endif
|
|
} else {
|
|
GST_WARNING ("unhandled bitdepth %d", bitdepth_minus8[b_idx] + 8);
|
|
break;
|
|
}
|
|
num_support++;
|
|
gst_value_list_append_value (&format_list, &format);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
if (num_support == 0) {
|
|
GST_INFO ("device can not support %s", codec_map->codec_name);
|
|
goto done;
|
|
}
|
|
|
|
src_templ = gst_caps_new_simple ("video/x-raw",
|
|
"width", GST_TYPE_INT_RANGE, min_width, max_width,
|
|
"height", GST_TYPE_INT_RANGE, min_height, max_height,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL);
|
|
|
|
gst_caps_set_value (src_templ, "format", &format_list);
|
|
|
|
/* OpenGL specific */
|
|
#if HAVE_NVCODEC_GST_GL
|
|
{
|
|
GstCaps *gl_caps = gst_caps_copy (src_templ);
|
|
gst_caps_set_features_simple (gl_caps,
|
|
gst_caps_features_from_string (GST_CAPS_FEATURE_MEMORY_GL_MEMORY));
|
|
gst_caps_append (src_templ, gl_caps);
|
|
}
|
|
#endif
|
|
|
|
sink_templ = gst_caps_from_string (codec_map->sink_caps_string);
|
|
gst_caps_set_simple (sink_templ,
|
|
"width", GST_TYPE_INT_RANGE, min_width, max_width,
|
|
"height", GST_TYPE_INT_RANGE, min_height, max_height, NULL);
|
|
|
|
if (gst_nv_decocer_get_supported_codec_profiles (&profile_list, codec,
|
|
format_flags)) {
|
|
gst_caps_set_value (sink_templ, "profile", &profile_list);
|
|
}
|
|
|
|
GST_DEBUG ("sink template caps %" GST_PTR_FORMAT, sink_templ);
|
|
GST_DEBUG ("src template caps %" GST_PTR_FORMAT, src_templ);
|
|
|
|
CuCtxPopCurrent (NULL);
|
|
|
|
done:
|
|
g_value_unset (&format_list);
|
|
g_value_unset (&format);
|
|
g_value_unset (&profile_list);
|
|
|
|
if (!sink_templ || !src_templ) {
|
|
gst_clear_caps (&sink_templ);
|
|
gst_clear_caps (&src_templ);
|
|
|
|
ret = FALSE;
|
|
} else {
|
|
/* class data will be leaked if the element never gets instantiated */
|
|
GST_MINI_OBJECT_FLAG_SET (src_templ, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED);
|
|
GST_MINI_OBJECT_FLAG_SET (sink_templ, GST_MINI_OBJECT_FLAG_MAY_BE_LEAKED);
|
|
|
|
*src_template = src_templ;
|
|
*sink_template = sink_templ;
|
|
|
|
ret = TRUE;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
const gchar *
|
|
gst_cuda_video_codec_to_string (cudaVideoCodec codec)
|
|
{
|
|
gint i;
|
|
|
|
for (i = 0; i < G_N_ELEMENTS (codec_map_list); i++) {
|
|
if (codec_map_list[i].codec == codec)
|
|
return codec_map_list[i].codec_name;
|
|
}
|
|
|
|
return "unknown";
|
|
}
|