mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-03 16:09:39 +00:00
969ab3e664
Output texture of d3d11 decoder cannot have the bind flag D3D11_BIND_SHADER_RESOURCE (meaning that it cannot be used for shader input resource). So d3d11convert (and it's subclasses) was copying texture into another internal texture to use d3d11 shader. It's obviously overhead and we can avoid texture copy for colorspace conversion or resizing via ID3D11VideoProcessor as it supports decoder output texture. This commit would be a visible optimization for d3d11 decoder with d3d11compositor use case because we can avoid texture copy per frame. Part-of: <https://gitlab.freedesktop.org/gstreamer/gst-plugins-bad/-/merge_requests/1718>
543 lines
16 KiB
C
543 lines
16 KiB
C
/* GStreamer
|
|
* Copyright (C) <2020> Seungha Yang <seungha.yang@navercorp.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
# include <config.h>
|
|
#endif
|
|
|
|
#include "gstd3d11device.h"
|
|
#include "gstd3d11utils.h"
|
|
#include "gstd3d11format.h"
|
|
#include "gstd3d11memory.h"
|
|
#include "gstd3d11videoprocessor.h"
|
|
|
|
#include <string.h>
|
|
|
|
GST_DEBUG_CATEGORY_EXTERN (gst_d3d11_video_processor_debug);
|
|
#define GST_CAT_DEFAULT gst_d3d11_video_processor_debug
|
|
|
|
#if (D3D11_HEADER_VERSION >= 1 && DXGI_HEADER_VERSION >= 4)
|
|
#define HAVE_VIDEO_CONTEXT_ONE
|
|
#endif
|
|
|
|
#if (D3D11_HEADER_VERSION >= 4) && (DXGI_HEADER_VERSION >= 5)
|
|
#define HAVE_VIDEO_CONTEXT_TWO
|
|
#endif
|
|
|
|
struct _GstD3D11VideoProcessor
|
|
{
|
|
GstD3D11Device *device;
|
|
|
|
ID3D11VideoDevice *video_device;
|
|
ID3D11VideoContext *video_context;
|
|
#ifdef HAVE_VIDEO_CONTEXT_ONE
|
|
ID3D11VideoContext1 *video_context1;
|
|
#endif
|
|
#ifdef HAVE_VIDEO_CONTEXT_TWO
|
|
ID3D11VideoContext2 *video_context2;
|
|
#endif
|
|
ID3D11VideoProcessor *processor;
|
|
ID3D11VideoProcessorEnumerator *enumerator;
|
|
|
|
D3D11_VIDEO_PROCESSOR_CAPS processor_caps;
|
|
};
|
|
|
|
GstD3D11VideoProcessor *
|
|
gst_d3d11_video_processor_new (GstD3D11Device * device, guint in_width,
|
|
guint in_height, guint out_width, guint out_height)
|
|
{
|
|
GstD3D11VideoProcessor *self;
|
|
ID3D11Device *device_handle;
|
|
ID3D11DeviceContext *context_handle;
|
|
HRESULT hr;
|
|
D3D11_VIDEO_PROCESSOR_CONTENT_DESC desc = { 0, };
|
|
|
|
g_return_val_if_fail (GST_IS_D3D11_DEVICE (device), NULL);
|
|
|
|
device_handle = gst_d3d11_device_get_device_handle (device);
|
|
context_handle = gst_d3d11_device_get_device_context_handle (device);
|
|
|
|
self = g_new0 (GstD3D11VideoProcessor, 1);
|
|
self->device = gst_object_ref (device);
|
|
|
|
hr = ID3D11Device_QueryInterface (device_handle,
|
|
&IID_ID3D11VideoDevice, (void **) &self->video_device);
|
|
if (!gst_d3d11_result (hr, device))
|
|
goto fail;
|
|
|
|
hr = ID3D11DeviceContext_QueryInterface (context_handle,
|
|
&IID_ID3D11VideoContext, (void **) &self->video_context);
|
|
if (!gst_d3d11_result (hr, device))
|
|
goto fail;
|
|
|
|
/* FIXME: Add support intelace */
|
|
desc.InputFrameFormat = D3D11_VIDEO_FRAME_FORMAT_PROGRESSIVE;
|
|
desc.InputWidth = in_width;
|
|
desc.InputHeight = in_height;
|
|
desc.OutputWidth = out_width;
|
|
desc.OutputHeight = out_height;
|
|
/* TODO: make option for this */
|
|
desc.Usage = D3D11_VIDEO_USAGE_PLAYBACK_NORMAL;
|
|
|
|
hr = ID3D11VideoDevice_CreateVideoProcessorEnumerator (self->video_device,
|
|
&desc, &self->enumerator);
|
|
if (!gst_d3d11_result (hr, device))
|
|
goto fail;
|
|
|
|
hr = ID3D11VideoProcessorEnumerator_GetVideoProcessorCaps (self->enumerator,
|
|
&self->processor_caps);
|
|
if (!gst_d3d11_result (hr, device))
|
|
goto fail;
|
|
|
|
hr = ID3D11VideoDevice_CreateVideoProcessor (self->video_device,
|
|
self->enumerator, 0, &self->processor);
|
|
if (!gst_d3d11_result (hr, device))
|
|
goto fail;
|
|
|
|
#ifdef HAVE_VIDEO_CONTEXT_ONE
|
|
hr = ID3D11VideoContext_QueryInterface (self->video_context,
|
|
&IID_ID3D11VideoContext1, (void **) &self->video_context1);
|
|
if (gst_d3d11_result (hr, device)) {
|
|
GST_DEBUG ("ID3D11VideoContext1 interface available");
|
|
}
|
|
#endif
|
|
#ifdef HAVE_VIDEO_CONTEXT_TWO
|
|
hr = ID3D11VideoContext_QueryInterface (self->video_context,
|
|
&IID_ID3D11VideoContext2, (void **) &self->video_context2);
|
|
if (gst_d3d11_result (hr, device)) {
|
|
GST_DEBUG ("ID3D11VideoContext2 interface available");
|
|
}
|
|
#endif
|
|
|
|
return self;
|
|
|
|
fail:
|
|
gst_d3d11_video_processor_free (self);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
void
|
|
gst_d3d11_video_processor_free (GstD3D11VideoProcessor * processor)
|
|
{
|
|
g_return_if_fail (processor != NULL);
|
|
|
|
if (processor->video_device)
|
|
ID3D11VideoDevice_Release (processor->video_device);
|
|
if (processor->video_context)
|
|
ID3D11VideoContext_Release (processor->video_context);
|
|
#ifdef HAVE_VIDEO_CONTEXT_ONE
|
|
if (processor->video_context1)
|
|
ID3D11VideoContext1_Release (processor->video_context1);
|
|
#endif
|
|
#ifdef HAVE_VIDEO_CONTEXT_TWO
|
|
if (processor->video_context2)
|
|
ID3D11VideoContext2_Release (processor->video_context2);
|
|
#endif
|
|
if (processor->processor)
|
|
ID3D11VideoProcessor_Release (processor->processor);
|
|
if (processor->enumerator)
|
|
ID3D11VideoProcessorEnumerator_Release (processor->enumerator);
|
|
|
|
gst_clear_object (&processor->device);
|
|
g_free (processor);
|
|
}
|
|
|
|
static gboolean
|
|
gst_d3d11_video_processor_supports_format (GstD3D11VideoProcessor *
|
|
self, DXGI_FORMAT format, gboolean is_input)
|
|
{
|
|
HRESULT hr;
|
|
UINT flag;
|
|
|
|
if (is_input) {
|
|
/* D3D11_VIDEO_PROCESSOR_FORMAT_SUPPORT_INPUT, missing in mingw header */
|
|
flag = 1;
|
|
} else {
|
|
/* D3D11_VIDEO_PROCESSOR_FORMAT_SUPPORT_OUTPUT, missing in mingw header */
|
|
flag = 2;
|
|
}
|
|
|
|
hr = ID3D11VideoProcessorEnumerator_CheckVideoProcessorFormat
|
|
(self->enumerator, format, &flag);
|
|
|
|
return gst_d3d11_result (hr, self->device);
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_supports_input_format (GstD3D11VideoProcessor *
|
|
processor, DXGI_FORMAT format)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
|
|
if (format == DXGI_FORMAT_UNKNOWN)
|
|
return FALSE;
|
|
|
|
return gst_d3d11_video_processor_supports_format (processor, format, TRUE);
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_supports_output_format (GstD3D11VideoProcessor *
|
|
processor, DXGI_FORMAT format)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
|
|
if (format == DXGI_FORMAT_UNKNOWN)
|
|
return FALSE;
|
|
|
|
return gst_d3d11_video_processor_supports_format (processor, format, FALSE);
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_get_caps (GstD3D11VideoProcessor * processor,
|
|
D3D11_VIDEO_PROCESSOR_CAPS * caps)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (caps != NULL, FALSE);
|
|
|
|
*caps = processor->processor_caps;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static void
|
|
video_processor_color_space_from_gst (GstD3D11VideoProcessor * self,
|
|
GstVideoColorimetry * color, D3D11_VIDEO_PROCESSOR_COLOR_SPACE * colorspace)
|
|
{
|
|
/* D3D11_VIDEO_PROCESSOR_DEVICE_CAPS_xvYCC */
|
|
UINT can_xvYCC = 2;
|
|
|
|
/* 0: playback, 1: video processing */
|
|
colorspace->Usage = 0;
|
|
|
|
if (color->range == GST_VIDEO_COLOR_RANGE_0_255) {
|
|
colorspace->RGB_Range = 0;
|
|
colorspace->Nominal_Range = D3D11_VIDEO_PROCESSOR_NOMINAL_RANGE_0_255;
|
|
} else {
|
|
/* 16-235 */
|
|
colorspace->RGB_Range = 1;
|
|
colorspace->Nominal_Range = D3D11_VIDEO_PROCESSOR_NOMINAL_RANGE_16_235;
|
|
}
|
|
|
|
if (color->matrix == GST_VIDEO_COLOR_MATRIX_BT601) {
|
|
colorspace->YCbCr_Matrix = 0;
|
|
} else {
|
|
/* BT.709, no other options (such as BT2020) */
|
|
colorspace->YCbCr_Matrix = 1;
|
|
}
|
|
|
|
if ((self->processor_caps.DeviceCaps & can_xvYCC) == can_xvYCC) {
|
|
colorspace->YCbCr_xvYCC = 1;
|
|
} else {
|
|
colorspace->YCbCr_xvYCC = 0;
|
|
}
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_set_input_color_space (GstD3D11VideoProcessor *
|
|
processor, GstVideoColorimetry * color)
|
|
{
|
|
D3D11_VIDEO_PROCESSOR_COLOR_SPACE color_space;
|
|
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (color != NULL, FALSE);
|
|
|
|
video_processor_color_space_from_gst (processor, color, &color_space);
|
|
|
|
ID3D11VideoContext_VideoProcessorSetStreamColorSpace
|
|
(processor->video_context, processor->processor, 0, &color_space);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_set_output_color_space (GstD3D11VideoProcessor *
|
|
processor, GstVideoColorimetry * color)
|
|
{
|
|
D3D11_VIDEO_PROCESSOR_COLOR_SPACE color_space;
|
|
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (color != NULL, FALSE);
|
|
|
|
video_processor_color_space_from_gst (processor, color, &color_space);
|
|
|
|
ID3D11VideoContext_VideoProcessorSetOutputColorSpace
|
|
(processor->video_context, processor->processor, &color_space);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
#if (DXGI_HEADER_VERSION >= 4)
|
|
gboolean
|
|
gst_d3d11_video_processor_set_input_dxgi_color_space (GstD3D11VideoProcessor *
|
|
processor, DXGI_COLOR_SPACE_TYPE color_space)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
|
|
#ifdef HAVE_VIDEO_CONTEXT_ONE
|
|
if (processor->video_context1) {
|
|
ID3D11VideoContext1_VideoProcessorSetStreamColorSpace1
|
|
(processor->video_context1, processor->processor, 0, color_space);
|
|
return TRUE;
|
|
}
|
|
#endif
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_set_output_dxgi_color_space (GstD3D11VideoProcessor *
|
|
processor, DXGI_COLOR_SPACE_TYPE color_space)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
|
|
#ifdef HAVE_VIDEO_CONTEXT_ONE
|
|
if (processor->video_context1) {
|
|
ID3D11VideoContext1_VideoProcessorSetOutputColorSpace1
|
|
(processor->video_context1, processor->processor, color_space);
|
|
return TRUE;
|
|
}
|
|
#endif
|
|
|
|
return FALSE;
|
|
}
|
|
#endif
|
|
|
|
#if (DXGI_HEADER_VERSION >= 5)
|
|
/* D3D11_VIDEO_PROCESSOR_FEATURE_CAPS_METADATA_HDR10
|
|
* missing in mingw header */
|
|
#define FEATURE_CAPS_METADATA_HDR10 (0x800)
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_set_input_hdr10_metadata (GstD3D11VideoProcessor *
|
|
processor, DXGI_HDR_METADATA_HDR10 * hdr10_meta)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
|
|
#ifdef HAVE_VIDEO_CONTEXT_TWO
|
|
if (processor->video_context2 && (processor->processor_caps.FeatureCaps &
|
|
FEATURE_CAPS_METADATA_HDR10)) {
|
|
if (hdr10_meta) {
|
|
ID3D11VideoContext2_VideoProcessorSetStreamHDRMetaData
|
|
(processor->video_context2, processor->processor, 0,
|
|
DXGI_HDR_METADATA_TYPE_HDR10, sizeof (DXGI_HDR_METADATA_HDR10),
|
|
hdr10_meta);
|
|
} else {
|
|
ID3D11VideoContext2_VideoProcessorSetStreamHDRMetaData
|
|
(processor->video_context2, processor->processor, 0,
|
|
DXGI_HDR_METADATA_TYPE_NONE, 0, NULL);
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
#endif
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_set_output_hdr10_metadata (GstD3D11VideoProcessor *
|
|
processor, DXGI_HDR_METADATA_HDR10 * hdr10_meta)
|
|
{
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
|
|
#ifdef HAVE_VIDEO_CONTEXT_TWO
|
|
if (processor->video_context2 && (processor->processor_caps.FeatureCaps &
|
|
FEATURE_CAPS_METADATA_HDR10)) {
|
|
if (hdr10_meta) {
|
|
ID3D11VideoContext2_VideoProcessorSetOutputHDRMetaData
|
|
(processor->video_context2, processor->processor,
|
|
DXGI_HDR_METADATA_TYPE_HDR10, sizeof (DXGI_HDR_METADATA_HDR10),
|
|
hdr10_meta);
|
|
} else {
|
|
ID3D11VideoContext2_VideoProcessorSetOutputHDRMetaData
|
|
(processor->video_context2, processor->processor,
|
|
DXGI_HDR_METADATA_TYPE_NONE, 0, NULL);
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
#endif
|
|
|
|
return FALSE;
|
|
}
|
|
#endif
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_create_input_view (GstD3D11VideoProcessor * processor,
|
|
D3D11_VIDEO_PROCESSOR_INPUT_VIEW_DESC * desc, ID3D11Resource * resource,
|
|
ID3D11VideoProcessorInputView ** view)
|
|
{
|
|
HRESULT hr;
|
|
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (desc != NULL, FALSE);
|
|
g_return_val_if_fail (resource != NULL, FALSE);
|
|
g_return_val_if_fail (view != NULL, FALSE);
|
|
|
|
hr = ID3D11VideoDevice_CreateVideoProcessorInputView (processor->video_device,
|
|
resource, processor->enumerator, desc, view);
|
|
if (!gst_d3d11_result (hr, processor->device))
|
|
return FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_ensure_input_view (GstD3D11VideoProcessor * processor,
|
|
GstD3D11Memory * mem)
|
|
{
|
|
return gst_d3d11_memory_ensure_processor_input_view (mem,
|
|
processor->video_device, processor->enumerator);
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_create_output_view (GstD3D11VideoProcessor *
|
|
processor, D3D11_VIDEO_PROCESSOR_OUTPUT_VIEW_DESC * desc,
|
|
ID3D11Resource * resource, ID3D11VideoProcessorOutputView ** view)
|
|
{
|
|
HRESULT hr;
|
|
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (desc != NULL, FALSE);
|
|
g_return_val_if_fail (resource != NULL, FALSE);
|
|
g_return_val_if_fail (view != NULL, FALSE);
|
|
|
|
hr = ID3D11VideoDevice_CreateVideoProcessorOutputView
|
|
(processor->video_device, resource, processor->enumerator, desc, view);
|
|
if (!gst_d3d11_result (hr, processor->device))
|
|
return FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_ensure_output_view (GstD3D11VideoProcessor *
|
|
processor, GstD3D11Memory * mem)
|
|
{
|
|
return gst_d3d11_memory_ensure_processor_output_view (mem,
|
|
processor->video_device, processor->enumerator);
|
|
}
|
|
|
|
void
|
|
gst_d3d11_video_processor_input_view_release (ID3D11VideoProcessorInputView *
|
|
view)
|
|
{
|
|
if (!view)
|
|
return;
|
|
|
|
ID3D11VideoProcessorInputView_Release (view);
|
|
}
|
|
|
|
void
|
|
gst_d3d11_video_processor_output_view_release (ID3D11VideoProcessorOutputView *
|
|
view)
|
|
{
|
|
if (!view)
|
|
return;
|
|
|
|
ID3D11VideoProcessorOutputView_Release (view);
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_render (GstD3D11VideoProcessor * processor,
|
|
RECT * in_rect, ID3D11VideoProcessorInputView * in_view,
|
|
RECT * out_rect, ID3D11VideoProcessorOutputView * out_view)
|
|
{
|
|
gboolean ret;
|
|
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (in_view != NULL, FALSE);
|
|
g_return_val_if_fail (out_view != NULL, FALSE);
|
|
|
|
gst_d3d11_device_lock (processor->device);
|
|
ret = gst_d3d11_video_processor_render_unlocked (processor, in_rect, in_view,
|
|
out_rect, out_view);
|
|
gst_d3d11_device_unlock (processor->device);
|
|
|
|
return ret;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_render_unlocked (GstD3D11VideoProcessor * processor,
|
|
RECT * in_rect, ID3D11VideoProcessorInputView * in_view,
|
|
RECT * out_rect, ID3D11VideoProcessorOutputView * out_view)
|
|
{
|
|
HRESULT hr;
|
|
D3D11_VIDEO_PROCESSOR_STREAM stream = { 0, };
|
|
|
|
g_return_val_if_fail (processor != NULL, FALSE);
|
|
g_return_val_if_fail (in_view != NULL, FALSE);
|
|
g_return_val_if_fail (out_view != NULL, FALSE);
|
|
|
|
stream.Enable = TRUE;
|
|
stream.pInputSurface = in_view;
|
|
|
|
if (in_rect) {
|
|
ID3D11VideoContext_VideoProcessorSetStreamSourceRect
|
|
(processor->video_context, processor->processor, 0, TRUE, in_rect);
|
|
} else {
|
|
ID3D11VideoContext_VideoProcessorSetStreamSourceRect
|
|
(processor->video_context, processor->processor, 0, FALSE, NULL);
|
|
}
|
|
|
|
if (out_rect) {
|
|
ID3D11VideoContext_VideoProcessorSetStreamDestRect
|
|
(processor->video_context, processor->processor, 0, TRUE, out_rect);
|
|
ID3D11VideoContext_VideoProcessorSetOutputTargetRect
|
|
(processor->video_context, processor->processor, TRUE, out_rect);
|
|
} else {
|
|
ID3D11VideoContext_VideoProcessorSetStreamDestRect
|
|
(processor->video_context, processor->processor, 0, FALSE, NULL);
|
|
ID3D11VideoContext_VideoProcessorSetOutputTargetRect
|
|
(processor->video_context, processor->processor, FALSE, NULL);
|
|
}
|
|
|
|
hr = ID3D11VideoContext_VideoProcessorBlt (processor->video_context,
|
|
processor->processor, out_view, 0, 1, &stream);
|
|
if (!gst_d3d11_result (hr, processor->device))
|
|
return FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_check_bind_flags_for_input_view (guint bind_flags)
|
|
{
|
|
static const guint compatible_flags = (D3D11_BIND_DECODER |
|
|
D3D11_BIND_VIDEO_ENCODER | D3D11_BIND_RENDER_TARGET |
|
|
D3D11_BIND_UNORDERED_ACCESS);
|
|
|
|
if (bind_flags == 0)
|
|
return TRUE;
|
|
|
|
if ((bind_flags & compatible_flags) != 0)
|
|
return TRUE;
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
gboolean
|
|
gst_d3d11_video_processor_check_bind_flags_for_output_view (guint bind_flags)
|
|
{
|
|
if ((bind_flags & D3D11_BIND_RENDER_TARGET) == D3D11_BIND_RENDER_TARGET)
|
|
return TRUE;
|
|
|
|
return FALSE;
|
|
}
|