mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-25 17:50:36 +00:00
c4cb51c63d
New video capture implementation using WinRT Media APIs for UWP app. Due to the strict permission policy of UWP, device enumeration and open should be done via new WinRT APIs and to get permission from users, it will invoke permission dialog on UI. Strictly saying, this implementation is not a part of MediaFoundation but structurally it's very similar to MediaFoundation API. So we can avoid some code duplication by adding this implementation into MediaFoundation plugin. This implementation requires UniversalApiContract version >= 6.0 which is part of Windows 10 version 1803 (Redstone 4) Part-of: <https://gitlab.freedesktop.org/gstreamer/gst-plugins-bad/-/merge_requests/1241>
391 lines
11 KiB
C
391 lines
11 KiB
C
/* GStreamer
|
|
* Copyright (C) 2019 Seungha Yang <seungha.yang@navercorp.com>
|
|
* Copyright (C) 2020 Seungha Yang <seungha@centricular.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
/**
|
|
* SECTION:element-mfvideosrc
|
|
* @title: mfvideosrc
|
|
*
|
|
* Provides video capture from the Microsoft Media Foundation API.
|
|
*
|
|
* ## Example pipelines
|
|
* |[
|
|
* gst-launch-1.0 -v mfvideosrc ! fakesink
|
|
* ]| Capture from the default video capture device and render to fakesink.
|
|
*
|
|
* |[
|
|
* gst-launch-1.0 -v mfvideosrc device-index=1 ! fakesink
|
|
* ]| Capture from the second video device (if available) and render to fakesink.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include "gstmfconfig.h"
|
|
|
|
#include "gstmfvideosrc.h"
|
|
#include "gstmfutils.h"
|
|
#if GST_MF_WINAPI_ONLY_APP
|
|
#include "gstmfcapturewinrt.h"
|
|
#else /* GST_MF_WINAPI_ONLY_APP */
|
|
#include "gstmfsourcereader.h"
|
|
#if GST_MF_HAVE_CAPTURE_ENGINE
|
|
#include "gstmfcaptureengine.h"
|
|
#endif /* GST_MF_HAVE_CAPTURE_ENGINE */
|
|
#endif /* !GST_MF_WINAPI_ONLY_APP */
|
|
#include <string.h>
|
|
|
|
GST_DEBUG_CATEGORY (gst_mf_video_src_debug);
|
|
#define GST_CAT_DEFAULT gst_mf_video_src_debug
|
|
|
|
static GstStaticPadTemplate src_template = GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (GST_MF_VIDEO_FORMATS))
|
|
);
|
|
|
|
struct _GstMFVideoSrc
|
|
{
|
|
GstPushSrc parent;
|
|
|
|
GstMFSourceObject *source;
|
|
gboolean started;
|
|
GstVideoInfo info;
|
|
|
|
GstClockTime first_pts;
|
|
guint64 n_frames;
|
|
|
|
/* properties */
|
|
gchar *device_path;
|
|
gchar *device_name;
|
|
gint device_index;
|
|
};
|
|
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_DEVICE_PATH,
|
|
PROP_DEVICE_NAME,
|
|
PROP_DEVICE_INDEX,
|
|
};
|
|
|
|
#define DEFAULT_DEVICE_PATH NULL
|
|
#define DEFAULT_DEVICE_NAME NULL
|
|
#define DEFAULT_DEVICE_INDEX -1
|
|
|
|
static void gst_mf_video_src_finalize (GObject * object);
|
|
static void gst_mf_video_src_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec);
|
|
static void gst_mf_video_src_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec);
|
|
|
|
static gboolean gst_mf_video_src_start (GstBaseSrc * src);
|
|
static gboolean gst_mf_video_src_stop (GstBaseSrc * src);
|
|
static gboolean gst_mf_video_src_set_caps (GstBaseSrc * src, GstCaps * caps);
|
|
static GstCaps *gst_mf_video_src_get_caps (GstBaseSrc * src, GstCaps * filter);
|
|
static GstCaps *gst_mf_video_src_fixate (GstBaseSrc * src, GstCaps * caps);
|
|
static gboolean gst_mf_video_src_unlock (GstBaseSrc * src);
|
|
static gboolean gst_mf_video_src_unlock_stop (GstBaseSrc * src);
|
|
|
|
static GstFlowReturn gst_mf_video_src_fill (GstPushSrc * pushsrc,
|
|
GstBuffer * buffer);
|
|
|
|
#define gst_mf_video_src_parent_class parent_class
|
|
G_DEFINE_TYPE (GstMFVideoSrc, gst_mf_video_src, GST_TYPE_PUSH_SRC);
|
|
|
|
static void
|
|
gst_mf_video_src_class_init (GstMFVideoSrcClass * klass)
|
|
{
|
|
GObjectClass *gobject_class = G_OBJECT_CLASS (klass);
|
|
GstElementClass *element_class = GST_ELEMENT_CLASS (klass);
|
|
GstBaseSrcClass *basesrc_class = GST_BASE_SRC_CLASS (klass);
|
|
GstPushSrcClass *pushsrc_class = GST_PUSH_SRC_CLASS (klass);
|
|
|
|
gobject_class->finalize = gst_mf_video_src_finalize;
|
|
gobject_class->get_property = gst_mf_video_src_get_property;
|
|
gobject_class->set_property = gst_mf_video_src_set_property;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_DEVICE_PATH,
|
|
g_param_spec_string ("device-path", "Device Path",
|
|
"The device path", DEFAULT_DEVICE_PATH,
|
|
G_PARAM_READWRITE | GST_PARAM_MUTABLE_READY |
|
|
G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_DEVICE_NAME,
|
|
g_param_spec_string ("device-name", "Device Name",
|
|
"The human-readable device name", DEFAULT_DEVICE_NAME,
|
|
G_PARAM_READWRITE | GST_PARAM_MUTABLE_READY |
|
|
G_PARAM_STATIC_STRINGS));
|
|
g_object_class_install_property (gobject_class, PROP_DEVICE_INDEX,
|
|
g_param_spec_int ("device-index", "Device Index",
|
|
"The zero-based device index", -1, G_MAXINT, DEFAULT_DEVICE_INDEX,
|
|
G_PARAM_READWRITE | GST_PARAM_MUTABLE_READY |
|
|
G_PARAM_STATIC_STRINGS));
|
|
|
|
gst_element_class_set_static_metadata (element_class,
|
|
"Media Foundation Video Source",
|
|
"Source/Video/Hardware",
|
|
"Capture video stream through Windows Media Foundation",
|
|
"Seungha Yang <seungha.yang@navercorp.com>");
|
|
|
|
gst_element_class_add_static_pad_template (element_class, &src_template);
|
|
|
|
basesrc_class->start = GST_DEBUG_FUNCPTR (gst_mf_video_src_start);
|
|
basesrc_class->stop = GST_DEBUG_FUNCPTR (gst_mf_video_src_stop);
|
|
basesrc_class->set_caps = GST_DEBUG_FUNCPTR (gst_mf_video_src_set_caps);
|
|
basesrc_class->get_caps = GST_DEBUG_FUNCPTR (gst_mf_video_src_get_caps);
|
|
basesrc_class->fixate = GST_DEBUG_FUNCPTR (gst_mf_video_src_fixate);
|
|
basesrc_class->unlock = GST_DEBUG_FUNCPTR (gst_mf_video_src_unlock);
|
|
basesrc_class->unlock_stop = GST_DEBUG_FUNCPTR (gst_mf_video_src_unlock_stop);
|
|
|
|
pushsrc_class->fill = GST_DEBUG_FUNCPTR (gst_mf_video_src_fill);
|
|
|
|
GST_DEBUG_CATEGORY_INIT (gst_mf_video_src_debug, "mfvideosrc", 0,
|
|
"mfvideosrc");
|
|
}
|
|
|
|
static void
|
|
gst_mf_video_src_init (GstMFVideoSrc * self)
|
|
{
|
|
gst_base_src_set_format (GST_BASE_SRC (self), GST_FORMAT_TIME);
|
|
gst_base_src_set_live (GST_BASE_SRC (self), TRUE);
|
|
gst_base_src_set_do_timestamp (GST_BASE_SRC (self), TRUE);
|
|
|
|
self->device_index = DEFAULT_DEVICE_INDEX;
|
|
}
|
|
|
|
static void
|
|
gst_mf_video_src_finalize (GObject * object)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (object);
|
|
|
|
g_free (self->device_name);
|
|
g_free (self->device_path);
|
|
|
|
G_OBJECT_CLASS (parent_class)->finalize (object);
|
|
}
|
|
|
|
static void
|
|
gst_mf_video_src_get_property (GObject * object, guint prop_id, GValue * value,
|
|
GParamSpec * pspec)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE_PATH:
|
|
g_value_set_string (value, self->device_path);
|
|
break;
|
|
case PROP_DEVICE_NAME:
|
|
g_value_set_string (value, self->device_name);
|
|
break;
|
|
case PROP_DEVICE_INDEX:
|
|
g_value_set_int (value, self->device_index);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_mf_video_src_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE_PATH:
|
|
g_free (self->device_path);
|
|
self->device_path = g_value_dup_string (value);
|
|
break;
|
|
case PROP_DEVICE_NAME:
|
|
g_free (self->device_name);
|
|
self->device_name = g_value_dup_string (value);
|
|
break;
|
|
case PROP_DEVICE_INDEX:
|
|
self->device_index = g_value_get_int (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_mf_video_src_start (GstBaseSrc * src)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (src);
|
|
|
|
GST_DEBUG_OBJECT (self, "Start");
|
|
|
|
#if GST_MF_WINAPI_ONLY_APP
|
|
self->source = gst_mf_capture_winrt_new (GST_MF_SOURCE_TYPE_VIDEO,
|
|
self->device_index, self->device_name, self->device_path);
|
|
#else /* GST_MF_WINAPI_ONLY_APP */
|
|
#if GST_MF_HAVE_CAPTURE_ENGINE
|
|
if (!self->source)
|
|
self->source = gst_mf_capture_engine_new (GST_MF_SOURCE_TYPE_VIDEO,
|
|
self->device_index, self->device_name, self->device_path);
|
|
#endif /* GST_MF_HAVE_CAPTURE_ENGINE */
|
|
if (!self->source)
|
|
self->source = gst_mf_source_reader_new (GST_MF_SOURCE_TYPE_VIDEO,
|
|
self->device_index, self->device_name, self->device_path);
|
|
#endif /* GST_MF_WINAPI_ONLY_APP */
|
|
|
|
self->first_pts = GST_CLOCK_TIME_NONE;
|
|
self->n_frames = 0;
|
|
|
|
return ! !self->source;
|
|
}
|
|
|
|
static gboolean
|
|
gst_mf_video_src_stop (GstBaseSrc * src)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (src);
|
|
|
|
GST_DEBUG_OBJECT (self, "Stop");
|
|
|
|
if (self->source) {
|
|
gst_mf_source_object_stop (self->source);
|
|
gst_object_unref (self->source);
|
|
self->source = NULL;
|
|
}
|
|
|
|
self->started = FALSE;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_mf_video_src_set_caps (GstBaseSrc * src, GstCaps * caps)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (src);
|
|
|
|
GST_DEBUG_OBJECT (self, "Set caps %" GST_PTR_FORMAT, caps);
|
|
|
|
if (!self->source) {
|
|
GST_ERROR_OBJECT (self, "No capture engine yet");
|
|
return FALSE;
|
|
}
|
|
|
|
if (!gst_mf_source_object_set_caps (self->source, caps)) {
|
|
GST_ERROR_OBJECT (self, "CaptureEngine couldn't accept caps");
|
|
return FALSE;
|
|
}
|
|
|
|
gst_video_info_from_caps (&self->info, caps);
|
|
gst_base_src_set_blocksize (src, GST_VIDEO_INFO_SIZE (&self->info));
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_mf_video_src_get_caps (GstBaseSrc * src, GstCaps * filter)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (src);
|
|
GstCaps *caps = NULL;
|
|
|
|
if (self->source)
|
|
caps = gst_mf_source_object_get_caps (self->source);
|
|
|
|
if (!caps)
|
|
caps = gst_pad_get_pad_template_caps (GST_BASE_SRC_PAD (src));
|
|
|
|
if (filter) {
|
|
GstCaps *filtered =
|
|
gst_caps_intersect_full (filter, caps, GST_CAPS_INTERSECT_FIRST);
|
|
gst_caps_unref (caps);
|
|
caps = filtered;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (self, "Returning caps %" GST_PTR_FORMAT, caps);
|
|
|
|
return caps;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_mf_video_src_fixate (GstBaseSrc * src, GstCaps * caps)
|
|
{
|
|
GstStructure *structure;
|
|
GstCaps *fixated_caps;
|
|
gint i;
|
|
|
|
fixated_caps = gst_caps_make_writable (caps);
|
|
|
|
for (i = 0; i < gst_caps_get_size (fixated_caps); ++i) {
|
|
structure = gst_caps_get_structure (fixated_caps, i);
|
|
gst_structure_fixate_field_nearest_int (structure, "width", G_MAXINT);
|
|
gst_structure_fixate_field_nearest_int (structure, "height", G_MAXINT);
|
|
gst_structure_fixate_field_nearest_fraction (structure, "framerate",
|
|
G_MAXINT, 1);
|
|
}
|
|
|
|
fixated_caps = gst_caps_fixate (fixated_caps);
|
|
|
|
return fixated_caps;
|
|
}
|
|
|
|
static gboolean
|
|
gst_mf_video_src_unlock (GstBaseSrc * src)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (src);
|
|
|
|
if (self->source)
|
|
gst_mf_source_object_set_flushing (self->source, TRUE);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_mf_video_src_unlock_stop (GstBaseSrc * src)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (src);
|
|
|
|
if (self->source)
|
|
gst_mf_source_object_set_flushing (self->source, FALSE);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_mf_video_src_fill (GstPushSrc * pushsrc, GstBuffer * buffer)
|
|
{
|
|
GstMFVideoSrc *self = GST_MF_VIDEO_SRC (pushsrc);
|
|
GstFlowReturn ret = GST_FLOW_OK;
|
|
|
|
if (!self->started) {
|
|
if (!gst_mf_source_object_start (self->source)) {
|
|
GST_ERROR_OBJECT (self, "Failed to start capture object");
|
|
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
|
|
self->started = TRUE;
|
|
}
|
|
|
|
ret = gst_mf_source_object_fill (self->source, buffer);
|
|
if (ret != GST_FLOW_OK)
|
|
return ret;
|
|
|
|
GST_BUFFER_OFFSET (buffer) = self->n_frames;
|
|
GST_BUFFER_OFFSET_END (buffer) = GST_BUFFER_OFFSET (buffer) + 1;
|
|
self->n_frames++;
|
|
|
|
return GST_FLOW_OK;
|
|
}
|