mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-20 23:36:38 +00:00
209d6105d7
Original commit message from CVS: Patch by: Ole André Vadla Ravnås <ole.andre.ravnas@tandberg.com> * gst-libs/gst/dshow/gstdshowfakesink.cpp: (CDshowFakeSink.CDshowFakeSink): * gst-libs/gst/dshow/gstdshowfakesink.h: (CDshowFakeSink.m_hres): Fix crasher in constructor due to the base class's constructor not necessarily being NULL-safe (depends on the SDK version used apparently; #492406). * sys/dshowsrcwrapper/gstdshowaudiosrc.c: (gst_dshowaudiosrc_prepare): * sys/dshowsrcwrapper/gstdshowvideosrc.c: (gst_dshowvideosrc_set_caps): Fix a couple of MSVC compiler warnings (#492406).
998 lines
29 KiB
C
998 lines
29 KiB
C
/* GStreamer
|
|
* Copyright (C) 2007 Sebastien Moutte <sebastien@moutte.net>
|
|
*
|
|
* gstdshowvideosrc.c:
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
|
* Boston, MA 02111-1307, USA.
|
|
*/
|
|
|
|
#include "gstdshowvideosrc.h"
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
static const GstElementDetails gst_dshowvideosrc_details =
|
|
GST_ELEMENT_DETAILS ("DirectShow video capture source",
|
|
"Source/Video",
|
|
"Receive data from a directshow video capture graph",
|
|
"Sebastien Moutte <sebastien@moutte.net>");
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (dshowvideosrc_debug);
|
|
#define GST_CAT_DEFAULT dshowvideosrc_debug
|
|
|
|
const GUID MEDIASUBTYPE_I420
|
|
= { 0x30323449, 0x0000, 0x0010, {0x80, 0x00, 0x00, 0xAA, 0x00, 0x38, 0x9B,
|
|
0x71}
|
|
};
|
|
|
|
static GstStaticPadTemplate src_template = GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("video/x-raw-rgb,"
|
|
"bpp = (int) 24,"
|
|
"depth = (int) 24,"
|
|
"width = (int) [ 1, MAX ],"
|
|
"height = (int) [ 1, MAX ],"
|
|
"framerate = (fraction) [ 0, MAX ];"
|
|
"video/x-dv,"
|
|
"systemstream = (boolean) FALSE,"
|
|
"width = (int) [ 1, MAX ],"
|
|
"height = (int) [ 1, MAX ],"
|
|
"framerate = (fraction) [ 0, MAX ],"
|
|
"format = (fourcc) dvsd;"
|
|
"video/x-dv,"
|
|
"systemstream = (boolean) TRUE;"
|
|
"video/x-raw-yuv,"
|
|
"width = (int) [ 1, MAX ],"
|
|
"height = (int) [ 1, MAX ],"
|
|
"framerate = (fraction) [ 0, MAX ]," "format = (fourcc) I420")
|
|
);
|
|
|
|
static void gst_dshowvideosrc_init_interfaces (GType type);
|
|
|
|
GST_BOILERPLATE_FULL (GstDshowVideoSrc, gst_dshowvideosrc, GstPushSrc,
|
|
GST_TYPE_PUSH_SRC, gst_dshowvideosrc_init_interfaces);
|
|
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_DEVICE,
|
|
PROP_DEVICE_NAME
|
|
};
|
|
|
|
static void gst_dshowvideosrc_probe_interface_init (GstPropertyProbeInterface *
|
|
iface);
|
|
static const GList *gst_dshowvideosrc_probe_get_properties (GstPropertyProbe *
|
|
probe);
|
|
static GValueArray *gst_dshowvideosrc_probe_get_values (GstPropertyProbe *
|
|
probe, guint prop_id, const GParamSpec * pspec);
|
|
static GValueArray *gst_dshowvideosrc_get_device_name_values (GstDshowVideoSrc *
|
|
src);
|
|
static gboolean gst_dshowvideosrc_probe_needs_probe (GstPropertyProbe * probe,
|
|
guint prop_id, const GParamSpec * pspec);
|
|
static void gst_dshowvideosrc_probe_probe_property (GstPropertyProbe * probe,
|
|
guint prop_id, const GParamSpec * pspec);
|
|
|
|
|
|
static void gst_dshowvideosrc_dispose (GObject * gobject);
|
|
static void gst_dshowvideosrc_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec);
|
|
static void gst_dshowvideosrc_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec);
|
|
static GstCaps *gst_dshowvideosrc_get_caps (GstBaseSrc * src);
|
|
static GstStateChangeReturn gst_dshowvideosrc_change_state (GstElement *
|
|
element, GstStateChange transition);
|
|
|
|
|
|
static gboolean gst_dshowvideosrc_start (GstBaseSrc * bsrc);
|
|
static gboolean gst_dshowvideosrc_stop (GstBaseSrc * bsrc);
|
|
static gboolean gst_dshowvideosrc_set_caps (GstBaseSrc * bsrc, GstCaps * caps);
|
|
static GstCaps *gst_dshowvideosrc_get_caps (GstBaseSrc * bsrc);
|
|
static GstFlowReturn gst_dshowvideosrc_create (GstPushSrc * psrc,
|
|
GstBuffer ** buf);
|
|
|
|
/*utils*/
|
|
static GstCaps *gst_dshowvideosrc_getcaps_from_streamcaps (GstDshowVideoSrc *
|
|
src, IPin * pin, IAMStreamConfig * streamcaps);
|
|
static gboolean gst_dshowvideosrc_push_buffer (byte * buffer, long size,
|
|
byte * src_object, UINT64 start, UINT64 stop);
|
|
|
|
static void
|
|
gst_dshowvideosrc_init_interfaces (GType type)
|
|
{
|
|
static const GInterfaceInfo dshowvideosrc_info = {
|
|
(GInterfaceInitFunc) gst_dshowvideosrc_probe_interface_init,
|
|
NULL,
|
|
NULL,
|
|
};
|
|
|
|
g_type_add_interface_static (type,
|
|
GST_TYPE_PROPERTY_PROBE, &dshowvideosrc_info);
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_probe_interface_init (GstPropertyProbeInterface * iface)
|
|
{
|
|
iface->get_properties = gst_dshowvideosrc_probe_get_properties;
|
|
iface->needs_probe = gst_dshowvideosrc_probe_needs_probe;
|
|
iface->probe_property = gst_dshowvideosrc_probe_probe_property;
|
|
iface->get_values = gst_dshowvideosrc_probe_get_values;
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_base_init (gpointer klass)
|
|
{
|
|
GstElementClass *element_class = GST_ELEMENT_CLASS (klass);
|
|
|
|
gst_element_class_add_pad_template (element_class,
|
|
gst_static_pad_template_get (&src_template));
|
|
|
|
gst_element_class_set_details (element_class, &gst_dshowvideosrc_details);
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_class_init (GstDshowVideoSrcClass * klass)
|
|
{
|
|
GObjectClass *gobject_class;
|
|
GstElementClass *gstelement_class;
|
|
GstBaseSrcClass *gstbasesrc_class;
|
|
GstPushSrcClass *gstpushsrc_class;
|
|
|
|
gobject_class = (GObjectClass *) klass;
|
|
gstelement_class = (GstElementClass *) klass;
|
|
gstbasesrc_class = (GstBaseSrcClass *) klass;
|
|
gstpushsrc_class = (GstPushSrcClass *) klass;
|
|
|
|
gobject_class->dispose = GST_DEBUG_FUNCPTR (gst_dshowvideosrc_dispose);
|
|
gobject_class->set_property =
|
|
GST_DEBUG_FUNCPTR (gst_dshowvideosrc_set_property);
|
|
gobject_class->get_property =
|
|
GST_DEBUG_FUNCPTR (gst_dshowvideosrc_get_property);
|
|
|
|
gstelement_class->change_state =
|
|
GST_DEBUG_FUNCPTR (gst_dshowvideosrc_change_state);
|
|
|
|
gstbasesrc_class->get_caps = GST_DEBUG_FUNCPTR (gst_dshowvideosrc_get_caps);
|
|
gstbasesrc_class->set_caps = GST_DEBUG_FUNCPTR (gst_dshowvideosrc_set_caps);
|
|
gstbasesrc_class->start = GST_DEBUG_FUNCPTR (gst_dshowvideosrc_start);
|
|
gstbasesrc_class->stop = GST_DEBUG_FUNCPTR (gst_dshowvideosrc_stop);
|
|
|
|
gstpushsrc_class->create = GST_DEBUG_FUNCPTR (gst_dshowvideosrc_create);
|
|
|
|
g_object_class_install_property
|
|
(gobject_class, PROP_DEVICE,
|
|
g_param_spec_string ("device", "Device",
|
|
"Directshow device path (@..classID/name)", NULL, G_PARAM_READWRITE));
|
|
|
|
g_object_class_install_property
|
|
(gobject_class, PROP_DEVICE_NAME,
|
|
g_param_spec_string ("device_name", "Device name",
|
|
"Human-readable name of the sound device", NULL, G_PARAM_READWRITE));
|
|
|
|
GST_DEBUG_CATEGORY_INIT (dshowvideosrc_debug, "dshowvideosrc", 0,
|
|
"Directshow video source");
|
|
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_init (GstDshowVideoSrc * src, GstDshowVideoSrcClass * klass)
|
|
{
|
|
src->device = NULL;
|
|
src->device_name = NULL;
|
|
src->video_cap_filter = NULL;
|
|
src->dshow_fakesink = NULL;
|
|
src->media_filter = NULL;
|
|
src->filter_graph = NULL;
|
|
src->caps = NULL;
|
|
src->pins_mediatypes = NULL;
|
|
src->is_rgb = FALSE;
|
|
|
|
src->async_queue = g_async_queue_new ();
|
|
|
|
CoInitializeEx (NULL, COINIT_MULTITHREADED);
|
|
|
|
gst_base_src_set_live (GST_BASE_SRC (src), TRUE);
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_dispose (GObject * gobject)
|
|
{
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (gobject);
|
|
|
|
if (src->device) {
|
|
g_free (src->device);
|
|
src->device = NULL;
|
|
}
|
|
|
|
if (src->device_name) {
|
|
g_free (src->device_name);
|
|
src->device_name = NULL;
|
|
}
|
|
|
|
if (src->caps) {
|
|
gst_caps_unref (src->caps);
|
|
src->caps = NULL;
|
|
}
|
|
|
|
if (src->pins_mediatypes) {
|
|
gst_dshow_free_pins_mediatypes (src->pins_mediatypes);
|
|
src->pins_mediatypes = NULL;
|
|
}
|
|
|
|
/* clean dshow */
|
|
if (src->video_cap_filter) {
|
|
IBaseFilter_Release (src->video_cap_filter);
|
|
src->video_cap_filter = NULL;
|
|
}
|
|
|
|
if (src->async_queue) {
|
|
g_async_queue_unref (src->async_queue);
|
|
src->async_queue = NULL;
|
|
}
|
|
|
|
CoUninitialize ();
|
|
|
|
G_OBJECT_CLASS (parent_class)->dispose (object);
|
|
}
|
|
|
|
static gboolean
|
|
gst_dshowvideosrc_probe_needs_probe (GstPropertyProbe * probe,
|
|
guint prop_id, const GParamSpec * pspec)
|
|
{
|
|
static gboolean init = FALSE;
|
|
gboolean ret = FALSE;
|
|
|
|
if (!init) {
|
|
ret = TRUE;
|
|
init = TRUE;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_probe_probe_property (GstPropertyProbe * probe,
|
|
guint prop_id, const GParamSpec * pspec)
|
|
{
|
|
GObjectClass *klass = G_OBJECT_GET_CLASS (probe);
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE_NAME:
|
|
//gst_v4l_class_probe_devices (klass, FALSE);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (probe, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static const GList *
|
|
gst_dshowvideosrc_probe_get_properties (GstPropertyProbe * probe)
|
|
{
|
|
GObjectClass *klass = G_OBJECT_GET_CLASS (probe);
|
|
static GList *props = NULL;
|
|
|
|
if (!props) {
|
|
GParamSpec *pspec;
|
|
|
|
pspec = g_object_class_find_property (klass, "device_name");
|
|
props = g_list_append (props, pspec);
|
|
}
|
|
|
|
return props;
|
|
}
|
|
|
|
static GValueArray *
|
|
gst_dshowvideosrc_get_device_name_values (GstDshowVideoSrc * src)
|
|
{
|
|
GValueArray *array = g_value_array_new (0);
|
|
GValue value = { 0 };
|
|
ICreateDevEnum *devices_enum = NULL;
|
|
IEnumMoniker *moniker_enum = NULL;
|
|
IMoniker *moniker = NULL;
|
|
HRESULT hres = S_FALSE;
|
|
ULONG fetched;
|
|
|
|
g_value_init (&value, G_TYPE_STRING);
|
|
|
|
hres = CoCreateInstance (&CLSID_SystemDeviceEnum, NULL, CLSCTX_INPROC_SERVER,
|
|
&IID_ICreateDevEnum, (void **) &devices_enum);
|
|
if (hres != S_OK) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't create an instance of the system device enumerator (error=%d)",
|
|
hres);
|
|
array = NULL;
|
|
goto clean;
|
|
}
|
|
|
|
hres =
|
|
ICreateDevEnum_CreateClassEnumerator (devices_enum,
|
|
&CLSID_VideoInputDeviceCategory, &moniker_enum, 0);
|
|
if (hres != S_OK || !moniker_enum) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't get enumeration of video devices (error=%d)", hres);
|
|
array = NULL;
|
|
goto clean;
|
|
}
|
|
|
|
IEnumMoniker_Reset (moniker_enum);
|
|
|
|
while (hres = IEnumMoniker_Next (moniker_enum, 1, &moniker, &fetched),
|
|
hres == S_OK) {
|
|
IPropertyBag *property_bag = NULL;
|
|
|
|
hres =
|
|
IMoniker_BindToStorage (moniker, NULL, NULL, &IID_IPropertyBag,
|
|
(void **) &property_bag);
|
|
if (SUCCEEDED (hres) && property_bag) {
|
|
VARIANT varFriendlyName;
|
|
|
|
VariantInit (&varFriendlyName);
|
|
hres =
|
|
IPropertyBag_Read (property_bag, L"FriendlyName", &varFriendlyName,
|
|
NULL);
|
|
if (hres == S_OK && varFriendlyName.bstrVal) {
|
|
gchar *friendly_name =
|
|
g_utf16_to_utf8 ((const gunichar2 *) varFriendlyName.bstrVal,
|
|
wcslen (varFriendlyName.bstrVal), NULL, NULL, NULL);
|
|
|
|
g_value_set_string (&value, friendly_name);
|
|
g_value_array_append (array, &value);
|
|
g_value_unset (&value);
|
|
g_free (friendly_name);
|
|
SysFreeString (varFriendlyName.bstrVal);
|
|
}
|
|
IPropertyBag_Release (property_bag);
|
|
}
|
|
IMoniker_Release (moniker);
|
|
}
|
|
|
|
clean:
|
|
if (moniker_enum) {
|
|
IEnumMoniker_Release (moniker_enum);
|
|
}
|
|
|
|
if (devices_enum) {
|
|
ICreateDevEnum_Release (devices_enum);
|
|
}
|
|
|
|
return array;
|
|
}
|
|
|
|
static GValueArray *
|
|
gst_dshowvideosrc_probe_get_values (GstPropertyProbe * probe,
|
|
guint prop_id, const GParamSpec * pspec)
|
|
{
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (probe);
|
|
GValueArray *array = NULL;
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE_NAME:
|
|
array = gst_dshowvideosrc_get_device_name_values (src);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (probe, prop_id, pspec);
|
|
break;
|
|
}
|
|
|
|
return array;
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE:
|
|
{
|
|
if (src->device) {
|
|
g_free (src->device);
|
|
src->device = NULL;
|
|
}
|
|
if (g_value_get_string (value)) {
|
|
src->device = g_strdup (g_value_get_string (value));
|
|
}
|
|
break;
|
|
}
|
|
case PROP_DEVICE_NAME:
|
|
{
|
|
if (src->device_name) {
|
|
g_free (src->device_name);
|
|
src->device_name = NULL;
|
|
}
|
|
if (g_value_get_string (value)) {
|
|
src->device_name = g_strdup (g_value_get_string (value));
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_dshowvideosrc_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec)
|
|
{
|
|
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_dshowvideosrc_get_caps (GstBaseSrc * basesrc)
|
|
{
|
|
HRESULT hres = S_OK;
|
|
IBindCtx *lpbc = NULL;
|
|
IMoniker *videom;
|
|
DWORD dwEaten;
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (basesrc);
|
|
gunichar2 *unidevice = NULL;
|
|
|
|
if (src->device) {
|
|
g_free (src->device);
|
|
src->device = NULL;
|
|
}
|
|
|
|
src->device =
|
|
gst_dshow_getdevice_from_devicename (&CLSID_VideoInputDeviceCategory,
|
|
&src->device_name);
|
|
if (!src->device) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug, "No video device found.");
|
|
return NULL;
|
|
}
|
|
unidevice =
|
|
g_utf8_to_utf16 (src->device, strlen (src->device), NULL, NULL, NULL);
|
|
|
|
if (!src->video_cap_filter) {
|
|
hres = CreateBindCtx (0, &lpbc);
|
|
if (SUCCEEDED (hres)) {
|
|
hres = MkParseDisplayName (lpbc, unidevice, &dwEaten, &videom);
|
|
if (SUCCEEDED (hres)) {
|
|
hres =
|
|
IMoniker_BindToObject (videom, lpbc, NULL, &IID_IBaseFilter,
|
|
&src->video_cap_filter);
|
|
IMoniker_Release (videom);
|
|
}
|
|
IBindCtx_Release (lpbc);
|
|
}
|
|
}
|
|
|
|
if (src->video_cap_filter && !src->caps) {
|
|
/* get the capture pins supported types */
|
|
IPin *capture_pin = NULL;
|
|
IEnumPins *enumpins = NULL;
|
|
HRESULT hres;
|
|
|
|
hres = IBaseFilter_EnumPins (src->video_cap_filter, &enumpins);
|
|
if (SUCCEEDED (hres)) {
|
|
while (IEnumPins_Next (enumpins, 1, &capture_pin, NULL) == S_OK) {
|
|
IKsPropertySet *pKs = NULL;
|
|
|
|
hres =
|
|
IPin_QueryInterface (capture_pin, &IID_IKsPropertySet,
|
|
(void **) &pKs);
|
|
if (SUCCEEDED (hres) && pKs) {
|
|
DWORD cbReturned;
|
|
GUID pin_category;
|
|
RPC_STATUS rpcstatus;
|
|
|
|
hres =
|
|
IKsPropertySet_Get (pKs, &ROPSETID_Pin,
|
|
AMPROPERTY_PIN_CATEGORY, NULL, 0, &pin_category, sizeof (GUID),
|
|
&cbReturned);
|
|
|
|
/* we only want capture pins */
|
|
if (UuidCompare (&pin_category, &PIN_CATEGORY_CAPTURE,
|
|
&rpcstatus) == 0) {
|
|
IAMStreamConfig *streamcaps = NULL;
|
|
|
|
if (SUCCEEDED (IPin_QueryInterface (capture_pin,
|
|
&IID_IAMStreamConfig, (void **) &streamcaps))) {
|
|
src->caps =
|
|
gst_dshowvideosrc_getcaps_from_streamcaps (src, capture_pin,
|
|
streamcaps);
|
|
IAMStreamConfig_Release (streamcaps);
|
|
|
|
GST_CAT_LOG (dshowvideosrc_debug,
|
|
"get_cap returned %" GST_PTR_FORMAT, src->caps);
|
|
}
|
|
}
|
|
|
|
IKsPropertySet_Release (pKs);
|
|
}
|
|
|
|
IPin_Release (capture_pin);
|
|
}
|
|
IEnumPins_Release (enumpins);
|
|
}
|
|
}
|
|
|
|
if (unidevice) {
|
|
g_free (unidevice);
|
|
}
|
|
|
|
if (src->caps) {
|
|
return gst_caps_ref (src->caps);
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static GstStateChangeReturn
|
|
gst_dshowvideosrc_change_state (GstElement * element, GstStateChange transition)
|
|
{
|
|
HRESULT hres = S_FALSE;
|
|
IAMVfwCaptureDialogs *dialog = NULL;
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (element);
|
|
|
|
switch (transition) {
|
|
case GST_STATE_CHANGE_NULL_TO_READY:
|
|
break;
|
|
case GST_STATE_CHANGE_READY_TO_PAUSED:
|
|
break;
|
|
case GST_STATE_CHANGE_PAUSED_TO_PLAYING:
|
|
if (src->media_filter)
|
|
hres = IMediaFilter_Run (src->media_filter, 0);
|
|
if (hres != S_OK) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't RUN the directshow capture graph (error=%d)", hres);
|
|
return GST_STATE_CHANGE_FAILURE;
|
|
}
|
|
break;
|
|
case GST_STATE_CHANGE_PLAYING_TO_PAUSED:
|
|
if (src->media_filter)
|
|
hres = IMediaFilter_Stop (src->media_filter);
|
|
if (hres != S_OK) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't STOP the directshow capture graph (error=%d)", hres);
|
|
return GST_STATE_CHANGE_FAILURE;
|
|
}
|
|
break;
|
|
case GST_STATE_CHANGE_PAUSED_TO_READY:
|
|
break;
|
|
case GST_STATE_CHANGE_READY_TO_NULL:
|
|
break;
|
|
}
|
|
|
|
return GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
|
|
}
|
|
|
|
static gboolean
|
|
gst_dshowvideosrc_start (GstBaseSrc * bsrc)
|
|
{
|
|
HRESULT hres = S_FALSE;
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (bsrc);
|
|
|
|
hres = CoCreateInstance (&CLSID_FilterGraph, NULL, CLSCTX_INPROC,
|
|
&IID_IFilterGraph, (LPVOID *) & src->filter_graph);
|
|
if (hres != S_OK || !src->filter_graph) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't create an instance of the dshow graph manager (error=%d)", hres);
|
|
goto error;
|
|
}
|
|
|
|
hres = IFilterGraph_QueryInterface (src->filter_graph, &IID_IMediaFilter,
|
|
(void **) &src->media_filter);
|
|
if (hres != S_OK || !src->media_filter) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't get IMediacontrol interface from the graph manager (error=%d)",
|
|
hres);
|
|
goto error;
|
|
}
|
|
|
|
hres = CoCreateInstance (&CLSID_DshowFakeSink, NULL, CLSCTX_INPROC,
|
|
&IID_IBaseFilter, (LPVOID *) & src->dshow_fakesink);
|
|
if (hres != S_OK || !src->dshow_fakesink) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't create an instance of our dshow fakesink filter (error=%d)",
|
|
hres);
|
|
goto error;
|
|
}
|
|
|
|
hres =
|
|
IFilterGraph_AddFilter (src->filter_graph, src->video_cap_filter,
|
|
L"capture");
|
|
if (hres != S_OK) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't add video capture filter to the graph (error=%d)", hres);
|
|
goto error;
|
|
}
|
|
|
|
hres =
|
|
IFilterGraph_AddFilter (src->filter_graph, src->dshow_fakesink, L"sink");
|
|
if (hres != S_OK) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't add our fakesink filter to the graph (error=%d)", hres);
|
|
goto error;
|
|
}
|
|
|
|
return TRUE;
|
|
|
|
error:
|
|
if (src->dshow_fakesink) {
|
|
IBaseFilter_Release (src->dshow_fakesink);
|
|
src->dshow_fakesink = NULL;
|
|
}
|
|
|
|
if (src->media_filter) {
|
|
IMediaFilter_Release (src->media_filter);
|
|
src->media_filter = NULL;
|
|
}
|
|
if (src->filter_graph) {
|
|
IFilterGraph_Release (src->filter_graph);
|
|
src->filter_graph = NULL;
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_dshowvideosrc_set_caps (GstBaseSrc * bsrc, GstCaps * caps)
|
|
{
|
|
HRESULT hres;
|
|
IGstDshowInterface *srcinterface = NULL;
|
|
IPin *input_pin = NULL;
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (bsrc);
|
|
GstStructure *s = gst_caps_get_structure (caps, 0);
|
|
|
|
/* search the negociated caps in our caps list to get its index and the corresponding mediatype */
|
|
if (gst_caps_is_subset (caps, src->caps)) {
|
|
guint i = 0;
|
|
gint res = -1;
|
|
|
|
for (; i < gst_caps_get_size (src->caps) && res == -1; i++) {
|
|
GstCaps *capstmp = gst_caps_copy_nth (src->caps, i);
|
|
|
|
if (gst_caps_is_subset (caps, capstmp)) {
|
|
res = i;
|
|
}
|
|
gst_caps_unref (capstmp);
|
|
}
|
|
|
|
if (res != -1 && src->pins_mediatypes) {
|
|
/* get the corresponding media type and build the dshow graph */
|
|
GstCapturePinMediaType *pin_mediatype = NULL;
|
|
gchar *caps_string = NULL;
|
|
GList *type = g_list_nth (src->pins_mediatypes, res);
|
|
|
|
if (type) {
|
|
pin_mediatype = (GstCapturePinMediaType *) type->data;
|
|
|
|
hres =
|
|
IBaseFilter_QueryInterface (src->dshow_fakesink,
|
|
&IID_IGstDshowInterface, (void **) &srcinterface);
|
|
|
|
if (hres != S_OK || !srcinterface) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't get IGstDshowInterface interface from our dshow fakesink filter (error=%d)",
|
|
hres);
|
|
goto error;
|
|
}
|
|
|
|
IGstDshowInterface_gst_set_media_type (srcinterface,
|
|
pin_mediatype->mediatype);
|
|
IGstDshowInterface_gst_set_buffer_callback (srcinterface,
|
|
(byte *) gst_dshowvideosrc_push_buffer, (byte *) src);
|
|
|
|
if (srcinterface) {
|
|
IGstDshowInterface_Release (srcinterface);
|
|
}
|
|
|
|
gst_dshow_get_pin_from_filter (src->dshow_fakesink, PINDIR_INPUT,
|
|
&input_pin);
|
|
if (!input_pin) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't get input pin from our dshow fakesink");
|
|
goto error;
|
|
}
|
|
|
|
hres =
|
|
IFilterGraph_ConnectDirect (src->filter_graph,
|
|
pin_mediatype->capture_pin, input_pin, NULL);
|
|
IPin_Release (input_pin);
|
|
|
|
if (hres != S_OK) {
|
|
GST_CAT_ERROR (dshowvideosrc_debug,
|
|
"Can't connect capture filter with fakesink filter (error=%d)",
|
|
hres);
|
|
goto error;
|
|
}
|
|
|
|
/* save width and height negociated */
|
|
gst_structure_get_int (s, "width", &src->width);
|
|
gst_structure_get_int (s, "height", &src->height);
|
|
|
|
src->is_rgb = FALSE;
|
|
caps_string = gst_caps_to_string (caps);
|
|
if (caps_string) {
|
|
if (strstr (caps_string, "video/x-raw-rgb")) {
|
|
src->is_rgb = TRUE;
|
|
} else {
|
|
src->is_rgb = FALSE;
|
|
}
|
|
g_free (caps_string);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return TRUE;
|
|
|
|
error:
|
|
if (srcinterface) {
|
|
IGstDshowInterface_Release (srcinterface);
|
|
}
|
|
|
|
return FALSE;
|
|
}
|
|
|
|
static gboolean
|
|
gst_dshowvideosrc_stop (GstBaseSrc * bsrc)
|
|
{
|
|
IPin *input_pin = NULL, *output_pin = NULL;
|
|
HRESULT hres = S_FALSE;
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (bsrc);
|
|
|
|
if (!src->filter_graph)
|
|
return TRUE;
|
|
|
|
/* disconnect filters */
|
|
gst_dshow_get_pin_from_filter (src->video_cap_filter, PINDIR_OUTPUT,
|
|
&output_pin);
|
|
if (output_pin) {
|
|
hres = IFilterGraph_Disconnect (src->filter_graph, output_pin);
|
|
IPin_Release (output_pin);
|
|
}
|
|
|
|
gst_dshow_get_pin_from_filter (src->dshow_fakesink, PINDIR_INPUT, &input_pin);
|
|
if (input_pin) {
|
|
hres = IFilterGraph_Disconnect (src->filter_graph, input_pin);
|
|
IPin_Release (input_pin);
|
|
}
|
|
|
|
/*remove filters from the graph */
|
|
IFilterGraph_RemoveFilter (src->filter_graph, src->video_cap_filter);
|
|
IFilterGraph_RemoveFilter (src->filter_graph, src->dshow_fakesink);
|
|
|
|
/*release our gstreamer dshow sink */
|
|
IBaseFilter_Release (src->dshow_fakesink);
|
|
src->dshow_fakesink = NULL;
|
|
|
|
/*release media filter interface */
|
|
IMediaFilter_Release (src->media_filter);
|
|
src->media_filter = NULL;
|
|
|
|
/*release the filter graph manager */
|
|
IFilterGraph_Release (src->filter_graph);
|
|
src->filter_graph = NULL;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_dshowvideosrc_create (GstPushSrc * psrc, GstBuffer ** buf)
|
|
{
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (psrc);
|
|
|
|
*buf = g_async_queue_pop (src->async_queue);
|
|
|
|
GST_CAT_DEBUG (dshowvideosrc_debug,
|
|
"dshowvideosrc_create => pts %" GST_TIME_FORMAT " duration %"
|
|
GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (*buf)),
|
|
GST_TIME_ARGS (GST_BUFFER_DURATION (*buf)));
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_dshowvideosrc_getcaps_from_streamcaps (GstDshowVideoSrc * src, IPin * pin,
|
|
IAMStreamConfig * streamcaps)
|
|
{
|
|
GstCaps *caps = NULL;
|
|
HRESULT hres = S_OK;
|
|
RPC_STATUS rpcstatus;
|
|
int icount = 0;
|
|
int isize = 0;
|
|
VIDEO_STREAM_CONFIG_CAPS vscc;
|
|
int i = 0;
|
|
|
|
if (!streamcaps)
|
|
return NULL;
|
|
|
|
IAMStreamConfig_GetNumberOfCapabilities (streamcaps, &icount, &isize);
|
|
|
|
if (isize != sizeof (vscc))
|
|
return NULL;
|
|
|
|
for (; i < icount; i++) {
|
|
GstCapturePinMediaType *pin_mediatype = g_new0 (GstCapturePinMediaType, 1);
|
|
|
|
IPin_AddRef (pin);
|
|
pin_mediatype->capture_pin = pin;
|
|
|
|
hres =
|
|
IAMStreamConfig_GetStreamCaps (streamcaps, i, &pin_mediatype->mediatype,
|
|
(BYTE *) & vscc);
|
|
if (hres == S_OK && pin_mediatype->mediatype) {
|
|
VIDEOINFOHEADER *video_info;
|
|
GstCaps *mediacaps = NULL;
|
|
|
|
if (!caps)
|
|
caps = gst_caps_new_empty ();
|
|
|
|
/* I420 */
|
|
if ((UuidCompare (&pin_mediatype->mediatype->subtype, &MEDIASUBTYPE_I420,
|
|
&rpcstatus) == 0 && rpcstatus == RPC_S_OK)
|
|
&& (UuidCompare (&pin_mediatype->mediatype->formattype,
|
|
&FORMAT_VideoInfo, &rpcstatus) == 0
|
|
&& rpcstatus == RPC_S_OK)) {
|
|
video_info = (VIDEOINFOHEADER *) pin_mediatype->mediatype->pbFormat;
|
|
|
|
mediacaps = gst_caps_new_simple ("video/x-raw-yuv",
|
|
"width", G_TYPE_INT, video_info->bmiHeader.biWidth,
|
|
"height", G_TYPE_INT, video_info->bmiHeader.biHeight,
|
|
"framerate", GST_TYPE_FRACTION,
|
|
(int) (10000000 / video_info->AvgTimePerFrame), 1, "format",
|
|
GST_TYPE_FOURCC, MAKEFOURCC ('I', '4', '2', '0'), NULL);
|
|
|
|
if (mediacaps) {
|
|
src->pins_mediatypes =
|
|
g_list_append (src->pins_mediatypes, pin_mediatype);
|
|
gst_caps_append (caps, mediacaps);
|
|
} else {
|
|
gst_dshow_free_pin_mediatype (pin_mediatype);
|
|
}
|
|
continue;
|
|
}
|
|
|
|
/* RGB24 */
|
|
if ((UuidCompare (&pin_mediatype->mediatype->subtype, &MEDIASUBTYPE_RGB24,
|
|
&rpcstatus) == 0 && rpcstatus == RPC_S_OK)
|
|
&& (UuidCompare (&pin_mediatype->mediatype->formattype,
|
|
&FORMAT_VideoInfo, &rpcstatus) == 0
|
|
&& rpcstatus == RPC_S_OK)) {
|
|
video_info = (VIDEOINFOHEADER *) pin_mediatype->mediatype->pbFormat;
|
|
|
|
/* ffmpegcolorspace handles RGB24 in BIG_ENDIAN */
|
|
mediacaps = gst_caps_new_simple ("video/x-raw-rgb",
|
|
"bpp", G_TYPE_INT, 24,
|
|
"depth", G_TYPE_INT, 24,
|
|
"width", G_TYPE_INT, video_info->bmiHeader.biWidth,
|
|
"height", G_TYPE_INT, video_info->bmiHeader.biHeight,
|
|
"framerate", GST_TYPE_FRACTION,
|
|
(int) (10000000 / video_info->AvgTimePerFrame), 1, "endianness",
|
|
G_TYPE_INT, G_BIG_ENDIAN, "red_mask", G_TYPE_INT, 255, "green_mask",
|
|
G_TYPE_INT, 65280, "blue_mask", G_TYPE_INT, 16711680, NULL);
|
|
|
|
if (mediacaps) {
|
|
src->pins_mediatypes =
|
|
g_list_append (src->pins_mediatypes, pin_mediatype);
|
|
gst_caps_append (caps, mediacaps);
|
|
} else {
|
|
gst_dshow_free_pin_mediatype (pin_mediatype);
|
|
}
|
|
continue;
|
|
}
|
|
|
|
/* DVSD */
|
|
if ((UuidCompare (&pin_mediatype->mediatype->subtype, &MEDIASUBTYPE_dvsd,
|
|
&rpcstatus) == 0 && rpcstatus == RPC_S_OK)
|
|
&& (UuidCompare (&pin_mediatype->mediatype->formattype,
|
|
&FORMAT_VideoInfo, &rpcstatus) == 0
|
|
&& rpcstatus == RPC_S_OK)) {
|
|
video_info = (VIDEOINFOHEADER *) pin_mediatype->mediatype->pbFormat;
|
|
|
|
mediacaps = gst_caps_new_simple ("video/x-dv",
|
|
"systemstream", G_TYPE_BOOLEAN, FALSE,
|
|
"format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('d', 'v', 's', 'd'),
|
|
"framerate", GST_TYPE_FRACTION,
|
|
(int) (10000000 / video_info->AvgTimePerFrame), 1, "width",
|
|
G_TYPE_INT, video_info->bmiHeader.biWidth, "height", G_TYPE_INT,
|
|
video_info->bmiHeader.biHeight, NULL);
|
|
|
|
if (mediacaps) {
|
|
src->pins_mediatypes =
|
|
g_list_append (src->pins_mediatypes, pin_mediatype);
|
|
gst_caps_append (caps, mediacaps);
|
|
} else {
|
|
gst_dshow_free_pin_mediatype (pin_mediatype);
|
|
}
|
|
continue;
|
|
}
|
|
|
|
/* DV stream */
|
|
if ((UuidCompare (&pin_mediatype->mediatype->subtype, &MEDIASUBTYPE_dvsd,
|
|
&rpcstatus) == 0 && rpcstatus == RPC_S_OK)
|
|
&& (UuidCompare (&pin_mediatype->mediatype->formattype,
|
|
&FORMAT_DvInfo, &rpcstatus) == 0 && rpcstatus == RPC_S_OK)) {
|
|
|
|
mediacaps = gst_caps_new_simple ("video/x-dv",
|
|
"systemstream", G_TYPE_BOOLEAN, TRUE, NULL);
|
|
|
|
if (mediacaps) {
|
|
src->pins_mediatypes =
|
|
g_list_append (src->pins_mediatypes, pin_mediatype);
|
|
gst_caps_append (caps, mediacaps);
|
|
} else {
|
|
gst_dshow_free_pin_mediatype (pin_mediatype);
|
|
}
|
|
continue;
|
|
}
|
|
} else {
|
|
gst_dshow_free_pin_mediatype (pin_mediatype);
|
|
}
|
|
}
|
|
|
|
if (caps && gst_caps_is_empty (caps)) {
|
|
gst_caps_unref (caps);
|
|
caps = NULL;
|
|
}
|
|
|
|
return caps;
|
|
}
|
|
|
|
static gboolean
|
|
gst_dshowvideosrc_push_buffer (byte * buffer, long size, byte * src_object,
|
|
UINT64 start, UINT64 stop)
|
|
{
|
|
GstDshowVideoSrc *src = GST_DSHOWVIDEOSRC (src_object);
|
|
GstBuffer *buf;
|
|
IPin *pPin = NULL;
|
|
HRESULT hres = S_FALSE;
|
|
AM_MEDIA_TYPE *pMediaType = NULL;
|
|
|
|
if (!buffer || size == 0 || !src) {
|
|
return FALSE;
|
|
}
|
|
|
|
/* create a new buffer assign to it the clock time as timestamp */
|
|
buf = gst_buffer_new_and_alloc (size);
|
|
|
|
GST_BUFFER_SIZE (buf) = size;
|
|
GST_BUFFER_TIMESTAMP (buf) = gst_clock_get_time (GST_ELEMENT (src)->clock);
|
|
GST_BUFFER_TIMESTAMP (buf) -= GST_ELEMENT (src)->base_time;
|
|
GST_BUFFER_DURATION (buf) = stop - start;
|
|
|
|
if (src->is_rgb) {
|
|
/* FOR RGB directshow decoder will return bottom-up BITMAP
|
|
* There is probably a way to get top-bottom video frames from
|
|
* the decoder...
|
|
*/
|
|
gint line = 0;
|
|
gint stride = size / src->height;
|
|
|
|
for (; line < src->height; line++) {
|
|
memcpy (GST_BUFFER_DATA (buf) + (line * stride),
|
|
buffer + (size - ((line + 1) * (stride))), stride);
|
|
}
|
|
} else {
|
|
memcpy (GST_BUFFER_DATA (buf), buffer, size);
|
|
}
|
|
|
|
GST_CAT_DEBUG (dshowvideosrc_debug,
|
|
"push_buffer => pts %" GST_TIME_FORMAT "duration %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (buf)), GST_TIME_ARGS (stop - start));
|
|
|
|
/* the negotiate() method already set caps on the source pad */
|
|
gst_buffer_set_caps (buf, GST_PAD_CAPS (GST_BASE_SRC_PAD (src)));
|
|
|
|
g_async_queue_push (src->async_queue, buf);
|
|
|
|
return TRUE;
|
|
}
|