mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-03 16:09:39 +00:00
7898bc5810
This commit introduces IOSGLMemory which is a GLMemory that falls back to GstAppleCoreVideoMemory for CPU access. This is a temporary solution until IOSurface gets exposed as a public framework on iOS and so we can use IOSurfaceMemory on both MacOS and iOS. https://bugzilla.gnome.org/show_bug.cgi?id=769210
740 lines
19 KiB
Objective-C
740 lines
19 KiB
Objective-C
/*
|
|
* Copyright (C) 2009 Ole André Vadla Ravnås <oleavr@soundrop.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#include "qtkitvideosrc.h"
|
|
#import "corevideobuffer.h"
|
|
#include "glcontexthelper.h"
|
|
|
|
#import <QTKit/QTKit.h>
|
|
|
|
#define DEFAULT_DEVICE_INDEX -1
|
|
|
|
#define DEVICE_YUV_FOURCC "UYVY"
|
|
#define DEVICE_FPS_N 30
|
|
#define DEVICE_FPS_D 1
|
|
|
|
#define FRAME_QUEUE_SIZE 2
|
|
|
|
GST_DEBUG_CATEGORY (gst_qtkit_video_src_debug);
|
|
#define GST_CAT_DEFAULT gst_qtkit_video_src_debug
|
|
|
|
static GstStaticPadTemplate src_template = GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS (
|
|
"video/x-raw, "
|
|
"format =" DEVICE_YUV_FOURCC ", "
|
|
"width = (int) 640, "
|
|
"height = (int) 480, "
|
|
"framerate = [0/1, 100/1], "
|
|
"pixel-aspect-ratio = (fraction) 1/1"
|
|
"; "
|
|
"video/x-raw, "
|
|
"format =" DEVICE_YUV_FOURCC ", "
|
|
"width = (int) 160, "
|
|
"height = (int) 120, "
|
|
"framerate = [0/1, 100/1], "
|
|
"pixel-aspect-ratio = (fraction) 1/1"
|
|
"; "
|
|
"video/x-raw, "
|
|
"format =" DEVICE_YUV_FOURCC ", "
|
|
"width = (int) 176, "
|
|
"height = (int) 144, "
|
|
"framerate = [0/1, 100/1], "
|
|
"pixel-aspect-ratio = (fraction) 12/11"
|
|
"; "
|
|
"video/x-raw, "
|
|
"format =" DEVICE_YUV_FOURCC ", "
|
|
"width = (int) 320, "
|
|
"height = (int) 240, "
|
|
"framerate = [0/1, 100/1], "
|
|
"pixel-aspect-ratio = (fraction) 1/1"
|
|
"; "
|
|
"video/x-raw, "
|
|
"format =" DEVICE_YUV_FOURCC ", "
|
|
"width = (int) 352, "
|
|
"height = (int) 288, "
|
|
"framerate = [0/1, 100/1], "
|
|
"pixel-aspect-ratio = (fraction) 12/11"
|
|
";"
|
|
)
|
|
);
|
|
|
|
typedef enum _QueueState {
|
|
NO_FRAMES = 1,
|
|
HAS_FRAME_OR_STOP_REQUEST,
|
|
} QueueState;
|
|
|
|
G_DEFINE_TYPE (GstQTKitVideoSrc, gst_qtkit_video_src, GST_TYPE_PUSH_SRC);
|
|
|
|
@interface GstQTKitVideoSrcImpl : NSObject {
|
|
GstElement *element;
|
|
GstBaseSrc *baseSrc;
|
|
GstPushSrc *pushSrc;
|
|
|
|
int deviceIndex;
|
|
|
|
QTCaptureSession *session;
|
|
QTCaptureDeviceInput *input;
|
|
QTCaptureDecompressedVideoOutput *output;
|
|
QTCaptureDevice *device;
|
|
|
|
NSConditionLock *queueLock;
|
|
NSMutableArray *queue;
|
|
BOOL stopRequest;
|
|
|
|
gint width, height;
|
|
gint fps_n, fps_d;
|
|
GstClockTime duration;
|
|
guint64 offset;
|
|
GstGLContextHelper *ctxh;
|
|
GstVideoTextureCache *textureCache;
|
|
GstVideoInfo outputInfo;
|
|
}
|
|
|
|
- (id)init;
|
|
- (id)initWithSrc:(GstPushSrc *)src;
|
|
|
|
@property int deviceIndex;
|
|
|
|
- (BOOL)openDevice;
|
|
- (void)closeDevice;
|
|
- (BOOL)setCaps:(GstCaps *)caps;
|
|
- (BOOL)start;
|
|
- (BOOL)stop;
|
|
- (BOOL)unlock;
|
|
- (BOOL)unlockStop;
|
|
- (GstCaps *)fixate:(GstCaps *)caps;
|
|
- (BOOL)query:(GstQuery *)query;
|
|
- (GstStateChangeReturn)changeState:(GstStateChange)transition;
|
|
- (GstFlowReturn)create:(GstBuffer **)buf;
|
|
- (void)timestampBuffer:(GstBuffer *)buf;
|
|
- (void)captureOutput:(QTCaptureOutput *)captureOutput
|
|
didOutputVideoFrame:(CVImageBufferRef)videoFrame
|
|
withSampleBuffer:(QTSampleBuffer *)sampleBuffer
|
|
fromConnection:(QTCaptureConnection *)connection;
|
|
|
|
@end
|
|
|
|
@implementation GstQTKitVideoSrcImpl
|
|
|
|
- (id)init
|
|
{
|
|
return [self initWithSrc:NULL];
|
|
}
|
|
|
|
- (id)initWithSrc:(GstPushSrc *)src
|
|
{
|
|
if ((self = [super init])) {
|
|
element = GST_ELEMENT_CAST (src);
|
|
baseSrc = GST_BASE_SRC_CAST (src);
|
|
pushSrc = src;
|
|
|
|
deviceIndex = DEFAULT_DEVICE_INDEX;
|
|
|
|
device = nil;
|
|
|
|
gst_base_src_set_live (baseSrc, TRUE);
|
|
gst_base_src_set_format (baseSrc, GST_FORMAT_TIME);
|
|
|
|
self->ctxh = NULL;
|
|
textureCache = NULL;
|
|
gst_video_info_init (&outputInfo);
|
|
}
|
|
|
|
return self;
|
|
}
|
|
|
|
@synthesize deviceIndex;
|
|
|
|
- (BOOL)openDevice
|
|
{
|
|
NSString *mediaType = QTMediaTypeVideo;
|
|
NSError *error = nil;
|
|
|
|
if (deviceIndex == -1) {
|
|
device = [QTCaptureDevice defaultInputDeviceWithMediaType:mediaType];
|
|
if (device == nil) {
|
|
GST_ELEMENT_ERROR (element, RESOURCE, NOT_FOUND,
|
|
("No video capture devices found"), (NULL));
|
|
goto openFailed;
|
|
}
|
|
} else {
|
|
NSArray *devices = [QTCaptureDevice inputDevicesWithMediaType:mediaType];
|
|
if (deviceIndex >= [devices count]) {
|
|
GST_ELEMENT_ERROR (element, RESOURCE, NOT_FOUND,
|
|
("Invalid video capture device index"), (NULL));
|
|
goto openFailed;
|
|
}
|
|
device = [devices objectAtIndex:deviceIndex];
|
|
}
|
|
[device retain];
|
|
|
|
GST_INFO ("Opening '%s'", [[device localizedDisplayName] UTF8String]);
|
|
|
|
if (![device open:&error]) {
|
|
GST_ELEMENT_ERROR (element, RESOURCE, NOT_FOUND,
|
|
("Failed to open device '%s'",
|
|
[[device localizedDisplayName] UTF8String]), (NULL));
|
|
goto openFailed;
|
|
}
|
|
|
|
return YES;
|
|
|
|
/* ERRORS */
|
|
openFailed:
|
|
{
|
|
[device release];
|
|
device = nil;
|
|
|
|
return NO;
|
|
}
|
|
}
|
|
|
|
- (void)closeDevice
|
|
{
|
|
g_assert (![session isRunning]);
|
|
|
|
[session release];
|
|
session = nil;
|
|
|
|
[input release];
|
|
input = nil;
|
|
|
|
[output release];
|
|
output = nil;
|
|
|
|
[device release];
|
|
device = nil;
|
|
}
|
|
|
|
- (BOOL)setCaps:(GstCaps *)caps
|
|
{
|
|
GstStructure *s;
|
|
NSDictionary *outputAttrs;
|
|
BOOL success;
|
|
NSTimeInterval interval;
|
|
|
|
g_assert (device != nil);
|
|
|
|
GST_INFO ("setting up session");
|
|
|
|
s = gst_caps_get_structure (caps, 0);
|
|
gst_structure_get_int (s, "width", &width);
|
|
gst_structure_get_int (s, "height", &height);
|
|
if (!gst_structure_get_fraction (s, "framerate", &fps_n, &fps_d)) {
|
|
fps_n = DEVICE_FPS_N;
|
|
fps_d = DEVICE_FPS_D;
|
|
}
|
|
|
|
GST_INFO ("got caps %dx%d %d/%d", width, height, fps_n, fps_d);
|
|
input = [[QTCaptureDeviceInput alloc] initWithDevice:device];
|
|
|
|
output = [[QTCaptureDecompressedVideoOutput alloc] init];
|
|
#if MAC_OS_X_VERSION_MIN_REQUIRED >= 1060
|
|
[output setAutomaticallyDropsLateVideoFrames:YES];
|
|
#endif
|
|
outputAttrs = [NSDictionary dictionaryWithObjectsAndKeys:
|
|
[NSNumber numberWithUnsignedInt:k2vuyPixelFormat],
|
|
(id)kCVPixelBufferPixelFormatTypeKey,
|
|
[NSNumber numberWithUnsignedInt:width],
|
|
(id)kCVPixelBufferWidthKey,
|
|
[NSNumber numberWithUnsignedInt:height],
|
|
(id)kCVPixelBufferHeightKey,
|
|
nil
|
|
];
|
|
[output setPixelBufferAttributes:outputAttrs];
|
|
|
|
if (fps_n != 0) {
|
|
duration = gst_util_uint64_scale (GST_SECOND, fps_d, fps_n);
|
|
gst_util_fraction_to_double (fps_d, fps_n, (gdouble *) &interval);
|
|
} else {
|
|
duration = GST_CLOCK_TIME_NONE;
|
|
interval = 0;
|
|
}
|
|
|
|
[output setMinimumVideoFrameInterval:interval];
|
|
|
|
session = [[QTCaptureSession alloc] init];
|
|
success = [session addInput:input
|
|
error:nil];
|
|
g_assert (success);
|
|
success = [session addOutput:output
|
|
error:nil];
|
|
g_assert (success);
|
|
|
|
gst_gl_context_helper_ensure_context (ctxh);
|
|
GST_INFO_OBJECT (element, "pushing textures, context %p old context %p",
|
|
ctxh->context, textureCache ? textureCache->ctx : NULL);
|
|
if (textureCache && textureCache->ctx != ctxh->context) {
|
|
gst_video_texture_cache_free (textureCache);
|
|
textureCache = NULL;
|
|
}
|
|
textureCache = gst_video_texture_cache_new (ctxh->context);
|
|
gst_video_texture_cache_set_format (textureCache, GST_VIDEO_FORMAT_UYVY, caps);
|
|
gst_video_info_set_format (&outputInfo, GST_VIDEO_FORMAT_UYVY, width, height);
|
|
|
|
[output setDelegate:self];
|
|
[session startRunning];
|
|
|
|
return YES;
|
|
}
|
|
|
|
- (BOOL)start
|
|
{
|
|
NSRunLoop *mainRunLoop;
|
|
|
|
queueLock = [[NSConditionLock alloc] initWithCondition:NO_FRAMES];
|
|
queue = [[NSMutableArray alloc] initWithCapacity:FRAME_QUEUE_SIZE];
|
|
stopRequest = NO;
|
|
|
|
offset = 0;
|
|
width = height = 0;
|
|
fps_n = 0;
|
|
fps_d = 1;
|
|
duration = GST_CLOCK_TIME_NONE;
|
|
ctxh = gst_gl_context_helper_new (element);
|
|
|
|
/* this will trigger negotiation and open the device in setCaps */
|
|
gst_base_src_start_complete (baseSrc, GST_FLOW_OK);
|
|
|
|
mainRunLoop = [NSRunLoop mainRunLoop];
|
|
if ([mainRunLoop currentMode] == nil) {
|
|
/* QTCaptureSession::addInput and QTCaptureSession::addOutput, called from
|
|
* setCaps, call NSObject::performSelectorOnMainThread internally. If the
|
|
* mainRunLoop is not running we need to run it for a while for those
|
|
* methods to complete.
|
|
*/
|
|
GST_INFO ("mainRunLoop not running");
|
|
[[NSRunLoop mainRunLoop] runUntilDate:[NSDate dateWithTimeIntervalSinceNow:1]];
|
|
}
|
|
|
|
return YES;
|
|
}
|
|
|
|
- (BOOL)stop
|
|
{
|
|
[session stopRunning];
|
|
[output setDelegate:nil];
|
|
|
|
[queueLock release];
|
|
queueLock = nil;
|
|
[queue release];
|
|
queue = nil;
|
|
|
|
gst_gl_context_helper_free (ctxh);
|
|
ctxh = NULL;
|
|
if (textureCache)
|
|
gst_video_texture_cache_free (textureCache);
|
|
textureCache = NULL;
|
|
|
|
return YES;
|
|
}
|
|
|
|
- (BOOL)query:(GstQuery *)query
|
|
{
|
|
BOOL result = NO;
|
|
|
|
if (GST_QUERY_TYPE (query) == GST_QUERY_LATENCY) {
|
|
if (device != nil) {
|
|
GstClockTime min_latency, max_latency;
|
|
|
|
min_latency = max_latency = duration; /* for now */
|
|
result = YES;
|
|
|
|
GST_DEBUG_OBJECT (element, "reporting latency of min %" GST_TIME_FORMAT
|
|
" max %" GST_TIME_FORMAT,
|
|
GST_TIME_ARGS (min_latency), GST_TIME_ARGS (max_latency));
|
|
gst_query_set_latency (query, TRUE, min_latency, max_latency);
|
|
}
|
|
} else {
|
|
result = GST_BASE_SRC_CLASS (gst_qtkit_video_src_parent_class)->query (baseSrc, query);
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
- (BOOL)unlock
|
|
{
|
|
[queueLock lock];
|
|
stopRequest = YES;
|
|
[queueLock unlockWithCondition:HAS_FRAME_OR_STOP_REQUEST];
|
|
|
|
return YES;
|
|
}
|
|
|
|
- (BOOL)unlockStop
|
|
{
|
|
[queueLock lock];
|
|
stopRequest = NO;
|
|
[queueLock unlockWithCondition:NO_FRAMES];
|
|
|
|
return YES;
|
|
}
|
|
|
|
- (GstCaps *)fixate:(GstCaps *)caps
|
|
{
|
|
GstStructure *structure;
|
|
|
|
caps = gst_caps_truncate (caps);
|
|
structure = gst_caps_get_structure (caps, 0);
|
|
if (gst_structure_has_field (structure, "framerate")) {
|
|
gst_structure_fixate_field_nearest_fraction (structure, "framerate",
|
|
DEVICE_FPS_N, DEVICE_FPS_D);
|
|
}
|
|
|
|
return caps;
|
|
}
|
|
|
|
- (GstStateChangeReturn)changeState:(GstStateChange)transition
|
|
{
|
|
GstStateChangeReturn ret;
|
|
|
|
if (transition == GST_STATE_CHANGE_NULL_TO_READY) {
|
|
if (![self openDevice])
|
|
return GST_STATE_CHANGE_FAILURE;
|
|
}
|
|
|
|
ret = GST_ELEMENT_CLASS (gst_qtkit_video_src_parent_class)->change_state (element, transition);
|
|
|
|
if (transition == GST_STATE_CHANGE_READY_TO_NULL)
|
|
[self closeDevice];
|
|
|
|
return ret;
|
|
}
|
|
|
|
- (void)captureOutput:(QTCaptureOutput *)captureOutput
|
|
didOutputVideoFrame:(CVImageBufferRef)videoFrame
|
|
withSampleBuffer:(QTSampleBuffer *)sampleBuffer
|
|
fromConnection:(QTCaptureConnection *)connection
|
|
{
|
|
[queueLock lock];
|
|
|
|
if (stopRequest) {
|
|
[queueLock unlock];
|
|
return;
|
|
}
|
|
|
|
GST_INFO ("got new frame");
|
|
|
|
if ([queue count] == FRAME_QUEUE_SIZE)
|
|
[queue removeLastObject];
|
|
|
|
[queue insertObject:(id)videoFrame
|
|
atIndex:0];
|
|
|
|
[queueLock unlockWithCondition:HAS_FRAME_OR_STOP_REQUEST];
|
|
}
|
|
|
|
- (GstFlowReturn)create:(GstBuffer **)buf
|
|
{
|
|
CVPixelBufferRef frame;
|
|
|
|
[queueLock lockWhenCondition:HAS_FRAME_OR_STOP_REQUEST];
|
|
if (stopRequest) {
|
|
[queueLock unlock];
|
|
return GST_FLOW_FLUSHING;
|
|
}
|
|
|
|
frame = (CVPixelBufferRef) [queue lastObject];
|
|
CVBufferRetain (frame);
|
|
[queue removeLastObject];
|
|
[queueLock unlockWithCondition:
|
|
([queue count] == 0) ? NO_FRAMES : HAS_FRAME_OR_STOP_REQUEST];
|
|
|
|
*buf = gst_core_video_buffer_new ((CVBufferRef)frame, &outputInfo, textureCache);
|
|
CVBufferRelease (frame);
|
|
|
|
[self timestampBuffer:*buf];
|
|
|
|
return GST_FLOW_OK;
|
|
}
|
|
|
|
- (void)timestampBuffer:(GstBuffer *)buf
|
|
{
|
|
GstClock *clock;
|
|
GstClockTime timestamp;
|
|
|
|
GST_OBJECT_LOCK (element);
|
|
clock = GST_ELEMENT_CLOCK (element);
|
|
if (clock != NULL) {
|
|
gst_object_ref (clock);
|
|
timestamp = element->base_time;
|
|
} else {
|
|
timestamp = GST_CLOCK_TIME_NONE;
|
|
}
|
|
GST_OBJECT_UNLOCK (element);
|
|
|
|
if (clock != NULL) {
|
|
timestamp = gst_clock_get_time (clock) - timestamp;
|
|
if (timestamp > duration)
|
|
timestamp -= duration;
|
|
else
|
|
timestamp = 0;
|
|
|
|
gst_object_unref (clock);
|
|
clock = NULL;
|
|
}
|
|
|
|
GST_BUFFER_OFFSET (buf) = offset++;
|
|
GST_BUFFER_OFFSET_END (buf) = GST_BUFFER_OFFSET (buf) + 1;
|
|
GST_BUFFER_TIMESTAMP (buf) = timestamp;
|
|
GST_BUFFER_DURATION (buf) = duration;
|
|
}
|
|
|
|
@end
|
|
|
|
/*
|
|
* Glue code
|
|
*/
|
|
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_DEVICE_INDEX
|
|
};
|
|
|
|
static void gst_qtkit_video_src_finalize (GObject * obj);
|
|
static void gst_qtkit_video_src_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec);
|
|
static void gst_qtkit_video_src_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec);
|
|
static GstStateChangeReturn gst_qtkit_video_src_change_state (
|
|
GstElement * element, GstStateChange transition);
|
|
static gboolean gst_qtkit_video_src_set_caps (GstBaseSrc * basesrc,
|
|
GstCaps * caps);
|
|
static gboolean gst_qtkit_video_src_start (GstBaseSrc * basesrc);
|
|
static gboolean gst_qtkit_video_src_stop (GstBaseSrc * basesrc);
|
|
static gboolean gst_qtkit_video_src_query (GstBaseSrc * basesrc,
|
|
GstQuery * query);
|
|
static gboolean gst_qtkit_video_src_unlock (GstBaseSrc * basesrc);
|
|
static gboolean gst_qtkit_video_src_unlock_stop (GstBaseSrc * basesrc);
|
|
static GstFlowReturn gst_qtkit_video_src_create (GstPushSrc * pushsrc,
|
|
GstBuffer ** buf);
|
|
static GstCaps * gst_qtkit_video_src_fixate (GstBaseSrc * basesrc, GstCaps * caps);
|
|
|
|
static void
|
|
gst_qtkit_video_src_class_init (GstQTKitVideoSrcClass * klass)
|
|
{
|
|
GObjectClass *gobject_class = G_OBJECT_CLASS (klass);
|
|
GstElementClass *gstelement_class = GST_ELEMENT_CLASS (klass);
|
|
GstBaseSrcClass *gstbasesrc_class = GST_BASE_SRC_CLASS (klass);
|
|
GstPushSrcClass *gstpushsrc_class = GST_PUSH_SRC_CLASS (klass);
|
|
|
|
gst_element_class_set_metadata (gstelement_class,
|
|
"Video Source (QTKit)", "Source/Video",
|
|
"Reads frames from a Mac OS X QTKit device",
|
|
"Ole André Vadla Ravnås <oleavr@soundrop.com>");
|
|
|
|
gst_element_class_add_static_pad_template (gstelement_class, &src_template);
|
|
|
|
gobject_class->finalize = gst_qtkit_video_src_finalize;
|
|
gobject_class->get_property = gst_qtkit_video_src_get_property;
|
|
gobject_class->set_property = gst_qtkit_video_src_set_property;
|
|
|
|
gstelement_class->change_state = gst_qtkit_video_src_change_state;
|
|
|
|
gstbasesrc_class->set_caps = gst_qtkit_video_src_set_caps;
|
|
gstbasesrc_class->start = gst_qtkit_video_src_start;
|
|
gstbasesrc_class->stop = gst_qtkit_video_src_stop;
|
|
gstbasesrc_class->query = gst_qtkit_video_src_query;
|
|
gstbasesrc_class->unlock = gst_qtkit_video_src_unlock;
|
|
gstbasesrc_class->unlock_stop = gst_qtkit_video_src_unlock_stop;
|
|
gstbasesrc_class->fixate = gst_qtkit_video_src_fixate;
|
|
|
|
gstpushsrc_class->create = gst_qtkit_video_src_create;
|
|
|
|
g_object_class_install_property (gobject_class, PROP_DEVICE_INDEX,
|
|
g_param_spec_int ("device-index", "Device Index",
|
|
"The zero-based device index",
|
|
-1, G_MAXINT, DEFAULT_DEVICE_INDEX,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
|
|
GST_DEBUG_CATEGORY_INIT (gst_qtkit_video_src_debug, "qtkitvideosrc",
|
|
0, "Mac OS X QTKit video source");
|
|
}
|
|
|
|
#define OBJC_CALLOUT_BEGIN() \
|
|
NSAutoreleasePool *pool; \
|
|
\
|
|
pool = [[NSAutoreleasePool alloc] init]
|
|
#define OBJC_CALLOUT_END() \
|
|
[pool release]
|
|
|
|
static void
|
|
gst_qtkit_video_src_init (GstQTKitVideoSrc * src)
|
|
{
|
|
OBJC_CALLOUT_BEGIN ();
|
|
src->impl = [[GstQTKitVideoSrcImpl alloc] initWithSrc:GST_PUSH_SRC (src)];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
/* pretend to be async so we can spin the mainRunLoop from the main thread if
|
|
* needed (see ::start) */
|
|
gst_base_src_set_async (GST_BASE_SRC (src), TRUE);
|
|
}
|
|
|
|
static void
|
|
gst_qtkit_video_src_finalize (GObject * obj)
|
|
{
|
|
OBJC_CALLOUT_BEGIN ();
|
|
[GST_QTKIT_VIDEO_SRC_IMPL (obj) release];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
G_OBJECT_CLASS (gst_qtkit_video_src_parent_class)->finalize (obj);
|
|
}
|
|
|
|
static void
|
|
gst_qtkit_video_src_get_property (GObject * object, guint prop_id, GValue * value,
|
|
GParamSpec * pspec)
|
|
{
|
|
GstQTKitVideoSrcImpl *impl = GST_QTKIT_VIDEO_SRC_IMPL (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE_INDEX:
|
|
g_value_set_int (value, impl.deviceIndex);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_qtkit_video_src_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstQTKitVideoSrcImpl *impl = GST_QTKIT_VIDEO_SRC_IMPL (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_DEVICE_INDEX:
|
|
impl.deviceIndex = g_value_get_int (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static GstStateChangeReturn
|
|
gst_qtkit_video_src_change_state (GstElement * element, GstStateChange transition)
|
|
{
|
|
GstStateChangeReturn ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (element) changeState: transition];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_qtkit_video_src_set_caps (GstBaseSrc * basesrc, GstCaps * caps)
|
|
{
|
|
gboolean ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) setCaps:caps];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_qtkit_video_src_start (GstBaseSrc * basesrc)
|
|
{
|
|
gboolean ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) start];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_qtkit_video_src_stop (GstBaseSrc * basesrc)
|
|
{
|
|
gboolean ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) stop];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_qtkit_video_src_query (GstBaseSrc * basesrc, GstQuery * query)
|
|
{
|
|
gboolean ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) query:query];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_qtkit_video_src_unlock (GstBaseSrc * basesrc)
|
|
{
|
|
gboolean ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) unlock];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
gst_qtkit_video_src_unlock_stop (GstBaseSrc * basesrc)
|
|
{
|
|
gboolean ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) unlockStop];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_qtkit_video_src_create (GstPushSrc * pushsrc, GstBuffer ** buf)
|
|
{
|
|
GstFlowReturn ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (pushsrc) create: buf];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstCaps *
|
|
gst_qtkit_video_src_fixate (GstBaseSrc * basesrc, GstCaps * caps)
|
|
{
|
|
GstCaps *ret;
|
|
|
|
OBJC_CALLOUT_BEGIN ();
|
|
ret = [GST_QTKIT_VIDEO_SRC_IMPL (basesrc) fixate: caps];
|
|
OBJC_CALLOUT_END ();
|
|
|
|
return ret;
|
|
}
|