mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-24 09:10:36 +00:00
2e0d8e8b93
Original commit message from CVS: * sys/v4l2/v4l2src_calls.c: (gst_v4l2src_probe_caps_for_format_and_size), (sort_by_frame_size), (gst_v4l2src_probe_caps_for_format): Turns out we don't always get the frame sizes in a predefined order from lowest to highest resolution, so let's just sort the list by frame size once we've queried the possible resolutions rather than assume any particular order. Fixes probed caps for the camera in my HP2133 mini notebook and makes v4l2src default to a decent size.
1623 lines
49 KiB
C
1623 lines
49 KiB
C
/* GStreamer
|
|
*
|
|
* Copyright (C) 2002 Ronald Bultje <rbultje@ronald.bitfreak.net>
|
|
* 2006 Edgard Lima <edgard.lima@indt.org.br>
|
|
*
|
|
* v4l2src.c - system calls
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
|
* Boston, MA 02111-1307, USA.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include <stdlib.h>
|
|
#include <sys/types.h>
|
|
#include <sys/stat.h>
|
|
#include <fcntl.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/mman.h>
|
|
#include <string.h>
|
|
#include <errno.h>
|
|
#include "v4l2src_calls.h"
|
|
#include <sys/time.h>
|
|
#include <unistd.h>
|
|
#ifdef __sun
|
|
/* Needed on older Solaris Nevada builds (72 at least) */
|
|
#include <stropts.h>
|
|
#include <sys/ioccom.h>
|
|
#endif
|
|
|
|
#include "gstv4l2tuner.h"
|
|
|
|
GST_DEBUG_CATEGORY_EXTERN (v4l2src_debug);
|
|
#define GST_CAT_DEFAULT v4l2src_debug
|
|
|
|
/* lalala... */
|
|
#define GST_V4L2_SET_ACTIVE(element) (element)->buffer = GINT_TO_POINTER (-1)
|
|
#define GST_V4L2_SET_INACTIVE(element) (element)->buffer = NULL
|
|
|
|
/* On some systems MAP_FAILED seems to be missing */
|
|
#ifndef MAP_FAILED
|
|
#define MAP_FAILED ((caddr_t) -1)
|
|
#endif
|
|
|
|
|
|
#define GST_TYPE_V4L2_BUFFER (gst_v4l2_buffer_get_type())
|
|
#define GST_IS_V4L2_BUFFER(obj) (G_TYPE_CHECK_INSTANCE_TYPE ((obj), GST_TYPE_V4L2_BUFFER))
|
|
#define GST_V4L2_BUFFER(obj) (G_TYPE_CHECK_INSTANCE_CAST ((obj), GST_TYPE_V4L2_BUFFER, GstV4l2Buffer))
|
|
|
|
static GstBufferClass *v4l2buffer_parent_class = NULL;
|
|
|
|
/* Local functions */
|
|
static gboolean
|
|
gst_v4l2src_get_nearest_size (GstV4l2Src * v4l2src, guint32 pixelformat,
|
|
gint * width, gint * height);
|
|
static void gst_v4l2_buffer_pool_destroy (GstV4l2BufferPool * pool);
|
|
|
|
static void
|
|
gst_v4l2_buffer_finalize (GstV4l2Buffer * buffer)
|
|
{
|
|
GstV4l2BufferPool *pool;
|
|
gboolean resuscitated = FALSE;
|
|
gint index;
|
|
|
|
pool = buffer->pool;
|
|
|
|
index = buffer->vbuffer.index;
|
|
|
|
GST_LOG ("finalizing buffer %p %d", buffer, index);
|
|
|
|
g_mutex_lock (pool->lock);
|
|
if (GST_BUFFER_SIZE (buffer) != 0)
|
|
/* BUFFER_SIZE is only set if the frame was dequeued */
|
|
pool->num_live_buffers--;
|
|
|
|
if (pool->running) {
|
|
if (v4l2_ioctl (pool->video_fd, VIDIOC_QBUF, &buffer->vbuffer) < 0) {
|
|
GST_WARNING ("could not requeue buffer %p %d", buffer, index);
|
|
} else {
|
|
/* FIXME: check that the caps didn't change */
|
|
GST_LOG ("reviving buffer %p, %d", buffer, index);
|
|
gst_buffer_ref (GST_BUFFER (buffer));
|
|
GST_BUFFER_SIZE (buffer) = 0;
|
|
pool->buffers[index] = buffer;
|
|
resuscitated = TRUE;
|
|
}
|
|
} else {
|
|
GST_LOG ("the pool is shutting down");
|
|
}
|
|
g_mutex_unlock (pool->lock);
|
|
|
|
if (!resuscitated) {
|
|
GST_LOG ("buffer %p not recovered, unmapping", buffer);
|
|
gst_mini_object_unref (GST_MINI_OBJECT (pool));
|
|
v4l2_munmap ((void *) GST_BUFFER_DATA (buffer), buffer->vbuffer.length);
|
|
|
|
GST_MINI_OBJECT_CLASS (v4l2buffer_parent_class)->finalize (GST_MINI_OBJECT
|
|
(buffer));
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_v4l2_buffer_class_init (gpointer g_class, gpointer class_data)
|
|
{
|
|
GstMiniObjectClass *mini_object_class = GST_MINI_OBJECT_CLASS (g_class);
|
|
|
|
v4l2buffer_parent_class = g_type_class_peek_parent (g_class);
|
|
|
|
mini_object_class->finalize = (GstMiniObjectFinalizeFunction)
|
|
gst_v4l2_buffer_finalize;
|
|
}
|
|
|
|
static GType
|
|
gst_v4l2_buffer_get_type (void)
|
|
{
|
|
static GType _gst_v4l2_buffer_type;
|
|
|
|
if (G_UNLIKELY (_gst_v4l2_buffer_type == 0)) {
|
|
static const GTypeInfo v4l2_buffer_info = {
|
|
sizeof (GstBufferClass),
|
|
NULL,
|
|
NULL,
|
|
gst_v4l2_buffer_class_init,
|
|
NULL,
|
|
NULL,
|
|
sizeof (GstV4l2Buffer),
|
|
0,
|
|
NULL,
|
|
NULL
|
|
};
|
|
_gst_v4l2_buffer_type = g_type_register_static (GST_TYPE_BUFFER,
|
|
"GstV4l2Buffer", &v4l2_buffer_info, 0);
|
|
}
|
|
return _gst_v4l2_buffer_type;
|
|
}
|
|
|
|
static GstV4l2Buffer *
|
|
gst_v4l2_buffer_new (GstV4l2BufferPool * pool, guint index, GstCaps * caps)
|
|
{
|
|
GstV4l2Buffer *ret;
|
|
guint8 *data;
|
|
|
|
ret = (GstV4l2Buffer *) gst_mini_object_new (GST_TYPE_V4L2_BUFFER);
|
|
|
|
GST_LOG ("creating buffer %u, %p in pool %p", index, ret, pool);
|
|
|
|
ret->pool =
|
|
(GstV4l2BufferPool *) gst_mini_object_ref (GST_MINI_OBJECT (pool));
|
|
|
|
ret->vbuffer.index = index;
|
|
ret->vbuffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
ret->vbuffer.memory = V4L2_MEMORY_MMAP;
|
|
|
|
if (v4l2_ioctl (pool->video_fd, VIDIOC_QUERYBUF, &ret->vbuffer) < 0)
|
|
goto querybuf_failed;
|
|
|
|
GST_LOG (" index: %u", ret->vbuffer.index);
|
|
GST_LOG (" type: %d", ret->vbuffer.type);
|
|
GST_LOG (" bytesused: %u", ret->vbuffer.bytesused);
|
|
GST_LOG (" flags: %08x", ret->vbuffer.flags);
|
|
GST_LOG (" field: %d", ret->vbuffer.field);
|
|
GST_LOG (" memory: %d", ret->vbuffer.memory);
|
|
if (ret->vbuffer.memory == V4L2_MEMORY_MMAP)
|
|
GST_LOG (" MMAP offset: %u", ret->vbuffer.m.offset);
|
|
GST_LOG (" length: %u", ret->vbuffer.length);
|
|
GST_LOG (" input: %u", ret->vbuffer.input);
|
|
|
|
data = (guint8 *) v4l2_mmap (0, ret->vbuffer.length,
|
|
PROT_READ | PROT_WRITE, MAP_SHARED, pool->video_fd,
|
|
ret->vbuffer.m.offset);
|
|
|
|
if (data == MAP_FAILED)
|
|
goto mmap_failed;
|
|
|
|
GST_BUFFER_DATA (ret) = data;
|
|
GST_BUFFER_SIZE (ret) = ret->vbuffer.length;
|
|
|
|
GST_BUFFER_FLAG_SET (ret, GST_BUFFER_FLAG_READONLY);
|
|
|
|
gst_buffer_set_caps (GST_BUFFER (ret), caps);
|
|
|
|
return ret;
|
|
|
|
/* ERRORS */
|
|
querybuf_failed:
|
|
{
|
|
gint errnosave = errno;
|
|
|
|
GST_WARNING ("Failed QUERYBUF: %s", g_strerror (errnosave));
|
|
gst_buffer_unref (GST_BUFFER (ret));
|
|
errno = errnosave;
|
|
return NULL;
|
|
}
|
|
mmap_failed:
|
|
{
|
|
gint errnosave = errno;
|
|
|
|
GST_WARNING ("Failed to mmap: %s", g_strerror (errnosave));
|
|
gst_buffer_unref (GST_BUFFER (ret));
|
|
errno = errnosave;
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
#define GST_TYPE_V4L2_BUFFER_POOL (gst_v4l2_buffer_pool_get_type())
|
|
#define GST_IS_V4L2_BUFFER_POOL(obj) (G_TYPE_CHECK_INSTANCE_TYPE ((obj), GST_TYPE_V4L2_BUFFER_POOL))
|
|
#define GST_V4L2_BUFFER_POOL(obj) (G_TYPE_CHECK_INSTANCE_CAST ((obj), GST_TYPE_V4L2_BUFFER_POOL, GstV4l2BufferPool))
|
|
|
|
static GstMiniObjectClass *buffer_pool_parent_class = NULL;
|
|
|
|
static void
|
|
gst_v4l2_buffer_pool_finalize (GstV4l2BufferPool * pool)
|
|
{
|
|
g_mutex_free (pool->lock);
|
|
pool->lock = NULL;
|
|
|
|
if (pool->video_fd >= 0)
|
|
v4l2_close (pool->video_fd);
|
|
|
|
if (pool->buffers) {
|
|
g_free (pool->buffers);
|
|
pool->buffers = NULL;
|
|
}
|
|
|
|
GST_MINI_OBJECT_CLASS (buffer_pool_parent_class)->finalize (GST_MINI_OBJECT
|
|
(pool));
|
|
}
|
|
|
|
static void
|
|
gst_v4l2_buffer_pool_init (GstV4l2BufferPool * pool, gpointer g_class)
|
|
{
|
|
pool->lock = g_mutex_new ();
|
|
pool->running = FALSE;
|
|
pool->num_live_buffers = 0;
|
|
}
|
|
|
|
static void
|
|
gst_v4l2_buffer_pool_class_init (gpointer g_class, gpointer class_data)
|
|
{
|
|
GstMiniObjectClass *mini_object_class = GST_MINI_OBJECT_CLASS (g_class);
|
|
|
|
buffer_pool_parent_class = g_type_class_peek_parent (g_class);
|
|
|
|
mini_object_class->finalize = (GstMiniObjectFinalizeFunction)
|
|
gst_v4l2_buffer_pool_finalize;
|
|
}
|
|
|
|
static GType
|
|
gst_v4l2_buffer_pool_get_type (void)
|
|
{
|
|
static GType _gst_v4l2_buffer_pool_type;
|
|
|
|
if (G_UNLIKELY (_gst_v4l2_buffer_pool_type == 0)) {
|
|
static const GTypeInfo v4l2_buffer_pool_info = {
|
|
sizeof (GstBufferClass),
|
|
NULL,
|
|
NULL,
|
|
gst_v4l2_buffer_pool_class_init,
|
|
NULL,
|
|
NULL,
|
|
sizeof (GstV4l2BufferPool),
|
|
0,
|
|
(GInstanceInitFunc) gst_v4l2_buffer_pool_init,
|
|
NULL
|
|
};
|
|
_gst_v4l2_buffer_pool_type = g_type_register_static (GST_TYPE_MINI_OBJECT,
|
|
"GstV4l2BufferPool", &v4l2_buffer_pool_info, 0);
|
|
}
|
|
return _gst_v4l2_buffer_pool_type;
|
|
}
|
|
|
|
static GstV4l2BufferPool *
|
|
gst_v4l2_buffer_pool_new (GstV4l2Src * v4l2src, gint fd, gint num_buffers,
|
|
GstCaps * caps)
|
|
{
|
|
GstV4l2BufferPool *pool;
|
|
gint n;
|
|
|
|
pool = (GstV4l2BufferPool *) gst_mini_object_new (GST_TYPE_V4L2_BUFFER_POOL);
|
|
|
|
pool->video_fd = v4l2_dup (fd);
|
|
if (pool->video_fd < 0)
|
|
goto dup_failed;
|
|
|
|
pool->buffer_count = num_buffers;
|
|
pool->buffers = g_new0 (GstV4l2Buffer *, num_buffers);
|
|
|
|
for (n = 0; n < num_buffers; n++) {
|
|
pool->buffers[n] = gst_v4l2_buffer_new (pool, n, caps);
|
|
if (!pool->buffers[n])
|
|
goto buffer_new_failed;
|
|
}
|
|
|
|
return pool;
|
|
|
|
/* ERRORS */
|
|
dup_failed:
|
|
{
|
|
gint errnosave = errno;
|
|
|
|
gst_mini_object_unref (GST_MINI_OBJECT (pool));
|
|
|
|
errno = errnosave;
|
|
|
|
return NULL;
|
|
}
|
|
buffer_new_failed:
|
|
{
|
|
gint errnosave = errno;
|
|
|
|
gst_v4l2_buffer_pool_destroy (pool);
|
|
|
|
errno = errnosave;
|
|
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
static gboolean
|
|
gst_v4l2_buffer_pool_activate (GstV4l2BufferPool * pool, GstV4l2Src * v4l2src)
|
|
{
|
|
gint n;
|
|
|
|
g_mutex_lock (pool->lock);
|
|
|
|
for (n = 0; n < pool->buffer_count; n++) {
|
|
struct v4l2_buffer *buf;
|
|
|
|
buf = &pool->buffers[n]->vbuffer;
|
|
|
|
GST_LOG ("enqueue pool buffer %d", n);
|
|
|
|
if (v4l2_ioctl (pool->video_fd, VIDIOC_QBUF, buf) < 0)
|
|
goto queue_failed;
|
|
}
|
|
pool->running = TRUE;
|
|
|
|
g_mutex_unlock (pool->lock);
|
|
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
queue_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ,
|
|
(_("Could not enqueue buffers in device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
("enqueing buffer %d/%d failed: %s",
|
|
n, v4l2src->num_buffers, g_strerror (errno)));
|
|
g_mutex_unlock (pool->lock);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_v4l2_buffer_pool_destroy (GstV4l2BufferPool * pool)
|
|
{
|
|
gint n;
|
|
|
|
g_mutex_lock (pool->lock);
|
|
pool->running = FALSE;
|
|
g_mutex_unlock (pool->lock);
|
|
|
|
GST_DEBUG ("destroy pool");
|
|
|
|
/* after this point, no more buffers will be queued or dequeued; no buffer
|
|
* from pool->buffers that is NULL will be set to a buffer, and no buffer that
|
|
* is not NULL will be pushed out. */
|
|
|
|
/* miniobjects have no dispose, so they can't break ref-cycles, as buffers ref
|
|
* the pool, we need to unref the buffer to properly finalize te pool */
|
|
for (n = 0; n < pool->buffer_count; n++) {
|
|
GstBuffer *buf;
|
|
|
|
g_mutex_lock (pool->lock);
|
|
buf = GST_BUFFER (pool->buffers[n]);
|
|
g_mutex_unlock (pool->lock);
|
|
|
|
if (buf)
|
|
/* we own the ref if the buffer is in pool->buffers; drop it. */
|
|
gst_buffer_unref (buf);
|
|
}
|
|
|
|
gst_mini_object_unref (GST_MINI_OBJECT (pool));
|
|
}
|
|
|
|
/* complete made up ranking, the values themselves are meaningless */
|
|
#define YUV_BASE_RANK 1000
|
|
#define JPEG_BASE_RANK 500
|
|
#define DV_BASE_RANK 200
|
|
#define RGB_BASE_RANK 100
|
|
#define YUV_ODD_BASE_RANK 50
|
|
#define RGB_ODD_BASE_RANK 25
|
|
#define BAYER_BASE_RANK 15
|
|
#define S910_BASE_RANK 10
|
|
#define GREY_BASE_RANK 5
|
|
#define PWC_BASE_RANK 1
|
|
|
|
static gint
|
|
gst_v4l2src_format_get_rank (guint32 fourcc)
|
|
{
|
|
switch (fourcc) {
|
|
case V4L2_PIX_FMT_MJPEG:
|
|
return JPEG_BASE_RANK;
|
|
case V4L2_PIX_FMT_JPEG:
|
|
return JPEG_BASE_RANK + 1;
|
|
|
|
case V4L2_PIX_FMT_RGB332:
|
|
case V4L2_PIX_FMT_RGB555:
|
|
case V4L2_PIX_FMT_RGB555X:
|
|
case V4L2_PIX_FMT_RGB565:
|
|
case V4L2_PIX_FMT_RGB565X:
|
|
return RGB_ODD_BASE_RANK;
|
|
|
|
case V4L2_PIX_FMT_RGB24:
|
|
case V4L2_PIX_FMT_BGR24:
|
|
return RGB_BASE_RANK - 1;
|
|
|
|
case V4L2_PIX_FMT_RGB32:
|
|
case V4L2_PIX_FMT_BGR32:
|
|
return RGB_BASE_RANK;
|
|
|
|
case V4L2_PIX_FMT_GREY: /* 8 Greyscale */
|
|
return GREY_BASE_RANK;
|
|
|
|
case V4L2_PIX_FMT_NV12: /* 12 Y/CbCr 4:2:0 */
|
|
case V4L2_PIX_FMT_NV21: /* 12 Y/CrCb 4:2:0 */
|
|
case V4L2_PIX_FMT_YYUV: /* 16 YUV 4:2:2 */
|
|
case V4L2_PIX_FMT_HI240: /* 8 8-bit color */
|
|
return YUV_ODD_BASE_RANK;
|
|
|
|
case V4L2_PIX_FMT_YVU410: /* YVU9, 9 bits per pixel */
|
|
return YUV_BASE_RANK + 3;
|
|
case V4L2_PIX_FMT_YUV410: /* YUV9, 9 bits per pixel */
|
|
return YUV_BASE_RANK + 2;
|
|
case V4L2_PIX_FMT_YUV420: /* I420, 12 bits per pixel */
|
|
return YUV_BASE_RANK + 7;
|
|
case V4L2_PIX_FMT_YUYV: /* YUY2, 16 bits per pixel */
|
|
return YUV_BASE_RANK + 10;
|
|
case V4L2_PIX_FMT_YVU420: /* YV12, 12 bits per pixel */
|
|
return YUV_BASE_RANK + 6;
|
|
case V4L2_PIX_FMT_UYVY: /* UYVY, 16 bits per pixel */
|
|
return YUV_BASE_RANK + 9;
|
|
case V4L2_PIX_FMT_Y41P: /* Y41P, 12 bits per pixel */
|
|
return YUV_BASE_RANK + 5;
|
|
case V4L2_PIX_FMT_YUV411P: /* Y41B, 12 bits per pixel */
|
|
return YUV_BASE_RANK + 4;
|
|
case V4L2_PIX_FMT_YUV422P: /* Y42B, 16 bits per pixel */
|
|
return YUV_BASE_RANK + 8;
|
|
|
|
case V4L2_PIX_FMT_DV:
|
|
return DV_BASE_RANK;
|
|
|
|
case V4L2_PIX_FMT_MPEG: /* MPEG */
|
|
case V4L2_PIX_FMT_WNVA: /* Winnov hw compres */
|
|
return 0;
|
|
|
|
#ifdef V4L2_PIX_FMT_SBGGR8
|
|
case V4L2_PIX_FMT_SBGGR8:
|
|
return BAYER_BASE_RANK;
|
|
#endif
|
|
|
|
#ifdef V4L2_PIX_FMT_SN9C10X
|
|
case V4L2_PIX_FMT_SN9C10X:
|
|
return S910_BASE_RANK;
|
|
#endif
|
|
|
|
#ifdef V4L2_PIX_FMT_PWC1
|
|
case V4L2_PIX_FMT_PWC1:
|
|
return PWC_BASE_RANK;
|
|
#endif
|
|
#ifdef V4L2_PIX_FMT_PWC2
|
|
case V4L2_PIX_FMT_PWC2:
|
|
return PWC_BASE_RANK;
|
|
#endif
|
|
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static gint
|
|
gst_v4l2src_format_cmp_func (gconstpointer a, gconstpointer b)
|
|
{
|
|
guint32 pf1 = ((struct v4l2_fmtdesc *) a)->pixelformat;
|
|
guint32 pf2 = ((struct v4l2_fmtdesc *) b)->pixelformat;
|
|
|
|
if (pf1 == pf2)
|
|
return 0;
|
|
|
|
return gst_v4l2src_format_get_rank (pf2) - gst_v4l2src_format_get_rank (pf1);
|
|
}
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_fill_format_list():
|
|
* create list of supported capture formats
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_fill_format_list (GstV4l2Src * v4l2src)
|
|
{
|
|
gint n;
|
|
struct v4l2_fmtdesc *format;
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "getting src format enumerations");
|
|
|
|
/* format enumeration */
|
|
for (n = 0;; n++) {
|
|
format = g_new0 (struct v4l2_fmtdesc, 1);
|
|
|
|
format->index = n;
|
|
format->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
|
|
if (v4l2_ioctl (v4l2src->v4l2object->video_fd, VIDIOC_ENUM_FMT, format) < 0) {
|
|
if (errno == EINVAL) {
|
|
g_free (format);
|
|
break; /* end of enumeration */
|
|
} else {
|
|
goto failed;
|
|
}
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src, "index: %u", format->index);
|
|
GST_LOG_OBJECT (v4l2src, "type: %d", format->type);
|
|
GST_LOG_OBJECT (v4l2src, "flags: %08x", format->flags);
|
|
GST_LOG_OBJECT (v4l2src, "description: '%s'", format->description);
|
|
GST_LOG_OBJECT (v4l2src, "pixelformat: %" GST_FOURCC_FORMAT,
|
|
GST_FOURCC_ARGS (format->pixelformat));
|
|
|
|
/* sort formats according to our preference; we do this, because caps
|
|
* are probed in the order the formats are in the list, and the order of
|
|
* formats in the final probed caps matters for things like fixation */
|
|
v4l2src->formats = g_slist_insert_sorted (v4l2src->formats, format,
|
|
(GCompareFunc) gst_v4l2src_format_cmp_func);
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "got %d format(s)", n);
|
|
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Failed to enumerate possible video formats device '%s' can work with"), v4l2src->v4l2object->videodev), ("Failed to get number %d in pixelformat enumeration for %s. (%d - %s)", n, v4l2src->v4l2object->videodev, errno, g_strerror (errno)));
|
|
g_free (format);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_clear_format_list():
|
|
* free list of supported capture formats
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_clear_format_list (GstV4l2Src * v4l2src)
|
|
{
|
|
g_slist_foreach (v4l2src->formats, (GFunc) g_free, NULL);
|
|
g_slist_free (v4l2src->formats);
|
|
v4l2src->formats = NULL;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/* The frame interval enumeration code first appeared in Linux 2.6.19. */
|
|
#ifdef VIDIOC_ENUM_FRAMEINTERVALS
|
|
static GstStructure *
|
|
gst_v4l2src_probe_caps_for_format_and_size (GstV4l2Src * v4l2src,
|
|
guint32 pixelformat,
|
|
guint32 width, guint32 height, const GstStructure * template)
|
|
{
|
|
gint fd = v4l2src->v4l2object->video_fd;
|
|
struct v4l2_frmivalenum ival;
|
|
guint32 num, denom;
|
|
GstStructure *s;
|
|
GValue rates = { 0, };
|
|
|
|
memset (&ival, 0, sizeof (struct v4l2_frmivalenum));
|
|
ival.index = 0;
|
|
ival.pixel_format = pixelformat;
|
|
ival.width = width;
|
|
ival.height = height;
|
|
|
|
GST_LOG_OBJECT (v4l2src, "get frame interval for %ux%u, %" GST_FOURCC_FORMAT,
|
|
width, height, GST_FOURCC_ARGS (pixelformat));
|
|
|
|
/* keep in mind that v4l2 gives us frame intervals (durations); we invert the
|
|
* fraction to get framerate */
|
|
if (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMEINTERVALS, &ival) < 0)
|
|
goto enum_frameintervals_failed;
|
|
|
|
if (ival.type == V4L2_FRMIVAL_TYPE_DISCRETE) {
|
|
GValue rate = { 0, };
|
|
|
|
g_value_init (&rates, GST_TYPE_LIST);
|
|
g_value_init (&rate, GST_TYPE_FRACTION);
|
|
|
|
do {
|
|
num = ival.discrete.numerator;
|
|
denom = ival.discrete.denominator;
|
|
|
|
if (num > G_MAXINT || denom > G_MAXINT) {
|
|
/* let us hope we don't get here... */
|
|
num >>= 1;
|
|
denom >>= 1;
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src, "adding discrete framerate: %d/%d", denom, num);
|
|
|
|
/* swap to get the framerate */
|
|
gst_value_set_fraction (&rate, denom, num);
|
|
gst_value_list_append_value (&rates, &rate);
|
|
|
|
ival.index++;
|
|
} while (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMEINTERVALS, &ival) >= 0);
|
|
} else if (ival.type == V4L2_FRMIVAL_TYPE_STEPWISE) {
|
|
GValue min = { 0, };
|
|
GValue step = { 0, };
|
|
GValue max = { 0, };
|
|
gboolean added = FALSE;
|
|
guint32 minnum, mindenom;
|
|
guint32 maxnum, maxdenom;
|
|
|
|
g_value_init (&rates, GST_TYPE_LIST);
|
|
|
|
g_value_init (&min, GST_TYPE_FRACTION);
|
|
g_value_init (&step, GST_TYPE_FRACTION);
|
|
g_value_init (&max, GST_TYPE_FRACTION);
|
|
|
|
/* get the min */
|
|
minnum = ival.stepwise.min.numerator;
|
|
mindenom = ival.stepwise.min.denominator;
|
|
if (minnum > G_MAXINT || mindenom > G_MAXINT) {
|
|
minnum >>= 1;
|
|
mindenom >>= 1;
|
|
}
|
|
GST_LOG_OBJECT (v4l2src, "stepwise min frame interval: %d/%d", minnum,
|
|
mindenom);
|
|
gst_value_set_fraction (&min, minnum, mindenom);
|
|
|
|
/* get the max */
|
|
maxnum = ival.stepwise.max.numerator;
|
|
maxdenom = ival.stepwise.max.denominator;
|
|
if (maxnum > G_MAXINT || maxdenom > G_MAXINT) {
|
|
maxnum >>= 1;
|
|
maxdenom >>= 1;
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src, "stepwise max frame interval: %d/%d", maxnum,
|
|
maxdenom);
|
|
gst_value_set_fraction (&max, maxnum, maxdenom);
|
|
|
|
/* get the step */
|
|
num = ival.stepwise.step.numerator;
|
|
denom = ival.stepwise.step.denominator;
|
|
if (num > G_MAXINT || denom > G_MAXINT) {
|
|
num >>= 1;
|
|
denom >>= 1;
|
|
}
|
|
|
|
if (num == 0 || denom == 0) {
|
|
/* in this case we have a wrong fraction or no step, set the step to max
|
|
* so that we only add the min value in the loop below */
|
|
num = maxnum;
|
|
denom = maxdenom;
|
|
}
|
|
|
|
/* since we only have gst_value_fraction_subtract and not add, negate the
|
|
* numerator */
|
|
GST_LOG_OBJECT (v4l2src, "stepwise step frame interval: %d/%d", num, denom);
|
|
gst_value_set_fraction (&step, -num, denom);
|
|
|
|
while (gst_value_compare (&min, &max) <= 0) {
|
|
GValue rate = { 0, };
|
|
|
|
num = gst_value_get_fraction_numerator (&min);
|
|
denom = gst_value_get_fraction_denominator (&min);
|
|
GST_LOG_OBJECT (v4l2src, "adding stepwise framerate: %d/%d", denom, num);
|
|
|
|
/* invert to get the framerate */
|
|
g_value_init (&rate, GST_TYPE_FRACTION);
|
|
gst_value_set_fraction (&rate, denom, num);
|
|
gst_value_list_append_value (&rates, &rate);
|
|
added = TRUE;
|
|
|
|
/* we're actually adding because step was negated above. This is because
|
|
* there is no _add function... */
|
|
if (!gst_value_fraction_subtract (&min, &min, &step)) {
|
|
GST_WARNING_OBJECT (v4l2src, "could not step fraction!");
|
|
break;
|
|
}
|
|
}
|
|
if (!added) {
|
|
/* no range was added, leave the default range from the template */
|
|
GST_WARNING_OBJECT (v4l2src, "no range added, leaving default");
|
|
g_value_unset (&rates);
|
|
}
|
|
} else if (ival.type == V4L2_FRMIVAL_TYPE_CONTINUOUS) {
|
|
guint32 maxnum, maxdenom;
|
|
|
|
g_value_init (&rates, GST_TYPE_FRACTION_RANGE);
|
|
|
|
num = ival.stepwise.min.numerator;
|
|
denom = ival.stepwise.min.denominator;
|
|
if (num > G_MAXINT || denom > G_MAXINT) {
|
|
num >>= 1;
|
|
denom >>= 1;
|
|
}
|
|
|
|
maxnum = ival.stepwise.max.numerator;
|
|
maxdenom = ival.stepwise.max.denominator;
|
|
if (maxnum > G_MAXINT || maxdenom > G_MAXINT) {
|
|
maxnum >>= 1;
|
|
maxdenom >>= 1;
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src, "continuous frame interval %d/%d to %d/%d",
|
|
maxdenom, maxnum, denom, num);
|
|
|
|
gst_value_set_fraction_range_full (&rates, maxdenom, maxnum, denom, num);
|
|
} else {
|
|
goto unknown_type;
|
|
}
|
|
|
|
return_data:
|
|
s = gst_structure_copy (template);
|
|
gst_structure_set (s, "width", G_TYPE_INT, (gint) width,
|
|
"height", G_TYPE_INT, (gint) height, NULL);
|
|
|
|
if (G_IS_VALUE (&rates)) {
|
|
/* only change the framerate on the template when we have a valid probed new
|
|
* value */
|
|
gst_structure_set_value (s, "framerate", &rates);
|
|
g_value_unset (&rates);
|
|
}
|
|
return s;
|
|
|
|
/* ERRORS */
|
|
enum_frameintervals_failed:
|
|
{
|
|
GST_DEBUG_OBJECT (v4l2src,
|
|
"Unable to enumerate intervals for %" GST_FOURCC_FORMAT "@%ux%u",
|
|
GST_FOURCC_ARGS (pixelformat), width, height);
|
|
goto return_data;
|
|
}
|
|
unknown_type:
|
|
{
|
|
/* I don't see how this is actually an error, we ignore the format then */
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"Unknown frame interval type at %" GST_FOURCC_FORMAT "@%ux%u: %u",
|
|
GST_FOURCC_ARGS (pixelformat), width, height, ival.type);
|
|
return NULL;
|
|
}
|
|
}
|
|
#endif /* defined VIDIOC_ENUM_FRAMEINTERVALS */
|
|
|
|
static gint
|
|
sort_by_frame_size (GstStructure * s1, GstStructure * s2)
|
|
{
|
|
int w1, h1, w2, h2;
|
|
|
|
gst_structure_get_int (s1, "width", &w1);
|
|
gst_structure_get_int (s1, "height", &h1);
|
|
gst_structure_get_int (s2, "width", &w2);
|
|
gst_structure_get_int (s2, "height", &h2);
|
|
|
|
/* I think it's safe to assume that this won't overflow for a while */
|
|
return ((w2 * h2) - (w1 * h1));
|
|
}
|
|
|
|
GstCaps *
|
|
gst_v4l2src_probe_caps_for_format (GstV4l2Src * v4l2src, guint32 pixelformat,
|
|
const GstStructure * template)
|
|
{
|
|
GstCaps *ret = gst_caps_new_empty ();
|
|
GstStructure *tmp;
|
|
|
|
#ifdef VIDIOC_ENUM_FRAMESIZES
|
|
gint fd = v4l2src->v4l2object->video_fd;
|
|
struct v4l2_frmsizeenum size;
|
|
GList *results = NULL;
|
|
guint32 w, h;
|
|
|
|
memset (&size, 0, sizeof (struct v4l2_frmsizeenum));
|
|
size.index = 0;
|
|
size.pixel_format = pixelformat;
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "Enumerating frame sizes");
|
|
|
|
if (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMESIZES, &size) < 0)
|
|
goto enum_framesizes_failed;
|
|
|
|
if (size.type == V4L2_FRMSIZE_TYPE_DISCRETE) {
|
|
do {
|
|
GST_LOG_OBJECT (v4l2src, "got discrete frame size %dx%d",
|
|
size.discrete.width, size.discrete.height);
|
|
|
|
w = MIN (size.discrete.width, G_MAXINT);
|
|
h = MIN (size.discrete.height, G_MAXINT);
|
|
|
|
tmp = gst_v4l2src_probe_caps_for_format_and_size (v4l2src, pixelformat,
|
|
w, h, template);
|
|
|
|
if (tmp)
|
|
results = g_list_prepend (results, tmp);
|
|
|
|
size.index++;
|
|
} while (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMESIZES, &size) >= 0);
|
|
GST_DEBUG_OBJECT (v4l2src, "done iterating discrete frame sizes");
|
|
} else if (size.type == V4L2_FRMSIZE_TYPE_STEPWISE) {
|
|
GST_DEBUG_OBJECT (v4l2src, "we have stepwise frame sizes:");
|
|
GST_DEBUG_OBJECT (v4l2src, "min width: %d", size.stepwise.min_width);
|
|
GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.min_height);
|
|
GST_DEBUG_OBJECT (v4l2src, "max width: %d", size.stepwise.max_width);
|
|
GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.max_height);
|
|
GST_DEBUG_OBJECT (v4l2src, "step width: %d", size.stepwise.step_width);
|
|
GST_DEBUG_OBJECT (v4l2src, "step height: %d", size.stepwise.step_height);
|
|
|
|
for (w = size.stepwise.min_width, h = size.stepwise.min_height;
|
|
w < size.stepwise.max_width && h < size.stepwise.max_height;
|
|
w += size.stepwise.step_width, h += size.stepwise.step_height) {
|
|
if (w == 0 || h == 0)
|
|
continue;
|
|
|
|
tmp = gst_v4l2src_probe_caps_for_format_and_size (v4l2src, pixelformat,
|
|
w, h, template);
|
|
|
|
if (tmp)
|
|
results = g_list_prepend (results, tmp);
|
|
}
|
|
GST_DEBUG_OBJECT (v4l2src, "done iterating stepwise frame sizes");
|
|
} else if (size.type == V4L2_FRMSIZE_TYPE_CONTINUOUS) {
|
|
guint32 maxw, maxh;
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "we have continuous frame sizes:");
|
|
GST_DEBUG_OBJECT (v4l2src, "min width: %d", size.stepwise.min_width);
|
|
GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.min_height);
|
|
GST_DEBUG_OBJECT (v4l2src, "max width: %d", size.stepwise.max_width);
|
|
GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.max_height);
|
|
|
|
w = MAX (size.stepwise.min_width, 1);
|
|
h = MAX (size.stepwise.min_height, 1);
|
|
maxw = MIN (size.stepwise.max_width, G_MAXINT);
|
|
maxh = MIN (size.stepwise.max_height, G_MAXINT);
|
|
|
|
tmp = gst_v4l2src_probe_caps_for_format_and_size (v4l2src, pixelformat,
|
|
w, h, template);
|
|
if (tmp) {
|
|
gst_structure_set (tmp, "width", GST_TYPE_INT_RANGE, (gint) w,
|
|
(gint) maxw, "height", GST_TYPE_INT_RANGE, (gint) h, (gint) maxh,
|
|
NULL);
|
|
|
|
/* no point using the results list here, since there's only one struct */
|
|
gst_caps_append_structure (ret, tmp);
|
|
}
|
|
} else {
|
|
goto unknown_type;
|
|
}
|
|
|
|
/* we use an intermediary list to store and then sort the results of the
|
|
* probing because we can't make any assumptions about the order in which
|
|
* the driver will give us the sizes, but we want the final caps to contain
|
|
* the results starting with the highest resolution and having the lowest
|
|
* resolution last, since order in caps matters for things like fixation. */
|
|
results = g_list_sort (results, (GCompareFunc) sort_by_frame_size);
|
|
while (results != NULL) {
|
|
gst_caps_append_structure (ret, GST_STRUCTURE (results->data));
|
|
results = g_list_delete_link (results, results);
|
|
}
|
|
|
|
if (gst_caps_is_empty (ret))
|
|
goto enum_framesizes_no_results;
|
|
|
|
return ret;
|
|
|
|
/* ERRORS */
|
|
enum_framesizes_failed:
|
|
{
|
|
/* I don't see how this is actually an error */
|
|
GST_DEBUG_OBJECT (v4l2src,
|
|
"Failed to enumerate frame sizes for pixelformat %" GST_FOURCC_FORMAT
|
|
" (%s)", GST_FOURCC_ARGS (pixelformat), g_strerror (errno));
|
|
goto default_frame_sizes;
|
|
}
|
|
enum_framesizes_no_results:
|
|
{
|
|
/* it's possible that VIDIOC_ENUM_FRAMESIZES is defined but the driver in
|
|
* question doesn't actually support it yet */
|
|
GST_DEBUG_OBJECT (v4l2src, "No results for pixelformat %" GST_FOURCC_FORMAT
|
|
" enumerating frame sizes, trying fallback",
|
|
GST_FOURCC_ARGS (pixelformat));
|
|
goto default_frame_sizes;
|
|
}
|
|
unknown_type:
|
|
{
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"Unknown frame sizeenum type for pixelformat %" GST_FOURCC_FORMAT
|
|
": %u", GST_FOURCC_ARGS (pixelformat), size.type);
|
|
goto default_frame_sizes;
|
|
}
|
|
default_frame_sizes:
|
|
#endif /* defined VIDIOC_ENUM_FRAMESIZES */
|
|
{
|
|
gint min_w, max_w, min_h, max_h, fix_num = 0, fix_denom = 0;
|
|
|
|
/* This code is for Linux < 2.6.19 */
|
|
min_w = min_h = 1;
|
|
max_w = max_h = GST_V4L2_MAX_SIZE;
|
|
if (!gst_v4l2src_get_nearest_size (v4l2src, pixelformat, &min_w, &min_h)) {
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"Could not probe minimum capture size for pixelformat %"
|
|
GST_FOURCC_FORMAT, GST_FOURCC_ARGS (pixelformat));
|
|
}
|
|
if (!gst_v4l2src_get_nearest_size (v4l2src, pixelformat, &max_w, &max_h)) {
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"Could not probe maximum capture size for pixelformat %"
|
|
GST_FOURCC_FORMAT, GST_FOURCC_ARGS (pixelformat));
|
|
}
|
|
|
|
/* Since we can't get framerate directly, try to use the current norm */
|
|
if (v4l2src->v4l2object->norm && v4l2src->v4l2object->norms) {
|
|
GList *norms;
|
|
GstTunerNorm *norm;
|
|
|
|
for (norms = v4l2src->v4l2object->norms; norms != NULL;
|
|
norms = norms->next) {
|
|
norm = (GstTunerNorm *) norms->data;
|
|
if (!strcmp (norm->label, v4l2src->v4l2object->norm))
|
|
break;
|
|
}
|
|
/* If it's possible, set framerate to that (discrete) value */
|
|
if (norm) {
|
|
fix_num = gst_value_get_fraction_numerator (&norm->framerate);
|
|
fix_denom = gst_value_get_fraction_denominator (&norm->framerate);
|
|
}
|
|
}
|
|
|
|
tmp = gst_structure_copy (template);
|
|
if (fix_num) {
|
|
gst_structure_set (tmp,
|
|
"width", GST_TYPE_INT_RANGE, min_w, max_w,
|
|
"height", GST_TYPE_INT_RANGE, min_h, max_h,
|
|
"framerate", GST_TYPE_FRACTION, fix_num, fix_denom, NULL);
|
|
} else {
|
|
/* if norm can't be used, copy the template framerate */
|
|
gst_structure_set (tmp,
|
|
"width", GST_TYPE_INT_RANGE, min_w, max_w,
|
|
"height", GST_TYPE_INT_RANGE, min_h, max_h,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 100, 1, NULL);
|
|
}
|
|
gst_caps_append_structure (ret, tmp);
|
|
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_grab_frame ():
|
|
* grab a frame for capturing
|
|
* return value: GST_FLOW_OK or GST_FLOW_ERROR
|
|
******************************************************/
|
|
GstFlowReturn
|
|
gst_v4l2src_grab_frame (GstV4l2Src * v4l2src, GstBuffer ** buf)
|
|
{
|
|
#define NUM_TRIALS 50
|
|
struct v4l2_buffer buffer;
|
|
gint32 trials = NUM_TRIALS;
|
|
GstBuffer *pool_buffer;
|
|
gboolean need_copy;
|
|
gint index;
|
|
|
|
memset (&buffer, 0x00, sizeof (buffer));
|
|
buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
buffer.memory = V4L2_MEMORY_MMAP;
|
|
|
|
while (v4l2_ioctl (v4l2src->v4l2object->video_fd, VIDIOC_DQBUF, &buffer) < 0) {
|
|
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"problem grabbing frame %d (ix=%d), trials=%d, pool-ct=%d, buf.flags=%d",
|
|
buffer.sequence, buffer.index, trials,
|
|
GST_MINI_OBJECT_REFCOUNT (v4l2src->pool), buffer.flags);
|
|
|
|
/* if the sync() got interrupted, we can retry */
|
|
switch (errno) {
|
|
case EAGAIN:
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"Non-blocking I/O has been selected using O_NONBLOCK and"
|
|
" no buffer was in the outgoing queue. device %s",
|
|
v4l2src->v4l2object->videodev);
|
|
break;
|
|
case EINVAL:
|
|
goto einval;
|
|
case ENOMEM:
|
|
goto enomem;
|
|
case EIO:
|
|
GST_INFO_OBJECT (v4l2src,
|
|
"VIDIOC_DQBUF failed due to an internal error."
|
|
" Can also indicate temporary problems like signal loss."
|
|
" Note the driver might dequeue an (empty) buffer despite"
|
|
" returning an error, or even stop capturing."
|
|
" device %s", v4l2src->v4l2object->videodev);
|
|
/* have we de-queued a buffer ? */
|
|
if (!(buffer.flags & (V4L2_BUF_FLAG_QUEUED | V4L2_BUF_FLAG_DONE))) {
|
|
/* this fails
|
|
if ((buffer.index >= 0) && (buffer.index < v4l2src->breq.count)) {
|
|
GST_DEBUG_OBJECT (v4l2src, "reenqueing buffer (ix=%ld)", buffer.index);
|
|
gst_v4l2src_queue_frame (v4l2src, buffer.index);
|
|
}
|
|
else {
|
|
*/
|
|
GST_DEBUG_OBJECT (v4l2src, "reenqueing buffer");
|
|
/* FIXME: this is not a good idea, as drivers usualy return the buffer
|
|
* with index-number set to 0, thus the re-enque will fail unless it
|
|
* was incidentialy 0.
|
|
* We could try to re-enque all buffers without handling the ioctl
|
|
* return.
|
|
*/
|
|
/*
|
|
if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_QBUF, &buffer) < 0) {
|
|
goto qbuf_failed;
|
|
}
|
|
*/
|
|
/*} */
|
|
}
|
|
break;
|
|
case EINTR:
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"could not sync on a buffer on device %s",
|
|
v4l2src->v4l2object->videodev);
|
|
break;
|
|
default:
|
|
GST_WARNING_OBJECT (v4l2src,
|
|
"Grabbing frame got interrupted on %s. No expected reason.",
|
|
v4l2src->v4l2object->videodev);
|
|
break;
|
|
}
|
|
|
|
/* check nr. of attempts to capture */
|
|
if (--trials == -1) {
|
|
goto too_many_trials;
|
|
} else {
|
|
memset (&buffer, 0x00, sizeof (buffer));
|
|
buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
buffer.memory = V4L2_MEMORY_MMAP;
|
|
}
|
|
}
|
|
|
|
g_mutex_lock (v4l2src->pool->lock);
|
|
|
|
index = buffer.index;
|
|
|
|
/* get our GstBuffer with that index from the pool, if the buffer was
|
|
* outstanding we have a serious problem. */
|
|
pool_buffer = GST_BUFFER (v4l2src->pool->buffers[index]);
|
|
|
|
if (pool_buffer == NULL)
|
|
goto no_buffer;
|
|
|
|
GST_LOG_OBJECT (v4l2src, "grabbed buffer %p at index %d", pool_buffer, index);
|
|
|
|
/* we have the buffer now, mark the spot in the pool empty */
|
|
v4l2src->pool->buffers[index] = NULL;
|
|
v4l2src->pool->num_live_buffers++;
|
|
/* if we are handing out the last buffer in the pool, we need to make a
|
|
* copy and bring the buffer back in the pool. */
|
|
need_copy = v4l2src->always_copy
|
|
|| (v4l2src->pool->num_live_buffers == v4l2src->pool->buffer_count);
|
|
|
|
g_mutex_unlock (v4l2src->pool->lock);
|
|
|
|
/* this can change at every frame, esp. with jpeg */
|
|
GST_BUFFER_SIZE (pool_buffer) = buffer.bytesused;
|
|
|
|
GST_BUFFER_OFFSET (pool_buffer) = v4l2src->offset++;
|
|
GST_BUFFER_OFFSET_END (pool_buffer) = v4l2src->offset;
|
|
|
|
/* timestamps, LOCK to get clock and base time. */
|
|
{
|
|
GstClock *clock;
|
|
GstClockTime timestamp;
|
|
|
|
GST_OBJECT_LOCK (v4l2src);
|
|
if ((clock = GST_ELEMENT_CLOCK (v4l2src))) {
|
|
/* we have a clock, get base time and ref clock */
|
|
timestamp = GST_ELEMENT (v4l2src)->base_time;
|
|
gst_object_ref (clock);
|
|
} else {
|
|
/* no clock, can't set timestamps */
|
|
timestamp = GST_CLOCK_TIME_NONE;
|
|
}
|
|
GST_OBJECT_UNLOCK (v4l2src);
|
|
|
|
if (clock) {
|
|
/* the time now is the time of the clock minus the base time */
|
|
timestamp = gst_clock_get_time (clock) - timestamp;
|
|
gst_object_unref (clock);
|
|
}
|
|
|
|
/* FIXME: use the timestamp from the buffer itself! */
|
|
GST_BUFFER_TIMESTAMP (pool_buffer) = timestamp;
|
|
}
|
|
|
|
if (G_UNLIKELY (need_copy)) {
|
|
*buf = gst_buffer_copy (pool_buffer);
|
|
GST_BUFFER_FLAG_UNSET (*buf, GST_BUFFER_FLAG_READONLY);
|
|
/* this will requeue */
|
|
gst_buffer_unref (pool_buffer);
|
|
} else {
|
|
*buf = pool_buffer;
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src, "grabbed frame %d (ix=%d), flags %08x, pool-ct=%d",
|
|
buffer.sequence, buffer.index, buffer.flags,
|
|
v4l2src->pool->num_live_buffers);
|
|
|
|
return GST_FLOW_OK;
|
|
|
|
/* ERRORS */
|
|
einval:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED,
|
|
(_("Failed trying to get video frames from device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
(_("The buffer type is not supported, or the index is out of bounds,"
|
|
" or no buffers have been allocated yet, or the userptr"
|
|
" or length are invalid. device %s"),
|
|
v4l2src->v4l2object->videodev));
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
enomem:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED,
|
|
(_("Failed trying to get video frames from device '%s'. Not enough memory."), v4l2src->v4l2object->videodev), (_("insufficient memory to enqueue a user pointer buffer. device %s."), v4l2src->v4l2object->videodev));
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
too_many_trials:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED,
|
|
(_("Failed trying to get video frames from device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
(_("Failed after %d tries. device %s. system error: %s"),
|
|
NUM_TRIALS, v4l2src->v4l2object->videodev, g_strerror (errno)));
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
no_buffer:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED,
|
|
(_("Failed trying to get video frames from device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
(_("No free buffers found in the pool at index %d."), index));
|
|
g_mutex_unlock (v4l2src->pool->lock);
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
/*
|
|
qbuf_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, WRITE,
|
|
(_("Could not exchange data with device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
("Error queueing buffer on device %s. system error: %s",
|
|
v4l2src->v4l2object->videodev, g_strerror (errno)));
|
|
return GST_FLOW_ERROR;
|
|
}
|
|
*/
|
|
}
|
|
|
|
/* Note about fraction simplification
|
|
* n1/d1 == n2/d2 is also written as n1 == ( n2 * d1 ) / d2
|
|
*/
|
|
#define fractions_are_equal(n1,d1,n2,d2) ((n1) == gst_util_uint64_scale_int((n2), (d1), (d2)))
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_set_capture():
|
|
* set capture parameters
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_set_capture (GstV4l2Src * v4l2src, guint32 pixelformat,
|
|
guint32 width, guint32 height, guint fps_n, guint fps_d)
|
|
{
|
|
gint fd = v4l2src->v4l2object->video_fd;
|
|
struct v4l2_format format;
|
|
struct v4l2_streamparm stream;
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "Setting capture format to %dx%d, format "
|
|
"%" GST_FOURCC_FORMAT, width, height, GST_FOURCC_ARGS (pixelformat));
|
|
|
|
GST_V4L2_CHECK_OPEN (v4l2src->v4l2object);
|
|
GST_V4L2_CHECK_NOT_ACTIVE (v4l2src->v4l2object);
|
|
|
|
memset (&format, 0x00, sizeof (struct v4l2_format));
|
|
format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
|
|
if (v4l2_ioctl (fd, VIDIOC_G_FMT, &format) < 0)
|
|
goto get_fmt_failed;
|
|
|
|
format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
format.fmt.pix.width = width;
|
|
format.fmt.pix.height = height;
|
|
format.fmt.pix.pixelformat = pixelformat;
|
|
/* request whole frames; change when gstreamer supports interlaced video
|
|
* (INTERLACED mode returns frames where the fields have already been
|
|
* combined, there are other modes for requesting fields individually) */
|
|
format.fmt.pix.field = V4L2_FIELD_INTERLACED;
|
|
|
|
if (v4l2_ioctl (fd, VIDIOC_S_FMT, &format) < 0) {
|
|
if (errno != EINVAL)
|
|
goto set_fmt_failed;
|
|
|
|
/* try again with progressive video */
|
|
format.fmt.pix.width = width;
|
|
format.fmt.pix.height = height;
|
|
format.fmt.pix.pixelformat = pixelformat;
|
|
format.fmt.pix.field = V4L2_FIELD_NONE;
|
|
if (v4l2_ioctl (fd, VIDIOC_S_FMT, &format) < 0)
|
|
goto set_fmt_failed;
|
|
}
|
|
|
|
if (format.fmt.pix.width != width || format.fmt.pix.height != height)
|
|
goto invalid_dimensions;
|
|
|
|
if (format.fmt.pix.pixelformat != pixelformat)
|
|
goto invalid_pixelformat;
|
|
|
|
/* Is there a reason we require the caller to always specify a framerate? */
|
|
GST_LOG_OBJECT (v4l2src, "Desired framerate: %u/%u", fps_n, fps_d);
|
|
|
|
memset (&stream, 0x00, sizeof (struct v4l2_streamparm));
|
|
stream.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
if (v4l2_ioctl (fd, VIDIOC_G_PARM, &stream) < 0) {
|
|
GST_ELEMENT_WARNING (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Could not get parameters on device '%s'"),
|
|
v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM);
|
|
goto done;
|
|
}
|
|
|
|
/* Note: V4L2 provides the frame interval, we have the frame rate */
|
|
if (fractions_are_equal (stream.parm.capture.timeperframe.numerator,
|
|
stream.parm.capture.timeperframe.denominator, fps_d, fps_n)) {
|
|
GST_LOG_OBJECT (v4l2src, "Desired framerate already set");
|
|
v4l2src->fps_n = fps_n;
|
|
v4l2src->fps_d = fps_d;
|
|
goto done;
|
|
}
|
|
|
|
/* We want to change the frame rate, so check whether we can. Some cheap USB
|
|
* cameras don't have the capability */
|
|
if ((stream.parm.capture.capability & V4L2_CAP_TIMEPERFRAME) == 0) {
|
|
GST_DEBUG_OBJECT (v4l2src, "Not setting framerate (not supported)");
|
|
goto done;
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src, "Setting framerate to %u/%u", fps_n, fps_d);
|
|
|
|
/* Note: V4L2 wants the frame interval, we have the frame rate */
|
|
stream.parm.capture.timeperframe.numerator = fps_d;
|
|
stream.parm.capture.timeperframe.denominator = fps_n;
|
|
|
|
/* some cheap USB cam's won't accept any change */
|
|
if (v4l2_ioctl (fd, VIDIOC_S_PARM, &stream) < 0) {
|
|
GST_ELEMENT_WARNING (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Video input device did not accept new frame rate setting.")),
|
|
GST_ERROR_SYSTEM);
|
|
goto done;
|
|
}
|
|
|
|
v4l2src->fps_n = fps_n;
|
|
v4l2src->fps_d = fps_d;
|
|
GST_INFO_OBJECT (v4l2src, "Set framerate to %u/%u", fps_n, fps_d);
|
|
|
|
done:
|
|
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
get_fmt_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Device '%s' does not support video capture"),
|
|
v4l2src->v4l2object->videodev),
|
|
("Call to G_FMT failed: (%s)", g_strerror (errno)));
|
|
return FALSE;
|
|
}
|
|
set_fmt_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Device '%s' cannot capture at %dx%d"),
|
|
v4l2src->v4l2object->videodev, width, height),
|
|
("Call to S_FMT failed for %" GST_FOURCC_FORMAT " @ %dx%d: %s",
|
|
GST_FOURCC_ARGS (pixelformat), width, height, g_strerror (errno)));
|
|
return FALSE;
|
|
}
|
|
invalid_dimensions:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Device '%s' cannot capture at %dx%d"),
|
|
v4l2src->v4l2object->videodev, width, height),
|
|
("Tried to capture at %dx%d, but device returned size %dx%d",
|
|
width, height, format.fmt.pix.width, format.fmt.pix.height));
|
|
return FALSE;
|
|
}
|
|
invalid_pixelformat:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS,
|
|
(_("Device '%s' cannot capture in the specified format"),
|
|
v4l2src->v4l2object->videodev),
|
|
("Tried to capture in %" GST_FOURCC_FORMAT
|
|
", but device returned format" " %" GST_FOURCC_FORMAT,
|
|
GST_FOURCC_ARGS (pixelformat),
|
|
GST_FOURCC_ARGS (format.fmt.pix.pixelformat)));
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_capture_init():
|
|
* initialize the capture system
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_capture_init (GstV4l2Src * v4l2src, GstCaps * caps)
|
|
{
|
|
gint fd = v4l2src->v4l2object->video_fd;
|
|
struct v4l2_requestbuffers breq;
|
|
|
|
memset (&breq, 0, sizeof (struct v4l2_requestbuffers));
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "initializing the capture system");
|
|
|
|
GST_V4L2_CHECK_OPEN (v4l2src->v4l2object);
|
|
GST_V4L2_CHECK_NOT_ACTIVE (v4l2src->v4l2object);
|
|
|
|
if (v4l2src->v4l2object->vcap.capabilities & V4L2_CAP_STREAMING) {
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "STREAMING, requesting %d MMAP CAPTURE buffers",
|
|
v4l2src->num_buffers);
|
|
|
|
breq.count = v4l2src->num_buffers;
|
|
breq.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
breq.memory = V4L2_MEMORY_MMAP;
|
|
|
|
if (v4l2_ioctl (fd, VIDIOC_REQBUFS, &breq) < 0)
|
|
goto reqbufs_failed;
|
|
|
|
GST_LOG_OBJECT (v4l2src, " count: %u", breq.count);
|
|
GST_LOG_OBJECT (v4l2src, " type: %d", breq.type);
|
|
GST_LOG_OBJECT (v4l2src, " memory: %d", breq.memory);
|
|
|
|
if (breq.count < GST_V4L2_MIN_BUFFERS)
|
|
goto no_buffers;
|
|
|
|
if (v4l2src->num_buffers != breq.count) {
|
|
GST_WARNING_OBJECT (v4l2src, "using %u buffers instead", breq.count);
|
|
v4l2src->num_buffers = breq.count;
|
|
g_object_notify (G_OBJECT (v4l2src), "queue-size");
|
|
}
|
|
|
|
/* Map the buffers */
|
|
GST_LOG_OBJECT (v4l2src, "initiating buffer pool");
|
|
|
|
if (!(v4l2src->pool = gst_v4l2_buffer_pool_new (v4l2src, fd,
|
|
v4l2src->num_buffers, caps)))
|
|
goto buffer_pool_new_failed;
|
|
|
|
GST_INFO_OBJECT (v4l2src, "capturing buffers via mmap()");
|
|
v4l2src->use_mmap = TRUE;
|
|
} else if (v4l2src->v4l2object->vcap.capabilities & V4L2_CAP_READWRITE) {
|
|
GST_INFO_OBJECT (v4l2src, "capturing buffers via read()");
|
|
v4l2src->use_mmap = FALSE;
|
|
v4l2src->pool = NULL;
|
|
} else {
|
|
goto no_supported_capture_method;
|
|
}
|
|
|
|
GST_V4L2_SET_ACTIVE (v4l2src->v4l2object);
|
|
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
reqbufs_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ,
|
|
(_("Could not get buffers from device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
("error requesting %d buffers: %s",
|
|
v4l2src->num_buffers, g_strerror (errno)));
|
|
return FALSE;
|
|
}
|
|
no_buffers:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ,
|
|
(_("Could not get enough buffers from device '%s'."),
|
|
v4l2src->v4l2object->videodev),
|
|
("we received %d from device '%s', we want at least %d",
|
|
breq.count, v4l2src->v4l2object->videodev, GST_V4L2_MIN_BUFFERS));
|
|
return FALSE;
|
|
}
|
|
buffer_pool_new_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ,
|
|
(_("Could not map buffers from device '%s'"),
|
|
v4l2src->v4l2object->videodev),
|
|
("Failed to create buffer pool: %s", g_strerror (errno)));
|
|
return FALSE;
|
|
}
|
|
no_supported_capture_method:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ,
|
|
(_("The driver of device '%s' does not support any known capture "
|
|
"method."), v4l2src->v4l2object->videodev), (NULL));
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_capture_start():
|
|
* start streaming capture
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_capture_start (GstV4l2Src * v4l2src)
|
|
{
|
|
gint type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
gint fd = v4l2src->v4l2object->video_fd;
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "starting the capturing");
|
|
//GST_V4L2_CHECK_OPEN (v4l2src->v4l2object);
|
|
GST_V4L2_CHECK_ACTIVE (v4l2src->v4l2object);
|
|
|
|
v4l2src->quit = FALSE;
|
|
|
|
if (v4l2src->use_mmap) {
|
|
if (!gst_v4l2_buffer_pool_activate (v4l2src->pool, v4l2src))
|
|
goto pool_activate_failed;
|
|
|
|
if (v4l2_ioctl (fd, VIDIOC_STREAMON, &type) < 0)
|
|
goto streamon_failed;
|
|
}
|
|
|
|
v4l2src->is_capturing = TRUE;
|
|
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
pool_activate_failed:
|
|
{
|
|
/* already errored */
|
|
return FALSE;
|
|
}
|
|
streamon_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, OPEN_READ,
|
|
(_("Error starting streaming capture from device '%s'."),
|
|
v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_capture_stop():
|
|
* stop streaming capture
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_capture_stop (GstV4l2Src * v4l2src)
|
|
{
|
|
gint type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
|
|
GST_DEBUG_OBJECT (v4l2src, "stopping capturing");
|
|
|
|
if (!GST_V4L2_IS_OPEN (v4l2src->v4l2object)) {
|
|
goto done;
|
|
}
|
|
if (!GST_V4L2_IS_ACTIVE (v4l2src->v4l2object)) {
|
|
goto done;
|
|
}
|
|
|
|
if (v4l2src->use_mmap) {
|
|
/* we actually need to sync on all queued buffers but not
|
|
* on the non-queued ones */
|
|
if (v4l2_ioctl (v4l2src->v4l2object->video_fd, VIDIOC_STREAMOFF, &type) < 0)
|
|
goto streamoff_failed;
|
|
}
|
|
|
|
done:
|
|
|
|
/* make an optional pending wait stop */
|
|
v4l2src->quit = TRUE;
|
|
v4l2src->is_capturing = FALSE;
|
|
|
|
return TRUE;
|
|
|
|
/* ERRORS */
|
|
streamoff_failed:
|
|
{
|
|
GST_ELEMENT_ERROR (v4l2src, RESOURCE, CLOSE,
|
|
(_("Error stopping streaming capture from device '%s'."),
|
|
v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
/******************************************************
|
|
* gst_v4l2src_capture_deinit():
|
|
* deinitialize the capture system
|
|
* return value: TRUE on success, FALSE on error
|
|
******************************************************/
|
|
gboolean
|
|
gst_v4l2src_capture_deinit (GstV4l2Src * v4l2src)
|
|
{
|
|
GST_DEBUG_OBJECT (v4l2src, "deinitting capture system");
|
|
|
|
if (!GST_V4L2_IS_OPEN (v4l2src->v4l2object)) {
|
|
return TRUE;
|
|
}
|
|
if (!GST_V4L2_IS_ACTIVE (v4l2src->v4l2object)) {
|
|
return TRUE;
|
|
}
|
|
|
|
if (v4l2src->pool) {
|
|
gst_v4l2_buffer_pool_destroy (v4l2src->pool);
|
|
v4l2src->pool = NULL;
|
|
}
|
|
|
|
GST_V4L2_SET_INACTIVE (v4l2src->v4l2object);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*
|
|
*/
|
|
static gboolean
|
|
gst_v4l2src_get_nearest_size (GstV4l2Src * v4l2src, guint32 pixelformat,
|
|
gint * width, gint * height)
|
|
{
|
|
struct v4l2_format fmt;
|
|
int fd;
|
|
int r;
|
|
|
|
g_return_val_if_fail (width != NULL, FALSE);
|
|
g_return_val_if_fail (height != NULL, FALSE);
|
|
|
|
GST_LOG_OBJECT (v4l2src,
|
|
"getting nearest size to %dx%d with format %" GST_FOURCC_FORMAT,
|
|
*width, *height, GST_FOURCC_ARGS (pixelformat));
|
|
|
|
fd = v4l2src->v4l2object->video_fd;
|
|
|
|
/* get size delimiters */
|
|
memset (&fmt, 0, sizeof (fmt));
|
|
fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
|
|
fmt.fmt.pix.width = *width;
|
|
fmt.fmt.pix.height = *height;
|
|
fmt.fmt.pix.pixelformat = pixelformat;
|
|
fmt.fmt.pix.field = V4L2_FIELD_INTERLACED;
|
|
|
|
r = v4l2_ioctl (fd, VIDIOC_TRY_FMT, &fmt);
|
|
if (r < 0 && errno == EINVAL) {
|
|
/* try again with progressive video */
|
|
fmt.fmt.pix.width = *width;
|
|
fmt.fmt.pix.height = *height;
|
|
fmt.fmt.pix.pixelformat = pixelformat;
|
|
fmt.fmt.pix.field = V4L2_FIELD_NONE;
|
|
r = v4l2_ioctl (fd, VIDIOC_TRY_FMT, &fmt);
|
|
}
|
|
|
|
if (r < 0) {
|
|
/* The driver might not implement TRY_FMT, in which case we will try
|
|
S_FMT to probe */
|
|
if (errno != ENOTTY)
|
|
return FALSE;
|
|
|
|
/* Only try S_FMT if we're not actively capturing yet, which we shouldn't
|
|
be, because we're still probing */
|
|
if (GST_V4L2_IS_ACTIVE (v4l2src->v4l2object))
|
|
return FALSE;
|
|
|
|
GST_LOG_OBJECT (v4l2src,
|
|
"Failed to probe size limit with VIDIOC_TRY_FMT, trying VIDIOC_S_FMT");
|
|
|
|
fmt.fmt.pix.width = *width;
|
|
fmt.fmt.pix.height = *height;
|
|
|
|
r = v4l2_ioctl (fd, VIDIOC_S_FMT, &fmt);
|
|
if (r < 0 && errno == EINVAL) {
|
|
/* try again with progressive video */
|
|
fmt.fmt.pix.width = *width;
|
|
fmt.fmt.pix.height = *height;
|
|
fmt.fmt.pix.pixelformat = pixelformat;
|
|
fmt.fmt.pix.field = V4L2_FIELD_NONE;
|
|
r = v4l2_ioctl (fd, VIDIOC_S_FMT, &fmt);
|
|
}
|
|
|
|
if (r < 0)
|
|
return FALSE;
|
|
}
|
|
|
|
GST_LOG_OBJECT (v4l2src,
|
|
"got nearest size %dx%d", fmt.fmt.pix.width, fmt.fmt.pix.height);
|
|
|
|
*width = fmt.fmt.pix.width;
|
|
*height = fmt.fmt.pix.height;
|
|
|
|
return TRUE;
|
|
}
|