/* GStreamer * * Copyright (C) 2002 Ronald Bultje * 2006 Edgard Lima * * v4l2src.c - system calls * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include #include #include #include #include "v4l2src_calls.h" #include #include #include "gstv4l2tuner.h" GST_DEBUG_CATEGORY_EXTERN (v4l2src_debug); #define GST_CAT_DEFAULT v4l2src_debug /* lalala... */ #define GST_V4L2_SET_ACTIVE(element) (element)->buffer = GINT_TO_POINTER (-1) #define GST_V4L2_SET_INACTIVE(element) (element)->buffer = NULL #define DEBUG(format, args...) \ GST_CAT_DEBUG_OBJECT (\ v4l2src_debug, v4l2src, \ "V4L2SRC: " format, ##args) /* On some systems MAP_FAILED seems to be missing */ #ifndef MAP_FAILED #define MAP_FAILED ( (caddr_t) -1 ) #endif /****************************************************** * gst_v4l2src_fill_format_list(): * create list of supported capture formats * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_fill_format_list (GstV4l2Src * v4l2src) { gint n; struct v4l2_fmtdesc *format; GST_DEBUG_OBJECT (v4l2src, "getting src format enumerations"); /* format enumeration */ for (n = 0;; n++) { format = g_new (struct v4l2_fmtdesc, 1); format->index = n; format->type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_ENUM_FMT, format) < 0) { if (errno == EINVAL) { break; /* end of enumeration */ } else { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("failed to get number %d in pixelformat enumeration for %s: %s"), n, v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); g_free (format); return FALSE; } } GST_LOG_OBJECT (v4l2src, "got format %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (format->pixelformat)); v4l2src->formats = g_slist_prepend (v4l2src->formats, format); } return TRUE; } /****************************************************** * gst_v4l2src_clear_format_list(): * free list of supported capture formats * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_clear_format_list (GstV4l2Src * v4l2src) { g_slist_foreach (v4l2src->formats, (GFunc) g_free, NULL); g_slist_free (v4l2src->formats); v4l2src->formats = NULL; return TRUE; } /****************************************************** * gst_v4l2src_queue_frame(): * queue a frame for capturing * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_queue_frame (GstV4l2Src * v4l2src, guint i) { GST_LOG_OBJECT (v4l2src, "queueing frame %u", i); if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_QBUF, &v4l2src->pool->buffers[i].buffer) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, WRITE, (_("Could not write to device '%s'."), v4l2src->v4l2object->videodev), ("Error queueing buffer %u on device %s. system error: %s", i, v4l2src->v4l2object->videodev, g_strerror (errno))); return FALSE; } return TRUE; } /****************************************************** * gst_v4l2src_grab_frame (): * grab a frame for capturing * return value: TRUE on success, FALSE on error ******************************************************/ gint gst_v4l2src_grab_frame (GstV4l2Src * v4l2src) { struct v4l2_buffer buffer; gint32 trials = 100; memset (&buffer, 0x00, sizeof (buffer)); buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; buffer.memory = v4l2src->breq.memory; while (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_DQBUF, &buffer) < 0) { /* if the sync() got interrupted, we can retry */ switch (errno) { case EAGAIN: GST_DEBUG_OBJECT (v4l2src, "Non-blocking I/O has been selected using O_NONBLOCK and" " no buffer was in the outgoing queue. device %s", v4l2src->v4l2object->videodev); break; case EINVAL: GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED, (_("Failed trying to get frames from device %s"), v4l2src->v4l2object->videodev), (_("The buffer type is not supported, or the index is out of bounds," " or no buffers have been allocated yet, or the userptr" " or length are invalid. device %s"), v4l2src->v4l2object->videodev)); break; case ENOMEM: GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED, (_("Failed trying to get frames from device %s. Not enough memory"), v4l2src->v4l2object->videodev), (_("isufficient memory to enqueue a user pointer buffer. device %s"), v4l2src->v4l2object->videodev)); break; case EIO: GST_DEBUG_OBJECT (v4l2src, "VIDIOC_DQBUF failed due to an internal error." " Can also indicate temporary problems like signal loss." " Note the driver might dequeue an (empty) buffer despite" " returning an error, or even stop capturing." " device %s", v4l2src->v4l2object->videodev); break; case EINTR: GST_DEBUG_OBJECT (v4l2src, "could not sync on a buffer on device %s", v4l2src->v4l2object->videodev); break; default: GST_ELEMENT_WARNING (v4l2src, RESOURCE, FAILED, (_("Grabbing frame got interrupted on %s. No expected reason"), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); break; } if (--trials == -1) { return -1; } else { ioctl (v4l2src->v4l2object->video_fd, VIDIOC_QBUF, &buffer); memset (&buffer, 0x00, sizeof (buffer)); buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; buffer.memory = v4l2src->breq.memory; } } GST_LOG_OBJECT (v4l2src, "grabbed frame %d", buffer.index); return buffer.index; } /****************************************************** * gst_v4l2src_get_capture(): * get capture parameters * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_get_capture (GstV4l2Src * v4l2src) { DEBUG ("Getting capture format"); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); memset (&v4l2src->format, 0, sizeof (struct v4l2_format)); v4l2src->format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_G_FMT, &v4l2src->format) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("failed to get pixelformat for device %s"), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); return FALSE; } return TRUE; } /****************************************************** * gst_v4l2src_set_capture(): * set capture parameters * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_set_capture (GstV4l2Src * v4l2src, struct v4l2_fmtdesc * fmt, gint * width, gint * height, guint * fps_n, guint * fps_d) { guint new_fps_n = *fps_n; guint new_fps_d = *fps_d; DEBUG ("Setting capture format to %dx%d, format %s", *width, *height, fmt->description); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_NOT_ACTIVE (v4l2src->v4l2object); if (!gst_v4l2src_get_capture (v4l2src)) { goto fail; } v4l2src->format.fmt.pix.width = *width; v4l2src->format.fmt.pix.height = *height; v4l2src->format.fmt.pix.pixelformat = fmt->pixelformat; v4l2src->format.fmt.pix.field = V4L2_FIELD_INTERLACED; v4l2src->format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_S_FMT, &v4l2src->format) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("failed to set pixelformat to %s @ %dx%d for device %s: %s"), fmt->description, *width, *height, v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); goto fail; } if (*width != v4l2src->format.fmt.pix.width || *height != v4l2src->format.fmt.pix.height) { DEBUG ("Updating size from %dx%d to %dx%d, format %s", *width, *height, v4l2src->format.fmt.pix.width, v4l2src->format.fmt.pix.height, fmt->description); } /* update internal info */ if (!gst_v4l2src_get_capture (v4l2src)) { goto fail; } if (fmt->pixelformat != v4l2src->format.fmt.pix.pixelformat) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("failed to set pixelformat to %s @ %dx%d for device %s: %s"), fmt->description, *width, *height, v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); goto fail; } if (*fps_n) { if (gst_v4l2src_set_fps (v4l2src, &new_fps_n, &new_fps_d)) { if (new_fps_n != *fps_n || new_fps_d != *fps_d) { DEBUG ("Updating framerate from %u/%u to %u%u", *fps_n, *fps_d, new_fps_n, new_fps_d); *fps_n = new_fps_n; *fps_d = new_fps_d; } } } else { if (gst_v4l2src_get_fps (v4l2src, &new_fps_n, &new_fps_d)) { DEBUG ("framerate is %u/%u", new_fps_n, new_fps_d); *fps_n = new_fps_n; *fps_d = new_fps_d; } } *width = v4l2src->format.fmt.pix.width; *height = v4l2src->format.fmt.pix.height; return TRUE; fail: return FALSE; } /****************************************************** * gst_v4l2src_capture_init(): * initialize the capture system * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_init (GstV4l2Src * v4l2src) { gint n; guint buffers; GST_DEBUG_OBJECT (v4l2src, "initting the capture system"); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_NOT_ACTIVE (v4l2src->v4l2object); /* request buffer info */ buffers = v4l2src->breq.count; if (v4l2src->breq.count > GST_V4L2_MAX_BUFFERS) { v4l2src->breq.count = GST_V4L2_MAX_BUFFERS; } if (v4l2src->breq.count < GST_V4L2_MIN_BUFFERS) { v4l2src->breq.count = GST_V4L2_MIN_BUFFERS; } v4l2src->breq.type = v4l2src->format.type; if (v4l2src->v4l2object->vcap.capabilities & V4L2_CAP_STREAMING) { v4l2src->breq.memory = V4L2_MEMORY_MMAP; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_REQBUFS, &v4l2src->breq) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not get buffers from device '%s'."), v4l2src->v4l2object->videodev), ("error requesting %d buffers. system error: %s", v4l2src->breq.count, g_strerror (errno))); return FALSE; } GST_LOG_OBJECT (v4l2src, "using default mmap method"); } else if (v4l2src->v4l2object->vcap.capabilities & V4L2_CAP_READWRITE) { v4l2src->breq.memory = 0; GST_INFO_OBJECT (v4l2src, "using fallback read method"); } else { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("the driver of device '%s' is broken."), v4l2src->v4l2object->videodev), ("no supported read capability from %s", v4l2src->v4l2object->videodev)); return FALSE; } /* Determine the device's framerate */ if (!gst_v4l2src_update_fps (v4l2src->v4l2object)) { GST_DEBUG_OBJECT (v4l2src, "frame rate is unknown."); v4l2src->fps_d = 1; v4l2src->fps_n = 0; } if (v4l2src->breq.memory > 0) { if (v4l2src->breq.count < GST_V4L2_MIN_BUFFERS) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not get enough buffers from device '%s'."), v4l2src->v4l2object->videodev), ("we received %d, we want at least %d", v4l2src->breq.count, GST_V4L2_MIN_BUFFERS)); v4l2src->breq.count = buffers; return FALSE; } if (v4l2src->breq.count != buffers) g_object_notify (G_OBJECT (v4l2src), "num_buffers"); GST_INFO_OBJECT (v4l2src, "Got %d buffers (%" GST_FOURCC_FORMAT ") of size %d KB", v4l2src->breq.count, GST_FOURCC_ARGS (v4l2src->format.fmt.pix.pixelformat), v4l2src->format.fmt.pix.sizeimage / 1024); /* Map the buffers */ GST_LOG_OBJECT (v4l2src, "initiating buffer pool"); v4l2src->pool = g_new (GstV4l2BufferPool, 1); gst_atomic_int_set (&v4l2src->pool->refcount, 1); v4l2src->pool->video_fd = v4l2src->v4l2object->video_fd; v4l2src->pool->buffer_count = v4l2src->breq.count; v4l2src->pool->buffers = g_new0 (GstV4l2Buffer, v4l2src->breq.count); for (n = 0; n < v4l2src->breq.count; n++) { GstV4l2Buffer *buffer = &v4l2src->pool->buffers[n]; gst_atomic_int_set (&buffer->refcount, 1); buffer->pool = v4l2src->pool; memset (&buffer->buffer, 0x00, sizeof (buffer->buffer)); buffer->buffer.index = n; buffer->buffer.type = v4l2src->breq.type; buffer->buffer.memory = v4l2src->breq.memory; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_QUERYBUF, &buffer->buffer) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not get buffer properties of buffer %d"), n), GST_ERROR_SYSTEM); gst_v4l2src_capture_deinit (v4l2src); return FALSE; } buffer->start = mmap (0, buffer->buffer.length, PROT_READ | PROT_WRITE, MAP_SHARED, v4l2src->v4l2object->video_fd, buffer->buffer.m.offset); if (buffer->start == MAP_FAILED) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not mmap video buffer %d: %s"), n), GST_ERROR_SYSTEM); buffer->start = 0; gst_v4l2src_capture_deinit (v4l2src); return FALSE; } buffer->length = buffer->buffer.length; if (!gst_v4l2src_queue_frame (v4l2src, n)) { gst_v4l2src_capture_deinit (v4l2src); return FALSE; } } } else { GST_LOG_OBJECT (v4l2src, "no buffer pool used"); v4l2src->pool = NULL; } GST_V4L2_SET_ACTIVE (v4l2src->v4l2object); return TRUE; } /****************************************************** * gst_v4l2src_capture_start(): * start streaming capture * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_start (GstV4l2Src * v4l2src) { gint type = V4L2_BUF_TYPE_VIDEO_CAPTURE; GST_DEBUG_OBJECT (v4l2src, "starting the capturing"); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); if (!GST_V4L2_IS_ACTIVE (v4l2src->v4l2object)) { /* gst_pad_renegotiate (v4l2src->srcpad); FIX: is it still required in 0.10 */ } GST_V4L2_CHECK_ACTIVE (v4l2src->v4l2object); v4l2src->quit = FALSE; if (v4l2src->breq.memory != 0) { if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_STREAMON, &type) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, OPEN_READ, (_("Error starting streaming capture from device %s"), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); return FALSE; } } v4l2src->is_capturing = TRUE; return TRUE; } /****************************************************** * gst_v4l2src_capture_stop(): * stop streaming capture * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_stop (GstV4l2Src * v4l2src) { gint type = V4L2_BUF_TYPE_VIDEO_CAPTURE; GST_DEBUG_OBJECT (v4l2src, "stopping capturing"); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_ACTIVE (v4l2src->v4l2object); if (v4l2src->breq.memory != 0) { /* we actually need to sync on all queued buffers but not * on the non-queued ones */ if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_STREAMOFF, &type) < 0) { GST_ELEMENT_ERROR (v4l2src, RESOURCE, CLOSE, (_("Error stopping streaming capture from device %s: %s"), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); return FALSE; } } /* make an optional pending wait stop */ v4l2src->quit = TRUE; v4l2src->is_capturing = FALSE; return TRUE; } static void gst_v4l2src_buffer_pool_free (GstV4l2BufferPool * pool, gboolean do_close) { guint i; for (i = 0; i < pool->buffer_count; i++) { gst_atomic_int_set (&pool->buffers[i].refcount, 0); munmap (pool->buffers[i].start, pool->buffers[i].length); } g_free (pool->buffers); gst_atomic_int_set (&pool->refcount, 0); if (do_close) close (pool->video_fd); g_free (pool); } #if 0 void gst_v4l2src_free_buffer (GstBuffer * buffer) { GstV4l2Buffer *buf = (GstV4l2Buffer *) GST_BUFFER_PRIVATE (buffer); GST_LOG ("freeing buffer %p (nr. %d)", buffer, buf->buffer.index); if (!g_atomic_int_dec_and_test (&buf->refcount)) { /* we're still in use, add to queue again note: this might fail because the device is already stopped (race) */ if (ioctl (buf->pool->video_fd, VIDIOC_QBUF, &buf->buffer) < 0) GST_INFO ("readding to queue failed, assuming video device is stopped"); } if (g_atomic_int_dec_and_test (&buf->pool->refcount)) { /* we're last thing that used all this */ gst_v4l2src_buffer_pool_free (buf->pool, TRUE); } } #endif /****************************************************** * gst_v4l2src_capture_deinit(): * deinitialize the capture system * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_deinit (GstV4l2Src * v4l2src) { gint i; gboolean try_reinit = FALSE; GST_DEBUG_OBJECT (v4l2src, "deinitting capture system"); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_ACTIVE (v4l2src->v4l2object); if (v4l2src->pool) { /* free the buffers */ for (i = 0; i < v4l2src->breq.count; i++) { if (g_atomic_int_dec_and_test (&v4l2src->pool->buffers[i].refcount)) { if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_DQBUF, &v4l2src->pool->buffers[i].buffer) < 0) GST_DEBUG_OBJECT (v4l2src, "Could not dequeue buffer on uninitialization." "system error: %s. Will try reinit instead", g_strerror (errno)); try_reinit = TRUE; } } if (g_atomic_int_dec_and_test (&v4l2src->pool->refcount)) { /* we're last thing that used all this */ gst_v4l2src_buffer_pool_free (v4l2src->pool, FALSE); } v4l2src->pool = NULL; /* This is our second try to get the buffers dequeued. * Since buffers are normally dequeued automatically when capturing is * stopped, but may be enqueued before capturing has started, you get * a problem when you abort before capturing started but have enqueued * the buffers. We avoid that by starting/stopping capturing once so * they get auto-dequeued. */ if (try_reinit) { if (!gst_v4l2src_capture_start (v4l2src) || !gst_v4l2src_capture_stop (v4l2src)) return FALSE; } } GST_V4L2_SET_INACTIVE (v4l2src->v4l2object); return TRUE; } /* */ gboolean gst_v4l2src_get_size_limits (GstV4l2Src * v4l2src, struct v4l2_fmtdesc * format, gint * min_w, gint * max_w, gint * min_h, gint * max_h) { struct v4l2_format fmt; GST_LOG_OBJECT (v4l2src, "getting size limits with format %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (format->pixelformat)); /* get size delimiters */ memset (&fmt, 0, sizeof (fmt)); fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; fmt.fmt.pix.width = 0; fmt.fmt.pix.height = 0; fmt.fmt.pix.pixelformat = format->pixelformat; fmt.fmt.pix.field = V4L2_FIELD_INTERLACED; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_TRY_FMT, &fmt) < 0) { return FALSE; } if (min_w) *min_w = fmt.fmt.pix.width; if (min_h) *min_h = fmt.fmt.pix.height; GST_LOG_OBJECT (v4l2src, "got min size %dx%d", fmt.fmt.pix.width, fmt.fmt.pix.height); fmt.fmt.pix.width = 4096; fmt.fmt.pix.height = 4096; if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_TRY_FMT, &fmt) < 0) { return FALSE; } if (max_w) *max_w = fmt.fmt.pix.width; if (max_h) *max_h = fmt.fmt.pix.height; GST_LOG_OBJECT (v4l2src, "got max size %dx%d", fmt.fmt.pix.width, fmt.fmt.pix.height); return TRUE; } gboolean gst_v4l2src_update_fps (GstV4l2Object * v4l2object) { GstV4l2Src *v4l2src = GST_V4L2SRC (v4l2object->element); return gst_v4l2src_get_fps (v4l2src, &v4l2src->fps_n, &v4l2src->fps_d); } gboolean gst_v4l2src_set_fps (GstV4l2Src * v4l2src, guint * fps_n, guint * fps_d) { GstV4l2Object *v4l2object = v4l2src->v4l2object; struct v4l2_streamparm stream; memset (&stream, 0x00, sizeof (struct v4l2_streamparm)); stream.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (ioctl (v4l2object->video_fd, VIDIOC_G_PARM, &stream) == 0 && stream.parm.capture.capability & V4L2_CAP_TIMEPERFRAME) { stream.parm.capture.timeperframe.denominator = *fps_n; stream.parm.capture.timeperframe.numerator = *fps_d; if (ioctl (v4l2object->video_fd, VIDIOC_S_PARM, &stream) == 0) { *fps_n = stream.parm.capture.timeperframe.denominator; *fps_d = stream.parm.capture.timeperframe.numerator; return TRUE; } } return FALSE; } gboolean gst_v4l2src_get_fps (GstV4l2Src * v4l2src, guint * fps_n, guint * fps_d) { GstV4l2Object *v4l2object = v4l2src->v4l2object; v4l2_std_id std; struct v4l2_streamparm stream; const GList *item; if (!GST_V4L2_IS_OPEN (v4l2object)) return FALSE; /* Try to get the frame rate directly from the device using VIDIOC_G_PARM */ memset (&stream, 0x00, sizeof (struct v4l2_streamparm)); stream.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (ioctl (v4l2object->video_fd, VIDIOC_G_PARM, &stream) == 0 && stream.parm.capture.capability & V4L2_CAP_TIMEPERFRAME) { /* Note: V4L2 gives us the frame interval, we need the frame rate */ *fps_n = stream.parm.capture.timeperframe.denominator; *fps_d = stream.parm.capture.timeperframe.numerator; GST_DEBUG_OBJECT (v4l2src, "frame rate returned by G_PARM: %d/%d fps", *fps_n, *fps_d); return TRUE; } /* If G_PARM failed, try to get the same information from the video standard */ if (!gst_v4l2_get_norm (v4l2object, &std)) return FALSE; for (item = v4l2object->stds; item != NULL; item = item->next) { GstV4l2TunerNorm *v4l2norm = item->data; if (v4l2norm->index == std) { *fps_n = gst_value_get_fraction_numerator (&GST_TUNER_NORM (v4l2norm)-> framerate); *fps_d = gst_value_get_fraction_denominator (&GST_TUNER_NORM (v4l2norm)-> framerate); GST_DEBUG_OBJECT (v4l2src, "frame rate returned by get_norm: %d/%d fps", *fps_n, *fps_d); return TRUE; } } return FALSE; } #define GST_TYPE_V4L2SRC_BUFFER (gst_v4l2src_buffer_get_type()) #define GST_IS_V4L2SRC_BUFFER(obj) (G_TYPE_CHECK_INSTANCE_TYPE ((obj), GST_TYPE_V4L2SRC_BUFFER)) #define GST_V4L2SRC_BUFFER(obj) (G_TYPE_CHECK_INSTANCE_CAST ((obj), GST_TYPE_V4L2SRC_BUFFER, GstV4l2SrcBuffer)) typedef struct _GstV4l2SrcBuffer { GstBuffer buffer; GstV4l2Buffer *buf; } GstV4l2SrcBuffer; static void gst_v4l2src_buffer_class_init (gpointer g_class, gpointer class_data); static void gst_v4l2src_buffer_init (GTypeInstance * instance, gpointer g_class); static void gst_v4l2src_buffer_finalize (GstV4l2SrcBuffer * v4l2src_buffer); GType gst_v4l2src_buffer_get_type (void) { static GType _gst_v4l2src_buffer_type; if (G_UNLIKELY (_gst_v4l2src_buffer_type == 0)) { static const GTypeInfo v4l2src_buffer_info = { sizeof (GstBufferClass), NULL, NULL, gst_v4l2src_buffer_class_init, NULL, NULL, sizeof (GstV4l2SrcBuffer), 0, gst_v4l2src_buffer_init, NULL }; _gst_v4l2src_buffer_type = g_type_register_static (GST_TYPE_BUFFER, "GstV4l2SrcBuffer", &v4l2src_buffer_info, 0); } return _gst_v4l2src_buffer_type; } static void gst_v4l2src_buffer_class_init (gpointer g_class, gpointer class_data) { GstMiniObjectClass *mini_object_class = GST_MINI_OBJECT_CLASS (g_class); mini_object_class->finalize = (GstMiniObjectFinalizeFunction) gst_v4l2src_buffer_finalize; } static void gst_v4l2src_buffer_init (GTypeInstance * instance, gpointer g_class) { } static void gst_v4l2src_buffer_finalize (GstV4l2SrcBuffer * v4l2src_buffer) { GstV4l2Buffer *buf = v4l2src_buffer->buf; if (buf) { GST_LOG ("freeing buffer %p (nr. %d)", buf, buf->buffer.index); if (!g_atomic_int_dec_and_test (&buf->refcount)) { /* we're still in use, add to queue again note: this might fail because the device is already stopped (race) */ if (ioctl (buf->pool->video_fd, VIDIOC_QBUF, &buf->buffer) < 0) GST_INFO ("readding to queue failed, assuming video device is stopped"); } if (g_atomic_int_dec_and_test (&buf->pool->refcount)) { /* we're last thing that used all this */ gst_v4l2src_buffer_pool_free (buf->pool, TRUE); } } } /* Create a V4l2Src buffer from our mmap'd data area */ GstBuffer * gst_v4l2src_buffer_new (GstV4l2Src * v4l2src, guint size, guint8 * data, GstV4l2Buffer * srcbuf) { GstBuffer *buf; if (data == NULL) { buf = gst_buffer_new_and_alloc (size); } else { buf = (GstBuffer *) gst_mini_object_new (GST_TYPE_V4L2SRC_BUFFER); GST_BUFFER_DATA (buf) = data; GST_V4L2SRC_BUFFER (buf)->buf = srcbuf; GST_LOG_OBJECT (v4l2src, "creating buffer %p (nr. %d)", srcbuf, srcbuf->buffer.index); } GST_BUFFER_SIZE (buf) = size; GST_BUFFER_TIMESTAMP (buf) = gst_clock_get_time (GST_ELEMENT (v4l2src)->clock); GST_BUFFER_TIMESTAMP (buf) -= GST_ELEMENT (v4l2src)->base_time; GST_BUFFER_OFFSET (buf) = v4l2src->offset++; GST_BUFFER_FLAG_SET (buf, GST_BUFFER_FLAG_READONLY); if (v4l2src->fps_n > 0) { GST_BUFFER_DURATION (buf) = gst_util_uint64_scale_int (GST_SECOND, v4l2src->fps_n, v4l2src->fps_d); } else { GST_BUFFER_DURATION (buf) = GST_CLOCK_TIME_NONE; } /* the negotiate() method already set caps on the source pad */ gst_buffer_set_caps (buf, GST_PAD_CAPS (GST_BASE_SRC_PAD (v4l2src))); return buf; }