/* GStreamer * * Copyright (C) 2002 Ronald Bultje * 2006 Edgard Lima * * v4l2src.c - system calls * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include #include #include #include #include #include #include "v4l2src_calls.h" #include #include #ifdef __sun /* Needed on older Solaris Nevada builds (72 at least) */ #include #include #endif #include "gstv4l2tuner.h" GST_DEBUG_CATEGORY_EXTERN (v4l2src_debug); #define GST_CAT_DEFAULT v4l2src_debug /* lalala... */ #define GST_V4L2_SET_ACTIVE(element) (element)->buffer = GINT_TO_POINTER (-1) #define GST_V4L2_SET_INACTIVE(element) (element)->buffer = NULL /* On some systems MAP_FAILED seems to be missing */ #ifndef MAP_FAILED #define MAP_FAILED ((caddr_t) -1) #endif #define GST_TYPE_V4L2_BUFFER (gst_v4l2_buffer_get_type()) #define GST_IS_V4L2_BUFFER(obj) (G_TYPE_CHECK_INSTANCE_TYPE ((obj), GST_TYPE_V4L2_BUFFER)) #define GST_V4L2_BUFFER(obj) (G_TYPE_CHECK_INSTANCE_CAST ((obj), GST_TYPE_V4L2_BUFFER, GstV4l2Buffer)) static GstBufferClass *v4l2buffer_parent_class = NULL; /* Local functions */ static gboolean gst_v4l2src_get_nearest_size (GstV4l2Src * v4l2src, guint32 pixelformat, gint * width, gint * height); static void gst_v4l2_buffer_pool_destroy (GstV4l2BufferPool * pool); static void gst_v4l2_buffer_finalize (GstV4l2Buffer * buffer) { GstV4l2BufferPool *pool; gboolean resuscitated = FALSE; gint index; pool = buffer->pool; index = buffer->vbuffer.index; GST_LOG ("finalizing buffer %p %d", buffer, index); g_mutex_lock (pool->lock); if (GST_BUFFER_SIZE (buffer) != 0) /* BUFFER_SIZE is only set if the frame was dequeued */ pool->num_live_buffers--; if (pool->running) { if (v4l2_ioctl (pool->video_fd, VIDIOC_QBUF, &buffer->vbuffer) < 0) { GST_WARNING ("could not requeue buffer %p %d", buffer, index); } else { /* FIXME: check that the caps didn't change */ GST_LOG ("reviving buffer %p, %d", buffer, index); gst_buffer_ref (GST_BUFFER (buffer)); GST_BUFFER_SIZE (buffer) = 0; pool->buffers[index] = buffer; resuscitated = TRUE; } } else { GST_LOG ("the pool is shutting down"); } g_mutex_unlock (pool->lock); if (!resuscitated) { GST_LOG ("buffer %p not recovered, unmapping", buffer); gst_mini_object_unref (GST_MINI_OBJECT (pool)); v4l2_munmap ((void *) GST_BUFFER_DATA (buffer), buffer->vbuffer.length); GST_MINI_OBJECT_CLASS (v4l2buffer_parent_class)->finalize (GST_MINI_OBJECT (buffer)); } } static void gst_v4l2_buffer_class_init (gpointer g_class, gpointer class_data) { GstMiniObjectClass *mini_object_class = GST_MINI_OBJECT_CLASS (g_class); v4l2buffer_parent_class = g_type_class_peek_parent (g_class); mini_object_class->finalize = (GstMiniObjectFinalizeFunction) gst_v4l2_buffer_finalize; } static GType gst_v4l2_buffer_get_type (void) { static GType _gst_v4l2_buffer_type; if (G_UNLIKELY (_gst_v4l2_buffer_type == 0)) { static const GTypeInfo v4l2_buffer_info = { sizeof (GstBufferClass), NULL, NULL, gst_v4l2_buffer_class_init, NULL, NULL, sizeof (GstV4l2Buffer), 0, NULL, NULL }; _gst_v4l2_buffer_type = g_type_register_static (GST_TYPE_BUFFER, "GstV4l2Buffer", &v4l2_buffer_info, 0); } return _gst_v4l2_buffer_type; } static GstV4l2Buffer * gst_v4l2_buffer_new (GstV4l2BufferPool * pool, guint index, GstCaps * caps) { GstV4l2Buffer *ret; guint8 *data; ret = (GstV4l2Buffer *) gst_mini_object_new (GST_TYPE_V4L2_BUFFER); GST_LOG ("creating buffer %u, %p in pool %p", index, ret, pool); ret->pool = (GstV4l2BufferPool *) gst_mini_object_ref (GST_MINI_OBJECT (pool)); ret->vbuffer.index = index; ret->vbuffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; ret->vbuffer.memory = V4L2_MEMORY_MMAP; if (v4l2_ioctl (pool->video_fd, VIDIOC_QUERYBUF, &ret->vbuffer) < 0) goto querybuf_failed; GST_LOG (" index: %u", ret->vbuffer.index); GST_LOG (" type: %d", ret->vbuffer.type); GST_LOG (" bytesused: %u", ret->vbuffer.bytesused); GST_LOG (" flags: %08x", ret->vbuffer.flags); GST_LOG (" field: %d", ret->vbuffer.field); GST_LOG (" memory: %d", ret->vbuffer.memory); if (ret->vbuffer.memory == V4L2_MEMORY_MMAP) GST_LOG (" MMAP offset: %u", ret->vbuffer.m.offset); GST_LOG (" length: %u", ret->vbuffer.length); GST_LOG (" input: %u", ret->vbuffer.input); data = (guint8 *) v4l2_mmap (0, ret->vbuffer.length, PROT_READ | PROT_WRITE, MAP_SHARED, pool->video_fd, ret->vbuffer.m.offset); if (data == MAP_FAILED) goto mmap_failed; GST_BUFFER_DATA (ret) = data; GST_BUFFER_SIZE (ret) = ret->vbuffer.length; GST_BUFFER_FLAG_SET (ret, GST_BUFFER_FLAG_READONLY); gst_buffer_set_caps (GST_BUFFER (ret), caps); return ret; /* ERRORS */ querybuf_failed: { gint errnosave = errno; GST_WARNING ("Failed QUERYBUF: %s", g_strerror (errnosave)); gst_buffer_unref (GST_BUFFER (ret)); errno = errnosave; return NULL; } mmap_failed: { gint errnosave = errno; GST_WARNING ("Failed to mmap: %s", g_strerror (errnosave)); gst_buffer_unref (GST_BUFFER (ret)); errno = errnosave; return NULL; } } #define GST_TYPE_V4L2_BUFFER_POOL (gst_v4l2_buffer_pool_get_type()) #define GST_IS_V4L2_BUFFER_POOL(obj) (G_TYPE_CHECK_INSTANCE_TYPE ((obj), GST_TYPE_V4L2_BUFFER_POOL)) #define GST_V4L2_BUFFER_POOL(obj) (G_TYPE_CHECK_INSTANCE_CAST ((obj), GST_TYPE_V4L2_BUFFER_POOL, GstV4l2BufferPool)) static GstMiniObjectClass *buffer_pool_parent_class = NULL; static void gst_v4l2_buffer_pool_finalize (GstV4l2BufferPool * pool) { g_mutex_free (pool->lock); pool->lock = NULL; if (pool->video_fd >= 0) v4l2_close (pool->video_fd); if (pool->buffers) { g_free (pool->buffers); pool->buffers = NULL; } GST_MINI_OBJECT_CLASS (buffer_pool_parent_class)->finalize (GST_MINI_OBJECT (pool)); } static void gst_v4l2_buffer_pool_init (GstV4l2BufferPool * pool, gpointer g_class) { pool->lock = g_mutex_new (); pool->running = FALSE; pool->num_live_buffers = 0; } static void gst_v4l2_buffer_pool_class_init (gpointer g_class, gpointer class_data) { GstMiniObjectClass *mini_object_class = GST_MINI_OBJECT_CLASS (g_class); buffer_pool_parent_class = g_type_class_peek_parent (g_class); mini_object_class->finalize = (GstMiniObjectFinalizeFunction) gst_v4l2_buffer_pool_finalize; } static GType gst_v4l2_buffer_pool_get_type (void) { static GType _gst_v4l2_buffer_pool_type; if (G_UNLIKELY (_gst_v4l2_buffer_pool_type == 0)) { static const GTypeInfo v4l2_buffer_pool_info = { sizeof (GstBufferClass), NULL, NULL, gst_v4l2_buffer_pool_class_init, NULL, NULL, sizeof (GstV4l2BufferPool), 0, (GInstanceInitFunc) gst_v4l2_buffer_pool_init, NULL }; _gst_v4l2_buffer_pool_type = g_type_register_static (GST_TYPE_MINI_OBJECT, "GstV4l2BufferPool", &v4l2_buffer_pool_info, 0); } return _gst_v4l2_buffer_pool_type; } static GstV4l2BufferPool * gst_v4l2_buffer_pool_new (GstV4l2Src * v4l2src, gint fd, gint num_buffers, GstCaps * caps) { GstV4l2BufferPool *pool; gint n; pool = (GstV4l2BufferPool *) gst_mini_object_new (GST_TYPE_V4L2_BUFFER_POOL); pool->video_fd = v4l2_dup (fd); if (pool->video_fd < 0) goto dup_failed; pool->buffer_count = num_buffers; pool->buffers = g_new0 (GstV4l2Buffer *, num_buffers); for (n = 0; n < num_buffers; n++) { pool->buffers[n] = gst_v4l2_buffer_new (pool, n, caps); if (!pool->buffers[n]) goto buffer_new_failed; } return pool; /* ERRORS */ dup_failed: { gint errnosave = errno; gst_mini_object_unref (GST_MINI_OBJECT (pool)); errno = errnosave; return NULL; } buffer_new_failed: { gint errnosave = errno; gst_v4l2_buffer_pool_destroy (pool); errno = errnosave; return NULL; } } static gboolean gst_v4l2_buffer_pool_activate (GstV4l2BufferPool * pool, GstV4l2Src * v4l2src) { gint n; g_mutex_lock (pool->lock); for (n = 0; n < pool->buffer_count; n++) { struct v4l2_buffer *buf; buf = &pool->buffers[n]->vbuffer; GST_LOG ("enqueue pool buffer %d", n); if (v4l2_ioctl (pool->video_fd, VIDIOC_QBUF, buf) < 0) goto queue_failed; } pool->running = TRUE; g_mutex_unlock (pool->lock); return TRUE; /* ERRORS */ queue_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not enqueue buffers in device '%s'."), v4l2src->v4l2object->videodev), ("enqueing buffer %d/%d failed: %s", n, v4l2src->num_buffers, g_strerror (errno))); g_mutex_unlock (pool->lock); return FALSE; } } static void gst_v4l2_buffer_pool_destroy (GstV4l2BufferPool * pool) { gint n; g_mutex_lock (pool->lock); pool->running = FALSE; g_mutex_unlock (pool->lock); GST_DEBUG ("destroy pool"); /* after this point, no more buffers will be queued or dequeued; no buffer * from pool->buffers that is NULL will be set to a buffer, and no buffer that * is not NULL will be pushed out. */ /* miniobjects have no dispose, so they can't break ref-cycles, as buffers ref * the pool, we need to unref the buffer to properly finalize te pool */ for (n = 0; n < pool->buffer_count; n++) { GstBuffer *buf; g_mutex_lock (pool->lock); buf = GST_BUFFER (pool->buffers[n]); g_mutex_unlock (pool->lock); if (buf) /* we own the ref if the buffer is in pool->buffers; drop it. */ gst_buffer_unref (buf); } gst_mini_object_unref (GST_MINI_OBJECT (pool)); } /* complete made up ranking, the values themselves are meaningless */ #define YUV_BASE_RANK 1000 #define JPEG_BASE_RANK 500 #define DV_BASE_RANK 200 #define RGB_BASE_RANK 100 #define YUV_ODD_BASE_RANK 50 #define RGB_ODD_BASE_RANK 25 #define BAYER_BASE_RANK 15 #define S910_BASE_RANK 10 #define GREY_BASE_RANK 5 #define PWC_BASE_RANK 1 static gint gst_v4l2src_format_get_rank (guint32 fourcc) { switch (fourcc) { case V4L2_PIX_FMT_MJPEG: return JPEG_BASE_RANK; case V4L2_PIX_FMT_JPEG: return JPEG_BASE_RANK + 1; case V4L2_PIX_FMT_RGB332: case V4L2_PIX_FMT_RGB555: case V4L2_PIX_FMT_RGB555X: case V4L2_PIX_FMT_RGB565: case V4L2_PIX_FMT_RGB565X: return RGB_ODD_BASE_RANK; case V4L2_PIX_FMT_RGB24: case V4L2_PIX_FMT_BGR24: return RGB_BASE_RANK - 1; case V4L2_PIX_FMT_RGB32: case V4L2_PIX_FMT_BGR32: return RGB_BASE_RANK; case V4L2_PIX_FMT_GREY: /* 8 Greyscale */ return GREY_BASE_RANK; case V4L2_PIX_FMT_NV12: /* 12 Y/CbCr 4:2:0 */ case V4L2_PIX_FMT_NV21: /* 12 Y/CrCb 4:2:0 */ case V4L2_PIX_FMT_YYUV: /* 16 YUV 4:2:2 */ case V4L2_PIX_FMT_HI240: /* 8 8-bit color */ return YUV_ODD_BASE_RANK; case V4L2_PIX_FMT_YVU410: /* YVU9, 9 bits per pixel */ return YUV_BASE_RANK + 3; case V4L2_PIX_FMT_YUV410: /* YUV9, 9 bits per pixel */ return YUV_BASE_RANK + 2; case V4L2_PIX_FMT_YUV420: /* I420, 12 bits per pixel */ return YUV_BASE_RANK + 7; case V4L2_PIX_FMT_YUYV: /* YUY2, 16 bits per pixel */ return YUV_BASE_RANK + 10; case V4L2_PIX_FMT_YVU420: /* YV12, 12 bits per pixel */ return YUV_BASE_RANK + 6; case V4L2_PIX_FMT_UYVY: /* UYVY, 16 bits per pixel */ return YUV_BASE_RANK + 9; case V4L2_PIX_FMT_Y41P: /* Y41P, 12 bits per pixel */ return YUV_BASE_RANK + 5; case V4L2_PIX_FMT_YUV411P: /* Y41B, 12 bits per pixel */ return YUV_BASE_RANK + 4; case V4L2_PIX_FMT_YUV422P: /* Y42B, 16 bits per pixel */ return YUV_BASE_RANK + 8; case V4L2_PIX_FMT_DV: return DV_BASE_RANK; case V4L2_PIX_FMT_MPEG: /* MPEG */ case V4L2_PIX_FMT_WNVA: /* Winnov hw compres */ return 0; #ifdef V4L2_PIX_FMT_SBGGR8 case V4L2_PIX_FMT_SBGGR8: return BAYER_BASE_RANK; #endif #ifdef V4L2_PIX_FMT_SN9C10X case V4L2_PIX_FMT_SN9C10X: return S910_BASE_RANK; #endif #ifdef V4L2_PIX_FMT_PWC1 case V4L2_PIX_FMT_PWC1: return PWC_BASE_RANK; #endif #ifdef V4L2_PIX_FMT_PWC2 case V4L2_PIX_FMT_PWC2: return PWC_BASE_RANK; #endif default: break; } return 0; } static gint gst_v4l2src_format_cmp_func (gconstpointer a, gconstpointer b) { guint32 pf1 = ((struct v4l2_fmtdesc *) a)->pixelformat; guint32 pf2 = ((struct v4l2_fmtdesc *) b)->pixelformat; if (pf1 == pf2) return 0; return gst_v4l2src_format_get_rank (pf2) - gst_v4l2src_format_get_rank (pf1); } /****************************************************** * gst_v4l2src_fill_format_list(): * create list of supported capture formats * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_fill_format_list (GstV4l2Src * v4l2src) { gint n; struct v4l2_fmtdesc *format; GST_DEBUG_OBJECT (v4l2src, "getting src format enumerations"); /* format enumeration */ for (n = 0;; n++) { format = g_new0 (struct v4l2_fmtdesc, 1); format->index = n; format->type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (v4l2_ioctl (v4l2src->v4l2object->video_fd, VIDIOC_ENUM_FMT, format) < 0) { if (errno == EINVAL) { g_free (format); break; /* end of enumeration */ } else { goto failed; } } GST_LOG_OBJECT (v4l2src, "index: %u", format->index); GST_LOG_OBJECT (v4l2src, "type: %d", format->type); GST_LOG_OBJECT (v4l2src, "flags: %08x", format->flags); GST_LOG_OBJECT (v4l2src, "description: '%s'", format->description); GST_LOG_OBJECT (v4l2src, "pixelformat: %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (format->pixelformat)); /* sort formats according to our preference; we do this, because caps * are probed in the order the formats are in the list, and the order of * formats in the final probed caps matters for things like fixation */ v4l2src->formats = g_slist_insert_sorted (v4l2src->formats, format, (GCompareFunc) gst_v4l2src_format_cmp_func); } GST_DEBUG_OBJECT (v4l2src, "got %d format(s)", n); return TRUE; /* ERRORS */ failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("Failed to enumerate possible video formats device '%s' can work with"), v4l2src->v4l2object->videodev), ("Failed to get number %d in pixelformat enumeration for %s. (%d - %s)", n, v4l2src->v4l2object->videodev, errno, g_strerror (errno))); g_free (format); return FALSE; } } /****************************************************** * gst_v4l2src_clear_format_list(): * free list of supported capture formats * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_clear_format_list (GstV4l2Src * v4l2src) { g_slist_foreach (v4l2src->formats, (GFunc) g_free, NULL); g_slist_free (v4l2src->formats); v4l2src->formats = NULL; return TRUE; } /* The frame interval enumeration code first appeared in Linux 2.6.19. */ #ifdef VIDIOC_ENUM_FRAMEINTERVALS static GstStructure * gst_v4l2src_probe_caps_for_format_and_size (GstV4l2Src * v4l2src, guint32 pixelformat, guint32 width, guint32 height, const GstStructure * template) { gint fd = v4l2src->v4l2object->video_fd; struct v4l2_frmivalenum ival; guint32 num, denom; GstStructure *s; GValue rates = { 0, }; memset (&ival, 0, sizeof (struct v4l2_frmivalenum)); ival.index = 0; ival.pixel_format = pixelformat; ival.width = width; ival.height = height; GST_LOG_OBJECT (v4l2src, "get frame interval for %ux%u, %" GST_FOURCC_FORMAT, width, height, GST_FOURCC_ARGS (pixelformat)); /* keep in mind that v4l2 gives us frame intervals (durations); we invert the * fraction to get framerate */ if (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMEINTERVALS, &ival) < 0) goto enum_frameintervals_failed; if (ival.type == V4L2_FRMIVAL_TYPE_DISCRETE) { GValue rate = { 0, }; g_value_init (&rates, GST_TYPE_LIST); g_value_init (&rate, GST_TYPE_FRACTION); do { num = ival.discrete.numerator; denom = ival.discrete.denominator; if (num > G_MAXINT || denom > G_MAXINT) { /* let us hope we don't get here... */ num >>= 1; denom >>= 1; } GST_LOG_OBJECT (v4l2src, "adding discrete framerate: %d/%d", denom, num); /* swap to get the framerate */ gst_value_set_fraction (&rate, denom, num); gst_value_list_append_value (&rates, &rate); ival.index++; } while (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMEINTERVALS, &ival) >= 0); } else if (ival.type == V4L2_FRMIVAL_TYPE_STEPWISE) { GValue min = { 0, }; GValue step = { 0, }; GValue max = { 0, }; gboolean added = FALSE; guint32 minnum, mindenom; guint32 maxnum, maxdenom; g_value_init (&rates, GST_TYPE_LIST); g_value_init (&min, GST_TYPE_FRACTION); g_value_init (&step, GST_TYPE_FRACTION); g_value_init (&max, GST_TYPE_FRACTION); /* get the min */ minnum = ival.stepwise.min.numerator; mindenom = ival.stepwise.min.denominator; if (minnum > G_MAXINT || mindenom > G_MAXINT) { minnum >>= 1; mindenom >>= 1; } GST_LOG_OBJECT (v4l2src, "stepwise min frame interval: %d/%d", minnum, mindenom); gst_value_set_fraction (&min, minnum, mindenom); /* get the max */ maxnum = ival.stepwise.max.numerator; maxdenom = ival.stepwise.max.denominator; if (maxnum > G_MAXINT || maxdenom > G_MAXINT) { maxnum >>= 1; maxdenom >>= 1; } GST_LOG_OBJECT (v4l2src, "stepwise max frame interval: %d/%d", maxnum, maxdenom); gst_value_set_fraction (&max, maxnum, maxdenom); /* get the step */ num = ival.stepwise.step.numerator; denom = ival.stepwise.step.denominator; if (num > G_MAXINT || denom > G_MAXINT) { num >>= 1; denom >>= 1; } if (num == 0 || denom == 0) { /* in this case we have a wrong fraction or no step, set the step to max * so that we only add the min value in the loop below */ num = maxnum; denom = maxdenom; } /* since we only have gst_value_fraction_subtract and not add, negate the * numerator */ GST_LOG_OBJECT (v4l2src, "stepwise step frame interval: %d/%d", num, denom); gst_value_set_fraction (&step, -num, denom); while (gst_value_compare (&min, &max) <= 0) { GValue rate = { 0, }; num = gst_value_get_fraction_numerator (&min); denom = gst_value_get_fraction_denominator (&min); GST_LOG_OBJECT (v4l2src, "adding stepwise framerate: %d/%d", denom, num); /* invert to get the framerate */ g_value_init (&rate, GST_TYPE_FRACTION); gst_value_set_fraction (&rate, denom, num); gst_value_list_append_value (&rates, &rate); added = TRUE; /* we're actually adding because step was negated above. This is because * there is no _add function... */ if (!gst_value_fraction_subtract (&min, &min, &step)) { GST_WARNING_OBJECT (v4l2src, "could not step fraction!"); break; } } if (!added) { /* no range was added, leave the default range from the template */ GST_WARNING_OBJECT (v4l2src, "no range added, leaving default"); g_value_unset (&rates); } } else if (ival.type == V4L2_FRMIVAL_TYPE_CONTINUOUS) { guint32 maxnum, maxdenom; g_value_init (&rates, GST_TYPE_FRACTION_RANGE); num = ival.stepwise.min.numerator; denom = ival.stepwise.min.denominator; if (num > G_MAXINT || denom > G_MAXINT) { num >>= 1; denom >>= 1; } maxnum = ival.stepwise.max.numerator; maxdenom = ival.stepwise.max.denominator; if (maxnum > G_MAXINT || maxdenom > G_MAXINT) { maxnum >>= 1; maxdenom >>= 1; } GST_LOG_OBJECT (v4l2src, "continuous frame interval %d/%d to %d/%d", maxdenom, maxnum, denom, num); gst_value_set_fraction_range_full (&rates, maxdenom, maxnum, denom, num); } else { goto unknown_type; } return_data: s = gst_structure_copy (template); gst_structure_set (s, "width", G_TYPE_INT, (gint) width, "height", G_TYPE_INT, (gint) height, NULL); if (G_IS_VALUE (&rates)) { /* only change the framerate on the template when we have a valid probed new * value */ gst_structure_set_value (s, "framerate", &rates); g_value_unset (&rates); } return s; /* ERRORS */ enum_frameintervals_failed: { GST_DEBUG_OBJECT (v4l2src, "Unable to enumerate intervals for %" GST_FOURCC_FORMAT "@%ux%u", GST_FOURCC_ARGS (pixelformat), width, height); goto return_data; } unknown_type: { /* I don't see how this is actually an error, we ignore the format then */ GST_WARNING_OBJECT (v4l2src, "Unknown frame interval type at %" GST_FOURCC_FORMAT "@%ux%u: %u", GST_FOURCC_ARGS (pixelformat), width, height, ival.type); return NULL; } } #endif /* defined VIDIOC_ENUM_FRAMEINTERVALS */ static gint sort_by_frame_size (GstStructure * s1, GstStructure * s2) { int w1, h1, w2, h2; gst_structure_get_int (s1, "width", &w1); gst_structure_get_int (s1, "height", &h1); gst_structure_get_int (s2, "width", &w2); gst_structure_get_int (s2, "height", &h2); /* I think it's safe to assume that this won't overflow for a while */ return ((w2 * h2) - (w1 * h1)); } GstCaps * gst_v4l2src_probe_caps_for_format (GstV4l2Src * v4l2src, guint32 pixelformat, const GstStructure * template) { GstCaps *ret = gst_caps_new_empty (); GstStructure *tmp; #ifdef VIDIOC_ENUM_FRAMESIZES gint fd = v4l2src->v4l2object->video_fd; struct v4l2_frmsizeenum size; GList *results = NULL; guint32 w, h; memset (&size, 0, sizeof (struct v4l2_frmsizeenum)); size.index = 0; size.pixel_format = pixelformat; GST_DEBUG_OBJECT (v4l2src, "Enumerating frame sizes"); if (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMESIZES, &size) < 0) goto enum_framesizes_failed; if (size.type == V4L2_FRMSIZE_TYPE_DISCRETE) { do { GST_LOG_OBJECT (v4l2src, "got discrete frame size %dx%d", size.discrete.width, size.discrete.height); w = MIN (size.discrete.width, G_MAXINT); h = MIN (size.discrete.height, G_MAXINT); tmp = gst_v4l2src_probe_caps_for_format_and_size (v4l2src, pixelformat, w, h, template); if (tmp) results = g_list_prepend (results, tmp); size.index++; } while (v4l2_ioctl (fd, VIDIOC_ENUM_FRAMESIZES, &size) >= 0); GST_DEBUG_OBJECT (v4l2src, "done iterating discrete frame sizes"); } else if (size.type == V4L2_FRMSIZE_TYPE_STEPWISE) { GST_DEBUG_OBJECT (v4l2src, "we have stepwise frame sizes:"); GST_DEBUG_OBJECT (v4l2src, "min width: %d", size.stepwise.min_width); GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.min_height); GST_DEBUG_OBJECT (v4l2src, "max width: %d", size.stepwise.max_width); GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.max_height); GST_DEBUG_OBJECT (v4l2src, "step width: %d", size.stepwise.step_width); GST_DEBUG_OBJECT (v4l2src, "step height: %d", size.stepwise.step_height); for (w = size.stepwise.min_width, h = size.stepwise.min_height; w < size.stepwise.max_width && h < size.stepwise.max_height; w += size.stepwise.step_width, h += size.stepwise.step_height) { if (w == 0 || h == 0) continue; tmp = gst_v4l2src_probe_caps_for_format_and_size (v4l2src, pixelformat, w, h, template); if (tmp) results = g_list_prepend (results, tmp); } GST_DEBUG_OBJECT (v4l2src, "done iterating stepwise frame sizes"); } else if (size.type == V4L2_FRMSIZE_TYPE_CONTINUOUS) { guint32 maxw, maxh; GST_DEBUG_OBJECT (v4l2src, "we have continuous frame sizes:"); GST_DEBUG_OBJECT (v4l2src, "min width: %d", size.stepwise.min_width); GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.min_height); GST_DEBUG_OBJECT (v4l2src, "max width: %d", size.stepwise.max_width); GST_DEBUG_OBJECT (v4l2src, "min height: %d", size.stepwise.max_height); w = MAX (size.stepwise.min_width, 1); h = MAX (size.stepwise.min_height, 1); maxw = MIN (size.stepwise.max_width, G_MAXINT); maxh = MIN (size.stepwise.max_height, G_MAXINT); tmp = gst_v4l2src_probe_caps_for_format_and_size (v4l2src, pixelformat, w, h, template); if (tmp) { gst_structure_set (tmp, "width", GST_TYPE_INT_RANGE, (gint) w, (gint) maxw, "height", GST_TYPE_INT_RANGE, (gint) h, (gint) maxh, NULL); /* no point using the results list here, since there's only one struct */ gst_caps_append_structure (ret, tmp); } } else { goto unknown_type; } /* we use an intermediary list to store and then sort the results of the * probing because we can't make any assumptions about the order in which * the driver will give us the sizes, but we want the final caps to contain * the results starting with the highest resolution and having the lowest * resolution last, since order in caps matters for things like fixation. */ results = g_list_sort (results, (GCompareFunc) sort_by_frame_size); while (results != NULL) { gst_caps_append_structure (ret, GST_STRUCTURE (results->data)); results = g_list_delete_link (results, results); } if (gst_caps_is_empty (ret)) goto enum_framesizes_no_results; return ret; /* ERRORS */ enum_framesizes_failed: { /* I don't see how this is actually an error */ GST_DEBUG_OBJECT (v4l2src, "Failed to enumerate frame sizes for pixelformat %" GST_FOURCC_FORMAT " (%s)", GST_FOURCC_ARGS (pixelformat), g_strerror (errno)); goto default_frame_sizes; } enum_framesizes_no_results: { /* it's possible that VIDIOC_ENUM_FRAMESIZES is defined but the driver in * question doesn't actually support it yet */ GST_DEBUG_OBJECT (v4l2src, "No results for pixelformat %" GST_FOURCC_FORMAT " enumerating frame sizes, trying fallback", GST_FOURCC_ARGS (pixelformat)); goto default_frame_sizes; } unknown_type: { GST_WARNING_OBJECT (v4l2src, "Unknown frame sizeenum type for pixelformat %" GST_FOURCC_FORMAT ": %u", GST_FOURCC_ARGS (pixelformat), size.type); goto default_frame_sizes; } default_frame_sizes: #endif /* defined VIDIOC_ENUM_FRAMESIZES */ { gint min_w, max_w, min_h, max_h, fix_num = 0, fix_denom = 0; /* This code is for Linux < 2.6.19 */ min_w = min_h = 1; max_w = max_h = GST_V4L2_MAX_SIZE; if (!gst_v4l2src_get_nearest_size (v4l2src, pixelformat, &min_w, &min_h)) { GST_WARNING_OBJECT (v4l2src, "Could not probe minimum capture size for pixelformat %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (pixelformat)); } if (!gst_v4l2src_get_nearest_size (v4l2src, pixelformat, &max_w, &max_h)) { GST_WARNING_OBJECT (v4l2src, "Could not probe maximum capture size for pixelformat %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (pixelformat)); } /* Since we can't get framerate directly, try to use the current norm */ if (v4l2src->v4l2object->norm && v4l2src->v4l2object->norms) { GList *norms; GstTunerNorm *norm; for (norms = v4l2src->v4l2object->norms; norms != NULL; norms = norms->next) { norm = (GstTunerNorm *) norms->data; if (!strcmp (norm->label, v4l2src->v4l2object->norm)) break; } /* If it's possible, set framerate to that (discrete) value */ if (norm) { fix_num = gst_value_get_fraction_numerator (&norm->framerate); fix_denom = gst_value_get_fraction_denominator (&norm->framerate); } } tmp = gst_structure_copy (template); if (fix_num) { gst_structure_set (tmp, "width", GST_TYPE_INT_RANGE, min_w, max_w, "height", GST_TYPE_INT_RANGE, min_h, max_h, "framerate", GST_TYPE_FRACTION, fix_num, fix_denom, NULL); } else { /* if norm can't be used, copy the template framerate */ gst_structure_set (tmp, "width", GST_TYPE_INT_RANGE, min_w, max_w, "height", GST_TYPE_INT_RANGE, min_h, max_h, "framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 100, 1, NULL); } gst_caps_append_structure (ret, tmp); return ret; } } /****************************************************** * gst_v4l2src_grab_frame (): * grab a frame for capturing * return value: GST_FLOW_OK or GST_FLOW_ERROR ******************************************************/ GstFlowReturn gst_v4l2src_grab_frame (GstV4l2Src * v4l2src, GstBuffer ** buf) { #define NUM_TRIALS 50 struct v4l2_buffer buffer; gint32 trials = NUM_TRIALS; GstBuffer *pool_buffer; gboolean need_copy; gint index; memset (&buffer, 0x00, sizeof (buffer)); buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; buffer.memory = V4L2_MEMORY_MMAP; while (v4l2_ioctl (v4l2src->v4l2object->video_fd, VIDIOC_DQBUF, &buffer) < 0) { GST_WARNING_OBJECT (v4l2src, "problem grabbing frame %d (ix=%d), trials=%d, pool-ct=%d, buf.flags=%d", buffer.sequence, buffer.index, trials, GST_MINI_OBJECT_REFCOUNT (v4l2src->pool), buffer.flags); /* if the sync() got interrupted, we can retry */ switch (errno) { case EAGAIN: GST_WARNING_OBJECT (v4l2src, "Non-blocking I/O has been selected using O_NONBLOCK and" " no buffer was in the outgoing queue. device %s", v4l2src->v4l2object->videodev); break; case EINVAL: goto einval; case ENOMEM: goto enomem; case EIO: GST_INFO_OBJECT (v4l2src, "VIDIOC_DQBUF failed due to an internal error." " Can also indicate temporary problems like signal loss." " Note the driver might dequeue an (empty) buffer despite" " returning an error, or even stop capturing." " device %s", v4l2src->v4l2object->videodev); /* have we de-queued a buffer ? */ if (!(buffer.flags & (V4L2_BUF_FLAG_QUEUED | V4L2_BUF_FLAG_DONE))) { /* this fails if ((buffer.index >= 0) && (buffer.index < v4l2src->breq.count)) { GST_DEBUG_OBJECT (v4l2src, "reenqueing buffer (ix=%ld)", buffer.index); gst_v4l2src_queue_frame (v4l2src, buffer.index); } else { */ GST_DEBUG_OBJECT (v4l2src, "reenqueing buffer"); /* FIXME: this is not a good idea, as drivers usualy return the buffer * with index-number set to 0, thus the re-enque will fail unless it * was incidentialy 0. * We could try to re-enque all buffers without handling the ioctl * return. */ /* if (ioctl (v4l2src->v4l2object->video_fd, VIDIOC_QBUF, &buffer) < 0) { goto qbuf_failed; } */ /*} */ } break; case EINTR: GST_WARNING_OBJECT (v4l2src, "could not sync on a buffer on device %s", v4l2src->v4l2object->videodev); break; default: GST_WARNING_OBJECT (v4l2src, "Grabbing frame got interrupted on %s. No expected reason.", v4l2src->v4l2object->videodev); break; } /* check nr. of attempts to capture */ if (--trials == -1) { goto too_many_trials; } else { memset (&buffer, 0x00, sizeof (buffer)); buffer.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; buffer.memory = V4L2_MEMORY_MMAP; } } g_mutex_lock (v4l2src->pool->lock); index = buffer.index; /* get our GstBuffer with that index from the pool, if the buffer was * outstanding we have a serious problem. */ pool_buffer = GST_BUFFER (v4l2src->pool->buffers[index]); if (pool_buffer == NULL) goto no_buffer; GST_LOG_OBJECT (v4l2src, "grabbed buffer %p at index %d", pool_buffer, index); /* we have the buffer now, mark the spot in the pool empty */ v4l2src->pool->buffers[index] = NULL; v4l2src->pool->num_live_buffers++; /* if we are handing out the last buffer in the pool, we need to make a * copy and bring the buffer back in the pool. */ need_copy = v4l2src->always_copy || (v4l2src->pool->num_live_buffers == v4l2src->pool->buffer_count); g_mutex_unlock (v4l2src->pool->lock); /* this can change at every frame, esp. with jpeg */ GST_BUFFER_SIZE (pool_buffer) = buffer.bytesused; GST_BUFFER_OFFSET (pool_buffer) = v4l2src->offset++; GST_BUFFER_OFFSET_END (pool_buffer) = v4l2src->offset; /* timestamps, LOCK to get clock and base time. */ { GstClock *clock; GstClockTime timestamp; GST_OBJECT_LOCK (v4l2src); if ((clock = GST_ELEMENT_CLOCK (v4l2src))) { /* we have a clock, get base time and ref clock */ timestamp = GST_ELEMENT (v4l2src)->base_time; gst_object_ref (clock); } else { /* no clock, can't set timestamps */ timestamp = GST_CLOCK_TIME_NONE; } GST_OBJECT_UNLOCK (v4l2src); if (clock) { /* the time now is the time of the clock minus the base time */ timestamp = gst_clock_get_time (clock) - timestamp; gst_object_unref (clock); } /* FIXME: use the timestamp from the buffer itself! */ GST_BUFFER_TIMESTAMP (pool_buffer) = timestamp; } if (G_UNLIKELY (need_copy)) { *buf = gst_buffer_copy (pool_buffer); GST_BUFFER_FLAG_UNSET (*buf, GST_BUFFER_FLAG_READONLY); /* this will requeue */ gst_buffer_unref (pool_buffer); } else { *buf = pool_buffer; } GST_LOG_OBJECT (v4l2src, "grabbed frame %d (ix=%d), flags %08x, pool-ct=%d", buffer.sequence, buffer.index, buffer.flags, v4l2src->pool->num_live_buffers); return GST_FLOW_OK; /* ERRORS */ einval: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED, (_("Failed trying to get video frames from device '%s'."), v4l2src->v4l2object->videodev), (_("The buffer type is not supported, or the index is out of bounds," " or no buffers have been allocated yet, or the userptr" " or length are invalid. device %s"), v4l2src->v4l2object->videodev)); return GST_FLOW_ERROR; } enomem: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED, (_("Failed trying to get video frames from device '%s'. Not enough memory."), v4l2src->v4l2object->videodev), (_("insufficient memory to enqueue a user pointer buffer. device %s."), v4l2src->v4l2object->videodev)); return GST_FLOW_ERROR; } too_many_trials: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED, (_("Failed trying to get video frames from device '%s'."), v4l2src->v4l2object->videodev), (_("Failed after %d tries. device %s. system error: %s"), NUM_TRIALS, v4l2src->v4l2object->videodev, g_strerror (errno))); return GST_FLOW_ERROR; } no_buffer: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, FAILED, (_("Failed trying to get video frames from device '%s'."), v4l2src->v4l2object->videodev), (_("No free buffers found in the pool at index %d."), index)); g_mutex_unlock (v4l2src->pool->lock); return GST_FLOW_ERROR; } /* qbuf_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, WRITE, (_("Could not exchange data with device '%s'."), v4l2src->v4l2object->videodev), ("Error queueing buffer on device %s. system error: %s", v4l2src->v4l2object->videodev, g_strerror (errno))); return GST_FLOW_ERROR; } */ } /* Note about fraction simplification * n1/d1 == n2/d2 is also written as n1 == ( n2 * d1 ) / d2 */ #define fractions_are_equal(n1,d1,n2,d2) ((n1) == gst_util_uint64_scale_int((n2), (d1), (d2))) /****************************************************** * gst_v4l2src_set_capture(): * set capture parameters * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_set_capture (GstV4l2Src * v4l2src, guint32 pixelformat, guint32 width, guint32 height, guint fps_n, guint fps_d) { gint fd = v4l2src->v4l2object->video_fd; struct v4l2_format format; struct v4l2_streamparm stream; GST_DEBUG_OBJECT (v4l2src, "Setting capture format to %dx%d, format " "%" GST_FOURCC_FORMAT, width, height, GST_FOURCC_ARGS (pixelformat)); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_NOT_ACTIVE (v4l2src->v4l2object); memset (&format, 0x00, sizeof (struct v4l2_format)); format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (v4l2_ioctl (fd, VIDIOC_G_FMT, &format) < 0) goto get_fmt_failed; format.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; format.fmt.pix.width = width; format.fmt.pix.height = height; format.fmt.pix.pixelformat = pixelformat; /* request whole frames; change when gstreamer supports interlaced video * (INTERLACED mode returns frames where the fields have already been * combined, there are other modes for requesting fields individually) */ format.fmt.pix.field = V4L2_FIELD_INTERLACED; if (v4l2_ioctl (fd, VIDIOC_S_FMT, &format) < 0) { if (errno != EINVAL) goto set_fmt_failed; /* try again with progressive video */ format.fmt.pix.width = width; format.fmt.pix.height = height; format.fmt.pix.pixelformat = pixelformat; format.fmt.pix.field = V4L2_FIELD_NONE; if (v4l2_ioctl (fd, VIDIOC_S_FMT, &format) < 0) goto set_fmt_failed; } if (format.fmt.pix.width != width || format.fmt.pix.height != height) goto invalid_dimensions; if (format.fmt.pix.pixelformat != pixelformat) goto invalid_pixelformat; /* Is there a reason we require the caller to always specify a framerate? */ GST_LOG_OBJECT (v4l2src, "Desired framerate: %u/%u", fps_n, fps_d); memset (&stream, 0x00, sizeof (struct v4l2_streamparm)); stream.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if (v4l2_ioctl (fd, VIDIOC_G_PARM, &stream) < 0) { GST_ELEMENT_WARNING (v4l2src, RESOURCE, SETTINGS, (_("Could not get parameters on device '%s'"), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); goto done; } /* Note: V4L2 provides the frame interval, we have the frame rate */ if (fractions_are_equal (stream.parm.capture.timeperframe.numerator, stream.parm.capture.timeperframe.denominator, fps_d, fps_n)) { GST_LOG_OBJECT (v4l2src, "Desired framerate already set"); v4l2src->fps_n = fps_n; v4l2src->fps_d = fps_d; goto done; } /* We want to change the frame rate, so check whether we can. Some cheap USB * cameras don't have the capability */ if ((stream.parm.capture.capability & V4L2_CAP_TIMEPERFRAME) == 0) { GST_DEBUG_OBJECT (v4l2src, "Not setting framerate (not supported)"); goto done; } GST_LOG_OBJECT (v4l2src, "Setting framerate to %u/%u", fps_n, fps_d); /* Note: V4L2 wants the frame interval, we have the frame rate */ stream.parm.capture.timeperframe.numerator = fps_d; stream.parm.capture.timeperframe.denominator = fps_n; /* some cheap USB cam's won't accept any change */ if (v4l2_ioctl (fd, VIDIOC_S_PARM, &stream) < 0) { GST_ELEMENT_WARNING (v4l2src, RESOURCE, SETTINGS, (_("Video input device did not accept new frame rate setting.")), GST_ERROR_SYSTEM); goto done; } v4l2src->fps_n = fps_n; v4l2src->fps_d = fps_d; GST_INFO_OBJECT (v4l2src, "Set framerate to %u/%u", fps_n, fps_d); done: return TRUE; /* ERRORS */ get_fmt_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("Device '%s' does not support video capture"), v4l2src->v4l2object->videodev), ("Call to G_FMT failed: (%s)", g_strerror (errno))); return FALSE; } set_fmt_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("Device '%s' cannot capture at %dx%d"), v4l2src->v4l2object->videodev, width, height), ("Call to S_FMT failed for %" GST_FOURCC_FORMAT " @ %dx%d: %s", GST_FOURCC_ARGS (pixelformat), width, height, g_strerror (errno))); return FALSE; } invalid_dimensions: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("Device '%s' cannot capture at %dx%d"), v4l2src->v4l2object->videodev, width, height), ("Tried to capture at %dx%d, but device returned size %dx%d", width, height, format.fmt.pix.width, format.fmt.pix.height)); return FALSE; } invalid_pixelformat: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, SETTINGS, (_("Device '%s' cannot capture in the specified format"), v4l2src->v4l2object->videodev), ("Tried to capture in %" GST_FOURCC_FORMAT ", but device returned format" " %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (pixelformat), GST_FOURCC_ARGS (format.fmt.pix.pixelformat))); return FALSE; } } /****************************************************** * gst_v4l2src_capture_init(): * initialize the capture system * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_init (GstV4l2Src * v4l2src, GstCaps * caps) { gint fd = v4l2src->v4l2object->video_fd; struct v4l2_requestbuffers breq; memset (&breq, 0, sizeof (struct v4l2_requestbuffers)); GST_DEBUG_OBJECT (v4l2src, "initializing the capture system"); GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_NOT_ACTIVE (v4l2src->v4l2object); if (v4l2src->v4l2object->vcap.capabilities & V4L2_CAP_STREAMING) { GST_DEBUG_OBJECT (v4l2src, "STREAMING, requesting %d MMAP CAPTURE buffers", v4l2src->num_buffers); breq.count = v4l2src->num_buffers; breq.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; breq.memory = V4L2_MEMORY_MMAP; if (v4l2_ioctl (fd, VIDIOC_REQBUFS, &breq) < 0) goto reqbufs_failed; GST_LOG_OBJECT (v4l2src, " count: %u", breq.count); GST_LOG_OBJECT (v4l2src, " type: %d", breq.type); GST_LOG_OBJECT (v4l2src, " memory: %d", breq.memory); if (breq.count < GST_V4L2_MIN_BUFFERS) goto no_buffers; if (v4l2src->num_buffers != breq.count) { GST_WARNING_OBJECT (v4l2src, "using %u buffers instead", breq.count); v4l2src->num_buffers = breq.count; g_object_notify (G_OBJECT (v4l2src), "queue-size"); } /* Map the buffers */ GST_LOG_OBJECT (v4l2src, "initiating buffer pool"); if (!(v4l2src->pool = gst_v4l2_buffer_pool_new (v4l2src, fd, v4l2src->num_buffers, caps))) goto buffer_pool_new_failed; GST_INFO_OBJECT (v4l2src, "capturing buffers via mmap()"); v4l2src->use_mmap = TRUE; } else if (v4l2src->v4l2object->vcap.capabilities & V4L2_CAP_READWRITE) { GST_INFO_OBJECT (v4l2src, "capturing buffers via read()"); v4l2src->use_mmap = FALSE; v4l2src->pool = NULL; } else { goto no_supported_capture_method; } GST_V4L2_SET_ACTIVE (v4l2src->v4l2object); return TRUE; /* ERRORS */ reqbufs_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not get buffers from device '%s'."), v4l2src->v4l2object->videodev), ("error requesting %d buffers: %s", v4l2src->num_buffers, g_strerror (errno))); return FALSE; } no_buffers: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not get enough buffers from device '%s'."), v4l2src->v4l2object->videodev), ("we received %d from device '%s', we want at least %d", breq.count, v4l2src->v4l2object->videodev, GST_V4L2_MIN_BUFFERS)); return FALSE; } buffer_pool_new_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("Could not map buffers from device '%s'"), v4l2src->v4l2object->videodev), ("Failed to create buffer pool: %s", g_strerror (errno))); return FALSE; } no_supported_capture_method: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, READ, (_("The driver of device '%s' does not support any known capture " "method."), v4l2src->v4l2object->videodev), (NULL)); return FALSE; } } /****************************************************** * gst_v4l2src_capture_start(): * start streaming capture * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_start (GstV4l2Src * v4l2src) { gint type = V4L2_BUF_TYPE_VIDEO_CAPTURE; gint fd = v4l2src->v4l2object->video_fd; GST_DEBUG_OBJECT (v4l2src, "starting the capturing"); //GST_V4L2_CHECK_OPEN (v4l2src->v4l2object); GST_V4L2_CHECK_ACTIVE (v4l2src->v4l2object); v4l2src->quit = FALSE; if (v4l2src->use_mmap) { if (!gst_v4l2_buffer_pool_activate (v4l2src->pool, v4l2src)) goto pool_activate_failed; if (v4l2_ioctl (fd, VIDIOC_STREAMON, &type) < 0) goto streamon_failed; } v4l2src->is_capturing = TRUE; return TRUE; /* ERRORS */ pool_activate_failed: { /* already errored */ return FALSE; } streamon_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, OPEN_READ, (_("Error starting streaming capture from device '%s'."), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); return FALSE; } } /****************************************************** * gst_v4l2src_capture_stop(): * stop streaming capture * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_stop (GstV4l2Src * v4l2src) { gint type = V4L2_BUF_TYPE_VIDEO_CAPTURE; GST_DEBUG_OBJECT (v4l2src, "stopping capturing"); if (!GST_V4L2_IS_OPEN (v4l2src->v4l2object)) { goto done; } if (!GST_V4L2_IS_ACTIVE (v4l2src->v4l2object)) { goto done; } if (v4l2src->use_mmap) { /* we actually need to sync on all queued buffers but not * on the non-queued ones */ if (v4l2_ioctl (v4l2src->v4l2object->video_fd, VIDIOC_STREAMOFF, &type) < 0) goto streamoff_failed; } done: /* make an optional pending wait stop */ v4l2src->quit = TRUE; v4l2src->is_capturing = FALSE; return TRUE; /* ERRORS */ streamoff_failed: { GST_ELEMENT_ERROR (v4l2src, RESOURCE, CLOSE, (_("Error stopping streaming capture from device '%s'."), v4l2src->v4l2object->videodev), GST_ERROR_SYSTEM); return FALSE; } } /****************************************************** * gst_v4l2src_capture_deinit(): * deinitialize the capture system * return value: TRUE on success, FALSE on error ******************************************************/ gboolean gst_v4l2src_capture_deinit (GstV4l2Src * v4l2src) { GST_DEBUG_OBJECT (v4l2src, "deinitting capture system"); if (!GST_V4L2_IS_OPEN (v4l2src->v4l2object)) { return TRUE; } if (!GST_V4L2_IS_ACTIVE (v4l2src->v4l2object)) { return TRUE; } if (v4l2src->pool) { gst_v4l2_buffer_pool_destroy (v4l2src->pool); v4l2src->pool = NULL; } GST_V4L2_SET_INACTIVE (v4l2src->v4l2object); return TRUE; } /* */ static gboolean gst_v4l2src_get_nearest_size (GstV4l2Src * v4l2src, guint32 pixelformat, gint * width, gint * height) { struct v4l2_format fmt; int fd; int r; g_return_val_if_fail (width != NULL, FALSE); g_return_val_if_fail (height != NULL, FALSE); GST_LOG_OBJECT (v4l2src, "getting nearest size to %dx%d with format %" GST_FOURCC_FORMAT, *width, *height, GST_FOURCC_ARGS (pixelformat)); fd = v4l2src->v4l2object->video_fd; /* get size delimiters */ memset (&fmt, 0, sizeof (fmt)); fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; fmt.fmt.pix.width = *width; fmt.fmt.pix.height = *height; fmt.fmt.pix.pixelformat = pixelformat; fmt.fmt.pix.field = V4L2_FIELD_INTERLACED; r = v4l2_ioctl (fd, VIDIOC_TRY_FMT, &fmt); if (r < 0 && errno == EINVAL) { /* try again with progressive video */ fmt.fmt.pix.width = *width; fmt.fmt.pix.height = *height; fmt.fmt.pix.pixelformat = pixelformat; fmt.fmt.pix.field = V4L2_FIELD_NONE; r = v4l2_ioctl (fd, VIDIOC_TRY_FMT, &fmt); } if (r < 0) { /* The driver might not implement TRY_FMT, in which case we will try S_FMT to probe */ if (errno != ENOTTY) return FALSE; /* Only try S_FMT if we're not actively capturing yet, which we shouldn't be, because we're still probing */ if (GST_V4L2_IS_ACTIVE (v4l2src->v4l2object)) return FALSE; GST_LOG_OBJECT (v4l2src, "Failed to probe size limit with VIDIOC_TRY_FMT, trying VIDIOC_S_FMT"); fmt.fmt.pix.width = *width; fmt.fmt.pix.height = *height; r = v4l2_ioctl (fd, VIDIOC_S_FMT, &fmt); if (r < 0 && errno == EINVAL) { /* try again with progressive video */ fmt.fmt.pix.width = *width; fmt.fmt.pix.height = *height; fmt.fmt.pix.pixelformat = pixelformat; fmt.fmt.pix.field = V4L2_FIELD_NONE; r = v4l2_ioctl (fd, VIDIOC_S_FMT, &fmt); } if (r < 0) return FALSE; } GST_LOG_OBJECT (v4l2src, "got nearest size %dx%d", fmt.fmt.pix.width, fmt.fmt.pix.height); *width = fmt.fmt.pix.width; *height = fmt.fmt.pix.height; return TRUE; }