gstreamer/gst-libs/gst/video/video.c

2044 lines
58 KiB
C
Raw Normal View History

/* GStreamer
* Copyright (C) <1999> Erik Walthinsen <omega@cse.ogi.edu>
* Library <2002> Ronald Bultje <rbultje@ronald.bitfreak.net>
* Copyright (C) 2007 David A. Schleef <ds@schleef.org>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
* Boston, MA 02111-1307, USA.
*/
#ifdef HAVE_CONFIG_H
# include "config.h"
#endif
2011-06-15 14:16:02 +00:00
#include <string.h>
#include "video.h"
typedef enum
{
VIDEO_FLAG_YUV = (1 << 0),
VIDEO_FLAG_RGB = (1 << 1),
VIDEO_FLAG_GRAY = (1 << 2),
VIDEO_FLAG_ALPHA = (1 << 3)
} VideoFlags;
2011-06-15 14:16:02 +00:00
typedef struct
{
const gchar *fmt;
GstVideoFormat format;
guint32 fourcc;
VideoFlags flags;
guint depth[GST_VIDEO_MAX_PLANES];
2011-06-15 14:16:02 +00:00
} VideoFormat;
#define COMP0 { 0, 0, 0, 0 }
#define COMP8 { 8, 0, 0, 0 }
#define COMP888 { 8, 8, 8, 0 }
#define COMP8888 { 8, 8, 8, 8 }
#define COMP10_10_10 { 10, 10, 10, 0 }
#define COMP16 { 16, 0, 0, 0 }
#define COMP16_16_16 { 16, 0, 0, 0 }
#define COMP16_16_16_16 { 16, 0, 0, 0 }
#define COMP555 { 5, 5, 5, 0 }
#define COMP565 { 5, 6, 5, 0 }
#define MAKE_YUV_FORMAT(name, fourcc, comp) \
{ G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, fourcc, VIDEO_FLAG_YUV, comp }
#define MAKE_YUVA_FORMAT(name, fourcc, comp) \
{ G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, fourcc, VIDEO_FLAG_YUV | VIDEO_FLAG_ALPHA, comp }
#define MAKE_RGB_FORMAT(name, comp) \
{ G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, 0x00000000, VIDEO_FLAG_RGB, comp }
#define MAKE_RGBA_FORMAT(name, comp) \
{ G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, 0x00000000, VIDEO_FLAG_RGB | VIDEO_FLAG_ALPHA, comp }
#define MAKE_GRAY_FORMAT(name, comp) \
{ G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, 0x00000000, VIDEO_FLAG_GRAY, comp }
2011-06-15 14:16:02 +00:00
static VideoFormat formats[] = {
{"UNKNOWN", GST_VIDEO_FORMAT_UNKNOWN, 0x00000000, 0, COMP0},
MAKE_YUV_FORMAT (I420, GST_MAKE_FOURCC ('I', '4', '2', '0'), COMP888),
MAKE_YUV_FORMAT (YV12, GST_MAKE_FOURCC ('Y', 'V', '1', '2'), COMP888),
MAKE_YUV_FORMAT (YUY2, GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'), COMP888),
MAKE_YUV_FORMAT (UYVY, GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'), COMP888),
MAKE_YUVA_FORMAT (AYUV, GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'), COMP8888),
MAKE_RGB_FORMAT (RGBx, COMP888),
MAKE_RGB_FORMAT (BGRx, COMP888),
MAKE_RGB_FORMAT (xRGB, COMP888),
MAKE_RGB_FORMAT (xBGR, COMP888),
MAKE_RGBA_FORMAT (RGBA, COMP8888),
MAKE_RGBA_FORMAT (BGRA, COMP8888),
MAKE_RGBA_FORMAT (ARGB, COMP8888),
MAKE_RGBA_FORMAT (ABGR, COMP8888),
MAKE_RGB_FORMAT (RGB, COMP888),
MAKE_RGB_FORMAT (BGR, COMP888),
MAKE_YUV_FORMAT (Y41B, GST_MAKE_FOURCC ('Y', '4', '1', 'B'), COMP888),
MAKE_YUV_FORMAT (Y42B, GST_MAKE_FOURCC ('Y', '4', '2', 'B'), COMP888),
MAKE_YUV_FORMAT (YVYU, GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'), COMP888),
MAKE_YUV_FORMAT (Y444, GST_MAKE_FOURCC ('Y', '4', '4', '4'), COMP888),
MAKE_YUV_FORMAT (v210, GST_MAKE_FOURCC ('v', '2', '1', '0'), COMP10_10_10),
MAKE_YUV_FORMAT (v216, GST_MAKE_FOURCC ('v', '2', '1', '6'), COMP16_16_16),
MAKE_YUV_FORMAT (NV12, GST_MAKE_FOURCC ('N', 'V', '1', '2'), COMP888),
MAKE_YUV_FORMAT (NV21, GST_MAKE_FOURCC ('N', 'V', '2', '1'), COMP888),
MAKE_GRAY_FORMAT (GRAY8, COMP8),
MAKE_GRAY_FORMAT (GRAY16_BE, COMP16),
MAKE_GRAY_FORMAT (GRAY16_LE, COMP16),
MAKE_YUV_FORMAT (v308, GST_MAKE_FOURCC ('v', '3', '0', '8'), COMP888),
MAKE_YUV_FORMAT (Y800, GST_MAKE_FOURCC ('Y', '8', '0', '0'), COMP8),
MAKE_YUV_FORMAT (Y16, GST_MAKE_FOURCC ('Y', '1', '6', ' '), COMP16),
MAKE_RGB_FORMAT (RGB16, COMP565),
MAKE_RGB_FORMAT (BGR16, COMP565),
MAKE_RGB_FORMAT (RGB15, COMP555),
MAKE_RGB_FORMAT (BGR15, COMP555),
MAKE_YUV_FORMAT (UYVP, GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'), COMP10_10_10),
MAKE_YUVA_FORMAT (A420, GST_MAKE_FOURCC ('A', '4', '2', '0'), COMP888),
MAKE_RGBA_FORMAT (RGB8_PALETTED, COMP8888),
MAKE_YUV_FORMAT (YUV9, GST_MAKE_FOURCC ('Y', 'U', 'V', '9'), COMP888),
MAKE_YUV_FORMAT (YVU9, GST_MAKE_FOURCC ('Y', 'V', 'U', '9'), COMP888),
MAKE_YUV_FORMAT (IYU1, GST_MAKE_FOURCC ('I', 'Y', 'U', '1'), COMP888),
MAKE_RGBA_FORMAT (ARGB64, COMP16_16_16_16),
MAKE_YUVA_FORMAT (AYUV64, 0x00000000, COMP16_16_16_16),
MAKE_YUV_FORMAT (r210, GST_MAKE_FOURCC ('r', '2', '1', '0'), COMP10_10_10),
2011-06-15 14:16:02 +00:00
};
/**
* SECTION:gstvideo
* @short_description: Support library for video operations
*
* <refsect2>
* <para>
2011-05-23 10:53:06 +00:00
* This library contains some helper functions and includes the
* videosink and videofilter base classes.
* </para>
* </refsect2>
*/
/**
* gst_video_frame_rate:
* @pad: pointer to a #GstPad
*
* A convenience function to retrieve a GValue holding the framerate
* from the caps on a pad.
2011-05-23 10:53:06 +00:00
*
* The pad needs to have negotiated caps containing a framerate property.
*
* Returns: NULL if the pad has no configured caps or the configured caps
* do not contain a framerate.
*
*/
Convert elements to use fractions for their framerate. Original commit message from CVS: * ext/libvisual/visual.c: (gst_visual_src_setcaps), (get_buffer), (gst_visual_chain): * ext/ogg/gstogmparse.c: (gst_ogm_parse_chain): * ext/theora/theoradec.c: (theora_handle_type_packet): * ext/theora/theoraenc.c: (theora_enc_sink_setcaps), (theora_enc_chain): * gst-libs/gst/riff/riff-media.c: (gst_riff_create_video_caps): * gst-libs/gst/video/video.c: (gst_video_frame_rate): * gst-libs/gst/video/video.h: * gst/ffmpegcolorspace/avcodec.h: * gst/ffmpegcolorspace/gstffmpegcodecmap.c: (gst_ffmpeg_caps_to_pixfmt): * gst/ffmpegcolorspace/gstffmpegcolorspace.c: (gst_ffmpegcsp_set_caps): * gst/videorate/gstvideorate.c: (gst_videorate_transformcaps), (gst_videorate_setcaps), (gst_videorate_blank_data), (gst_videorate_chain): * gst/videotestsrc/gstvideotestsrc.c: (gst_videotestsrc_src_fixate), (gst_videotestsrc_getcaps), (gst_videotestsrc_parse_caps), (gst_videotestsrc_setcaps), (gst_videotestsrc_event), (gst_videotestsrc_create): * gst/videotestsrc/gstvideotestsrc.h: * sys/ximage/ximagesink.c: (gst_ximagesink_xcontext_get), (gst_ximagesink_setcaps), (gst_ximagesink_change_state), (gst_ximagesink_get_times), (gst_ximagesink_init): * sys/ximage/ximagesink.h: * sys/xvimage/xvimagesink.c: (gst_xvimagesink_get_xv_support), (gst_xvimagesink_setcaps), (gst_xvimagesink_change_state), (gst_xvimagesink_get_times), (gst_xvimagesink_init): * sys/xvimage/xvimagesink.h: Convert elements to use fractions for their framerate. V4L elements to come later tonight.
2005-11-22 16:08:37 +00:00
const GValue *
gst_video_frame_rate (GstPad * pad)
{
Convert elements to use fractions for their framerate. Original commit message from CVS: * ext/libvisual/visual.c: (gst_visual_src_setcaps), (get_buffer), (gst_visual_chain): * ext/ogg/gstogmparse.c: (gst_ogm_parse_chain): * ext/theora/theoradec.c: (theora_handle_type_packet): * ext/theora/theoraenc.c: (theora_enc_sink_setcaps), (theora_enc_chain): * gst-libs/gst/riff/riff-media.c: (gst_riff_create_video_caps): * gst-libs/gst/video/video.c: (gst_video_frame_rate): * gst-libs/gst/video/video.h: * gst/ffmpegcolorspace/avcodec.h: * gst/ffmpegcolorspace/gstffmpegcodecmap.c: (gst_ffmpeg_caps_to_pixfmt): * gst/ffmpegcolorspace/gstffmpegcolorspace.c: (gst_ffmpegcsp_set_caps): * gst/videorate/gstvideorate.c: (gst_videorate_transformcaps), (gst_videorate_setcaps), (gst_videorate_blank_data), (gst_videorate_chain): * gst/videotestsrc/gstvideotestsrc.c: (gst_videotestsrc_src_fixate), (gst_videotestsrc_getcaps), (gst_videotestsrc_parse_caps), (gst_videotestsrc_setcaps), (gst_videotestsrc_event), (gst_videotestsrc_create): * gst/videotestsrc/gstvideotestsrc.h: * sys/ximage/ximagesink.c: (gst_ximagesink_xcontext_get), (gst_ximagesink_setcaps), (gst_ximagesink_change_state), (gst_ximagesink_get_times), (gst_ximagesink_init): * sys/ximage/ximagesink.h: * sys/xvimage/xvimagesink.c: (gst_xvimagesink_get_xv_support), (gst_xvimagesink_setcaps), (gst_xvimagesink_change_state), (gst_xvimagesink_get_times), (gst_xvimagesink_init): * sys/xvimage/xvimagesink.h: Convert elements to use fractions for their framerate. V4L elements to come later tonight.
2005-11-22 16:08:37 +00:00
const GValue *fps;
gchar *fps_string;
GstCaps *caps = NULL;
GstStructure *structure;
/* get pad caps */
caps = gst_pad_get_current_caps (pad);
if (caps == NULL)
goto no_caps;
structure = gst_caps_get_structure (caps, 0);
if ((fps = gst_structure_get_value (structure, "framerate")) == NULL)
goto no_framerate;
if (!GST_VALUE_HOLDS_FRACTION (fps))
goto no_fraction;
fps_string = gst_value_serialize (fps);
GST_DEBUG ("Framerate request on pad %s:%s: %s",
GST_DEBUG_PAD_NAME (pad), fps_string);
g_free (fps_string);
gst_caps_unref (caps);
return fps;
/* ERRORS */
no_caps:
{
g_warning ("gstvideo: failed to get caps of pad %s:%s",
GST_DEBUG_PAD_NAME (pad));
Convert elements to use fractions for their framerate. Original commit message from CVS: * ext/libvisual/visual.c: (gst_visual_src_setcaps), (get_buffer), (gst_visual_chain): * ext/ogg/gstogmparse.c: (gst_ogm_parse_chain): * ext/theora/theoradec.c: (theora_handle_type_packet): * ext/theora/theoraenc.c: (theora_enc_sink_setcaps), (theora_enc_chain): * gst-libs/gst/riff/riff-media.c: (gst_riff_create_video_caps): * gst-libs/gst/video/video.c: (gst_video_frame_rate): * gst-libs/gst/video/video.h: * gst/ffmpegcolorspace/avcodec.h: * gst/ffmpegcolorspace/gstffmpegcodecmap.c: (gst_ffmpeg_caps_to_pixfmt): * gst/ffmpegcolorspace/gstffmpegcolorspace.c: (gst_ffmpegcsp_set_caps): * gst/videorate/gstvideorate.c: (gst_videorate_transformcaps), (gst_videorate_setcaps), (gst_videorate_blank_data), (gst_videorate_chain): * gst/videotestsrc/gstvideotestsrc.c: (gst_videotestsrc_src_fixate), (gst_videotestsrc_getcaps), (gst_videotestsrc_parse_caps), (gst_videotestsrc_setcaps), (gst_videotestsrc_event), (gst_videotestsrc_create): * gst/videotestsrc/gstvideotestsrc.h: * sys/ximage/ximagesink.c: (gst_ximagesink_xcontext_get), (gst_ximagesink_setcaps), (gst_ximagesink_change_state), (gst_ximagesink_get_times), (gst_ximagesink_init): * sys/ximage/ximagesink.h: * sys/xvimage/xvimagesink.c: (gst_xvimagesink_get_xv_support), (gst_xvimagesink_setcaps), (gst_xvimagesink_change_state), (gst_xvimagesink_get_times), (gst_xvimagesink_init): * sys/xvimage/xvimagesink.h: Convert elements to use fractions for their framerate. V4L elements to come later tonight.
2005-11-22 16:08:37 +00:00
return NULL;
}
no_framerate:
{
g_warning ("gstvideo: failed to get framerate property of pad %s:%s",
GST_DEBUG_PAD_NAME (pad));
gst_caps_unref (caps);
Convert elements to use fractions for their framerate. Original commit message from CVS: * ext/libvisual/visual.c: (gst_visual_src_setcaps), (get_buffer), (gst_visual_chain): * ext/ogg/gstogmparse.c: (gst_ogm_parse_chain): * ext/theora/theoradec.c: (theora_handle_type_packet): * ext/theora/theoraenc.c: (theora_enc_sink_setcaps), (theora_enc_chain): * gst-libs/gst/riff/riff-media.c: (gst_riff_create_video_caps): * gst-libs/gst/video/video.c: (gst_video_frame_rate): * gst-libs/gst/video/video.h: * gst/ffmpegcolorspace/avcodec.h: * gst/ffmpegcolorspace/gstffmpegcodecmap.c: (gst_ffmpeg_caps_to_pixfmt): * gst/ffmpegcolorspace/gstffmpegcolorspace.c: (gst_ffmpegcsp_set_caps): * gst/videorate/gstvideorate.c: (gst_videorate_transformcaps), (gst_videorate_setcaps), (gst_videorate_blank_data), (gst_videorate_chain): * gst/videotestsrc/gstvideotestsrc.c: (gst_videotestsrc_src_fixate), (gst_videotestsrc_getcaps), (gst_videotestsrc_parse_caps), (gst_videotestsrc_setcaps), (gst_videotestsrc_event), (gst_videotestsrc_create): * gst/videotestsrc/gstvideotestsrc.h: * sys/ximage/ximagesink.c: (gst_ximagesink_xcontext_get), (gst_ximagesink_setcaps), (gst_ximagesink_change_state), (gst_ximagesink_get_times), (gst_ximagesink_init): * sys/ximage/ximagesink.h: * sys/xvimage/xvimagesink.c: (gst_xvimagesink_get_xv_support), (gst_xvimagesink_setcaps), (gst_xvimagesink_change_state), (gst_xvimagesink_get_times), (gst_xvimagesink_init): * sys/xvimage/xvimagesink.h: Convert elements to use fractions for their framerate. V4L elements to come later tonight.
2005-11-22 16:08:37 +00:00
return NULL;
}
no_fraction:
{
Convert elements to use fractions for their framerate. Original commit message from CVS: * ext/libvisual/visual.c: (gst_visual_src_setcaps), (get_buffer), (gst_visual_chain): * ext/ogg/gstogmparse.c: (gst_ogm_parse_chain): * ext/theora/theoradec.c: (theora_handle_type_packet): * ext/theora/theoraenc.c: (theora_enc_sink_setcaps), (theora_enc_chain): * gst-libs/gst/riff/riff-media.c: (gst_riff_create_video_caps): * gst-libs/gst/video/video.c: (gst_video_frame_rate): * gst-libs/gst/video/video.h: * gst/ffmpegcolorspace/avcodec.h: * gst/ffmpegcolorspace/gstffmpegcodecmap.c: (gst_ffmpeg_caps_to_pixfmt): * gst/ffmpegcolorspace/gstffmpegcolorspace.c: (gst_ffmpegcsp_set_caps): * gst/videorate/gstvideorate.c: (gst_videorate_transformcaps), (gst_videorate_setcaps), (gst_videorate_blank_data), (gst_videorate_chain): * gst/videotestsrc/gstvideotestsrc.c: (gst_videotestsrc_src_fixate), (gst_videotestsrc_getcaps), (gst_videotestsrc_parse_caps), (gst_videotestsrc_setcaps), (gst_videotestsrc_event), (gst_videotestsrc_create): * gst/videotestsrc/gstvideotestsrc.h: * sys/ximage/ximagesink.c: (gst_ximagesink_xcontext_get), (gst_ximagesink_setcaps), (gst_ximagesink_change_state), (gst_ximagesink_get_times), (gst_ximagesink_init): * sys/ximage/ximagesink.h: * sys/xvimage/xvimagesink.c: (gst_xvimagesink_get_xv_support), (gst_xvimagesink_setcaps), (gst_xvimagesink_change_state), (gst_xvimagesink_get_times), (gst_xvimagesink_init): * sys/xvimage/xvimagesink.h: Convert elements to use fractions for their framerate. V4L elements to come later tonight.
2005-11-22 16:08:37 +00:00
g_warning
("gstvideo: framerate property of pad %s:%s is not of type Fraction",
GST_DEBUG_PAD_NAME (pad));
gst_caps_unref (caps);
Convert elements to use fractions for their framerate. Original commit message from CVS: * ext/libvisual/visual.c: (gst_visual_src_setcaps), (get_buffer), (gst_visual_chain): * ext/ogg/gstogmparse.c: (gst_ogm_parse_chain): * ext/theora/theoradec.c: (theora_handle_type_packet): * ext/theora/theoraenc.c: (theora_enc_sink_setcaps), (theora_enc_chain): * gst-libs/gst/riff/riff-media.c: (gst_riff_create_video_caps): * gst-libs/gst/video/video.c: (gst_video_frame_rate): * gst-libs/gst/video/video.h: * gst/ffmpegcolorspace/avcodec.h: * gst/ffmpegcolorspace/gstffmpegcodecmap.c: (gst_ffmpeg_caps_to_pixfmt): * gst/ffmpegcolorspace/gstffmpegcolorspace.c: (gst_ffmpegcsp_set_caps): * gst/videorate/gstvideorate.c: (gst_videorate_transformcaps), (gst_videorate_setcaps), (gst_videorate_blank_data), (gst_videorate_chain): * gst/videotestsrc/gstvideotestsrc.c: (gst_videotestsrc_src_fixate), (gst_videotestsrc_getcaps), (gst_videotestsrc_parse_caps), (gst_videotestsrc_setcaps), (gst_videotestsrc_event), (gst_videotestsrc_create): * gst/videotestsrc/gstvideotestsrc.h: * sys/ximage/ximagesink.c: (gst_ximagesink_xcontext_get), (gst_ximagesink_setcaps), (gst_ximagesink_change_state), (gst_ximagesink_get_times), (gst_ximagesink_init): * sys/ximage/ximagesink.h: * sys/xvimage/xvimagesink.c: (gst_xvimagesink_get_xv_support), (gst_xvimagesink_setcaps), (gst_xvimagesink_change_state), (gst_xvimagesink_get_times), (gst_xvimagesink_init): * sys/xvimage/xvimagesink.h: Convert elements to use fractions for their framerate. V4L elements to come later tonight.
2005-11-22 16:08:37 +00:00
return NULL;
}
}
/**
* gst_video_calculate_display_ratio:
* @dar_n: Numerator of the calculated display_ratio
* @dar_d: Denominator of the calculated display_ratio
* @video_width: Width of the video frame in pixels
* @video_height: Height of the video frame in pixels
* @video_par_n: Numerator of the pixel aspect ratio of the input video.
* @video_par_d: Denominator of the pixel aspect ratio of the input video.
* @display_par_n: Numerator of the pixel aspect ratio of the display device
* @display_par_d: Denominator of the pixel aspect ratio of the display device
*
2011-05-23 10:53:06 +00:00
* Given the Pixel Aspect Ratio and size of an input video frame, and the
* pixel aspect ratio of the intended display device, calculates the actual
* display ratio the video will be rendered with.
*
2011-05-23 10:53:06 +00:00
* Returns: A boolean indicating success and a calculated Display Ratio in the
* dar_n and dar_d parameters.
* The return value is FALSE in the case of integer overflow or other error.
*
* Since: 0.10.7
*/
gboolean
gst_video_calculate_display_ratio (guint * dar_n, guint * dar_d,
guint video_width, guint video_height,
guint video_par_n, guint video_par_d,
guint display_par_n, guint display_par_d)
{
gint num, den;
gint tmp_n, tmp_d;
g_return_val_if_fail (dar_n != NULL, FALSE);
g_return_val_if_fail (dar_d != NULL, FALSE);
/* Calculate (video_width * video_par_n * display_par_d) /
* (video_height * video_par_d * display_par_n) */
if (!gst_util_fraction_multiply (video_width, video_height, video_par_n,
video_par_d, &tmp_n, &tmp_d))
goto error_overflow;
if (!gst_util_fraction_multiply (tmp_n, tmp_d, display_par_d, display_par_n,
&num, &den))
goto error_overflow;
g_return_val_if_fail (num > 0, FALSE);
g_return_val_if_fail (den > 0, FALSE);
*dar_n = num;
*dar_d = den;
return TRUE;
/* ERRORS */
error_overflow:
{
GST_WARNING ("overflow in multiply");
return FALSE;
}
}
/**
* gst_video_format_parse_caps_interlaced:
* @caps: the fixed #GstCaps to parse
* @interlaced: whether @caps represents interlaced video or not, may be NULL (output)
*
* Extracts whether the caps represents interlaced content or not and places it
* in @interlaced.
*
2009-02-19 16:40:45 +00:00
* Since: 0.10.23
*
* Returns: TRUE if @caps was parsed correctly.
*/
gboolean
gst_video_format_parse_caps_interlaced (GstCaps * caps, gboolean * interlaced)
{
GstStructure *structure;
if (!gst_caps_is_fixed (caps))
return FALSE;
structure = gst_caps_get_structure (caps, 0);
if (interlaced) {
if (!gst_structure_get_boolean (structure, "interlaced", interlaced))
*interlaced = FALSE;
}
return TRUE;
}
/**
* gst_video_parse_caps_color_matrix:
* @caps: the fixed #GstCaps to parse
*
* Extracts the color matrix used by the caps. Possible values are
* "sdtv" for the standard definition color matrix (as specified in
* Rec. ITU-R BT.470-6) or "hdtv" for the high definition color
* matrix (as specified in Rec. ITU-R BT.709)
*
* Since: 0.10.29
*
* Returns: a color matrix string, or NULL if no color matrix could be
* determined.
*/
const char *
gst_video_parse_caps_color_matrix (GstCaps * caps)
{
GstStructure *structure;
const char *s;
if (!gst_caps_is_fixed (caps))
return NULL;
structure = gst_caps_get_structure (caps, 0);
s = gst_structure_get_string (structure, "color-matrix");
if (s)
return s;
if (gst_structure_has_name (structure, "video/x-raw-yuv")) {
return "sdtv";
}
return NULL;
}
/**
* gst_video_parse_caps_chroma_site:
* @caps: the fixed #GstCaps to parse
*
* Extracts the chroma site used by the caps. Possible values are
* "mpeg2" for MPEG-2 style chroma siting (co-sited horizontally,
* halfway-sited vertically), "jpeg" for JPEG and Theora style
* chroma siting (halfway-sited both horizontally and vertically).
* Other chroma site values are possible, but uncommon.
2011-05-23 10:53:06 +00:00
*
* When no chroma site is specified in the caps, it should be assumed
* to be "mpeg2".
*
* Since: 0.10.29
*
* Returns: a chroma site string, or NULL if no chroma site could be
* determined.
*/
const char *
gst_video_parse_caps_chroma_site (GstCaps * caps)
{
GstStructure *structure;
const char *s;
if (!gst_caps_is_fixed (caps))
return NULL;
structure = gst_caps_get_structure (caps, 0);
s = gst_structure_get_string (structure, "chroma-site");
if (s)
return s;
if (gst_structure_has_name (structure, "video/x-raw-yuv")) {
return "mpeg2";
}
return NULL;
}
/**
* gst_video_parse_caps_framerate:
* @caps: pointer to a #GstCaps instance
* @fps_n: pointer to integer to hold numerator of frame rate (output)
* @fps_d: pointer to integer to hold denominator of frame rate (output)
*
* Extracts the frame rate from @caps and places the values in the locations
* pointed to by @fps_n and @fps_d. Returns TRUE if the values could be
* parsed correctly, FALSE if not.
*
* This function can be used with #GstCaps that have any media type; it
* is not limited to formats handled by #GstVideoFormat.
*
* Since: 0.10.16
*
* Returns: TRUE if @caps was parsed correctly.
*/
gboolean
gst_video_parse_caps_framerate (GstCaps * caps, int *fps_n, int *fps_d)
{
GstStructure *structure;
if (!gst_caps_is_fixed (caps))
return FALSE;
structure = gst_caps_get_structure (caps, 0);
return gst_structure_get_fraction (structure, "framerate", fps_n, fps_d);
}
/**
* gst_video_parse_caps_pixel_aspect_ratio:
* @caps: pointer to a #GstCaps instance
* @par_n: pointer to numerator of pixel aspect ratio (output)
* @par_d: pointer to denominator of pixel aspect ratio (output)
*
* Extracts the pixel aspect ratio from @caps and places the values in
* the locations pointed to by @par_n and @par_d. Returns TRUE if the
* values could be parsed correctly, FALSE if not.
*
* This function can be used with #GstCaps that have any media type; it
* is not limited to formats handled by #GstVideoFormat.
*
* Since: 0.10.16
*
* Returns: TRUE if @caps was parsed correctly.
*/
gboolean
gst_video_parse_caps_pixel_aspect_ratio (GstCaps * caps, int *par_n, int *par_d)
{
GstStructure *structure;
if (!gst_caps_is_fixed (caps))
return FALSE;
structure = gst_caps_get_structure (caps, 0);
if (!gst_structure_get_fraction (structure, "pixel-aspect-ratio",
par_n, par_d)) {
*par_n = 1;
*par_d = 1;
}
return TRUE;
}
/**
* gst_video_format_new_caps_interlaced:
* @format: the #GstVideoFormat describing the raw video format
* @width: width of video
* @height: height of video
* @framerate_n: numerator of frame rate
* @framerate_d: denominator of frame rate
* @par_n: numerator of pixel aspect ratio
* @par_d: denominator of pixel aspect ratio
* @interlaced: #TRUE if the format is interlaced
*
* Creates a new #GstCaps object based on the parameters provided.
*
2009-02-19 16:40:45 +00:00
* Since: 0.10.23
*
* Returns: a new #GstCaps object, or NULL if there was an error
*/
GstCaps *
gst_video_format_new_caps_interlaced (GstVideoFormat format,
int width, int height, int framerate_n, int framerate_d, int par_n,
int par_d, gboolean interlaced)
{
GstCaps *res;
res =
gst_video_format_new_caps (format, width, height, framerate_n,
framerate_d, par_n, par_d);
if (interlaced && (res != NULL))
gst_caps_set_simple (res, "interlaced", G_TYPE_BOOLEAN, TRUE, NULL);
return res;
}
static GstCaps *
gst_video_format_new_caps_raw (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
2011-06-15 14:16:02 +00:00
return gst_caps_new_simple ("video/x-raw",
"format", G_TYPE_STRING, gst_video_format_to_string (format), NULL);
}
/**
* gst_video_format_new_template_caps:
* @format: the #GstVideoFormat describing the raw video format
*
* Creates a new #GstCaps object based on the parameters provided.
* Size, frame rate, and pixel aspect ratio are set to the full
* range.
*
* Since: 0.10.33
*
* Returns: a new #GstCaps object, or NULL if there was an error
*/
GstCaps *
gst_video_format_new_template_caps (GstVideoFormat format)
{
GstCaps *caps;
GstStructure *structure;
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
caps = gst_video_format_new_caps_raw (format);
if (caps) {
GValue value = { 0 };
GValue v = { 0 };
structure = gst_caps_get_structure (caps, 0);
gst_structure_set (structure,
"width", GST_TYPE_INT_RANGE, 1, G_MAXINT,
"height", GST_TYPE_INT_RANGE, 1, G_MAXINT,
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1,
"pixel-aspect-ratio", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL);
g_value_init (&value, GST_TYPE_LIST);
g_value_init (&v, G_TYPE_BOOLEAN);
g_value_set_boolean (&v, TRUE);
gst_value_list_append_value (&value, &v);
g_value_set_boolean (&v, FALSE);
gst_value_list_append_value (&value, &v);
gst_structure_set_value (structure, "interlaced", &value);
g_value_reset (&value);
g_value_reset (&v);
}
return caps;
}
/**
* gst_video_format_new_caps:
* @format: the #GstVideoFormat describing the raw video format
* @width: width of video
* @height: height of video
* @framerate_n: numerator of frame rate
* @framerate_d: denominator of frame rate
* @par_n: numerator of pixel aspect ratio
* @par_d: denominator of pixel aspect ratio
*
* Creates a new #GstCaps object based on the parameters provided.
*
* Since: 0.10.16
*
* Returns: a new #GstCaps object, or NULL if there was an error
*/
GstCaps *
gst_video_format_new_caps (GstVideoFormat format, int width,
int height, int framerate_n, int framerate_d, int par_n, int par_d)
{
GstCaps *caps;
GstStructure *structure;
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
g_return_val_if_fail (width > 0 && height > 0, NULL);
caps = gst_video_format_new_caps_raw (format);
if (caps) {
structure = gst_caps_get_structure (caps, 0);
gst_structure_set (structure,
"width", G_TYPE_INT, width,
"height", G_TYPE_INT, height,
"framerate", GST_TYPE_FRACTION, framerate_n, framerate_d,
"pixel-aspect-ratio", GST_TYPE_FRACTION, par_n, par_d, NULL);
}
return caps;
}
static GstVideoFormat
gst_video_format_from_rgb32_masks (int red_mask, int green_mask, int blue_mask)
{
if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
blue_mask == 0x0000ff00) {
return GST_VIDEO_FORMAT_RGBx;
}
if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
blue_mask == 0xff000000) {
return GST_VIDEO_FORMAT_BGRx;
}
if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
blue_mask == 0x000000ff) {
return GST_VIDEO_FORMAT_xRGB;
}
if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
blue_mask == 0x00ff0000) {
return GST_VIDEO_FORMAT_xBGR;
}
return GST_VIDEO_FORMAT_UNKNOWN;
}
static GstVideoFormat
gst_video_format_from_rgba32_masks (int red_mask, int green_mask,
int blue_mask, int alpha_mask)
{
if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
blue_mask == 0x0000ff00 && alpha_mask == 0x000000ff) {
return GST_VIDEO_FORMAT_RGBA;
}
if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
blue_mask == 0xff000000 && alpha_mask == 0x000000ff) {
return GST_VIDEO_FORMAT_BGRA;
}
if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
blue_mask == 0x000000ff && alpha_mask == 0xff000000) {
return GST_VIDEO_FORMAT_ARGB;
}
if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
blue_mask == 0x00ff0000 && alpha_mask == 0xff000000) {
return GST_VIDEO_FORMAT_ABGR;
}
return GST_VIDEO_FORMAT_UNKNOWN;
}
static GstVideoFormat
gst_video_format_from_rgb24_masks (int red_mask, int green_mask, int blue_mask)
{
if (red_mask == 0xff0000 && green_mask == 0x00ff00 && blue_mask == 0x0000ff) {
return GST_VIDEO_FORMAT_RGB;
}
if (red_mask == 0x0000ff && green_mask == 0x00ff00 && blue_mask == 0xff0000) {
return GST_VIDEO_FORMAT_BGR;
}
return GST_VIDEO_FORMAT_UNKNOWN;
}
#define GST_VIDEO_COMP1_MASK_16_INT 0xf800
#define GST_VIDEO_COMP2_MASK_16_INT 0x07e0
#define GST_VIDEO_COMP3_MASK_16_INT 0x001f
#define GST_VIDEO_COMP1_MASK_15_INT 0x7c00
#define GST_VIDEO_COMP2_MASK_15_INT 0x03e0
#define GST_VIDEO_COMP3_MASK_15_INT 0x001f
static GstVideoFormat
gst_video_format_from_rgb16_masks (int red_mask, int green_mask, int blue_mask)
{
if (red_mask == GST_VIDEO_COMP1_MASK_16_INT
&& green_mask == GST_VIDEO_COMP2_MASK_16_INT
&& blue_mask == GST_VIDEO_COMP3_MASK_16_INT) {
return GST_VIDEO_FORMAT_RGB16;
}
if (red_mask == GST_VIDEO_COMP3_MASK_16_INT
&& green_mask == GST_VIDEO_COMP2_MASK_16_INT
&& blue_mask == GST_VIDEO_COMP1_MASK_16_INT) {
return GST_VIDEO_FORMAT_BGR16;
}
if (red_mask == GST_VIDEO_COMP1_MASK_15_INT
&& green_mask == GST_VIDEO_COMP2_MASK_15_INT
&& blue_mask == GST_VIDEO_COMP3_MASK_15_INT) {
return GST_VIDEO_FORMAT_RGB15;
}
if (red_mask == GST_VIDEO_COMP3_MASK_15_INT
&& green_mask == GST_VIDEO_COMP2_MASK_15_INT
&& blue_mask == GST_VIDEO_COMP1_MASK_15_INT) {
return GST_VIDEO_FORMAT_BGR15;
}
return GST_VIDEO_FORMAT_UNKNOWN;
}
GstVideoFormat
gst_video_format_from_masks (gint depth, gint bpp, gint endianness,
gint red_mask, gint green_mask, gint blue_mask, gint alpha_mask)
{
GstVideoFormat format;
/* our caps system handles 24/32bpp RGB as big-endian. */
if ((bpp == 24 || bpp == 32) && endianness == G_LITTLE_ENDIAN) {
red_mask = GUINT32_TO_BE (red_mask);
green_mask = GUINT32_TO_BE (green_mask);
blue_mask = GUINT32_TO_BE (blue_mask);
endianness = G_BIG_ENDIAN;
if (bpp == 24) {
red_mask >>= 8;
green_mask >>= 8;
blue_mask >>= 8;
}
}
if (depth == 30 && bpp == 32) {
format = GST_VIDEO_FORMAT_r210;
} else if (depth == 24 && bpp == 32) {
format = gst_video_format_from_rgb32_masks (red_mask, green_mask,
blue_mask);
} else if (depth == 32 && bpp == 32 && alpha_mask) {
format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
blue_mask, alpha_mask);
} else if (depth == 24 && bpp == 24) {
format = gst_video_format_from_rgb24_masks (red_mask, green_mask,
blue_mask);
} else if ((depth == 15 || depth == 16) && bpp == 16 &&
endianness == G_BYTE_ORDER) {
format = gst_video_format_from_rgb16_masks (red_mask, green_mask,
blue_mask);
} else if (depth == 8 && bpp == 8) {
format = GST_VIDEO_FORMAT_RGB8_PALETTED;
} else if (depth == 64 && bpp == 64) {
format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
blue_mask, alpha_mask);
if (format == GST_VIDEO_FORMAT_ARGB) {
format = GST_VIDEO_FORMAT_ARGB64;
} else {
format = GST_VIDEO_FORMAT_UNKNOWN;
}
} else {
format = GST_VIDEO_FORMAT_UNKNOWN;
}
return format;
}
/**
* gst_video_format_from_fourcc:
* @fourcc: a FOURCC value representing raw YUV video
*
* Converts a FOURCC value into the corresponding #GstVideoFormat.
* If the FOURCC cannot be represented by #GstVideoFormat,
* #GST_VIDEO_FORMAT_UNKNOWN is returned.
*
* Since: 0.10.16
*
* Returns: the #GstVideoFormat describing the FOURCC value
*/
GstVideoFormat
gst_video_format_from_fourcc (guint32 fourcc)
{
switch (fourcc) {
case GST_MAKE_FOURCC ('I', '4', '2', '0'):
return GST_VIDEO_FORMAT_I420;
case GST_MAKE_FOURCC ('Y', 'V', '1', '2'):
return GST_VIDEO_FORMAT_YV12;
case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'):
return GST_VIDEO_FORMAT_YUY2;
case GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'):
return GST_VIDEO_FORMAT_YVYU;
case GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'):
return GST_VIDEO_FORMAT_UYVY;
case GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'):
return GST_VIDEO_FORMAT_AYUV;
case GST_MAKE_FOURCC ('Y', '4', '1', 'B'):
return GST_VIDEO_FORMAT_Y41B;
case GST_MAKE_FOURCC ('Y', '4', '2', 'B'):
return GST_VIDEO_FORMAT_Y42B;
2009-05-15 23:18:18 +00:00
case GST_MAKE_FOURCC ('Y', '4', '4', '4'):
return GST_VIDEO_FORMAT_Y444;
case GST_MAKE_FOURCC ('v', '2', '1', '0'):
return GST_VIDEO_FORMAT_v210;
case GST_MAKE_FOURCC ('v', '2', '1', '6'):
return GST_VIDEO_FORMAT_v216;
2009-09-23 08:25:02 +00:00
case GST_MAKE_FOURCC ('N', 'V', '1', '2'):
return GST_VIDEO_FORMAT_NV12;
case GST_MAKE_FOURCC ('N', 'V', '2', '1'):
return GST_VIDEO_FORMAT_NV21;
case GST_MAKE_FOURCC ('v', '3', '0', '8'):
return GST_VIDEO_FORMAT_v308;
case GST_MAKE_FOURCC ('Y', '8', '0', '0'):
case GST_MAKE_FOURCC ('Y', '8', ' ', ' '):
case GST_MAKE_FOURCC ('G', 'R', 'E', 'Y'):
return GST_VIDEO_FORMAT_Y800;
case GST_MAKE_FOURCC ('Y', '1', '6', ' '):
return GST_VIDEO_FORMAT_Y16;
2010-09-13 02:34:28 +00:00
case GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'):
return GST_VIDEO_FORMAT_UYVP;
case GST_MAKE_FOURCC ('A', '4', '2', '0'):
return GST_VIDEO_FORMAT_A420;
case GST_MAKE_FOURCC ('Y', 'U', 'V', '9'):
return GST_VIDEO_FORMAT_YUV9;
case GST_MAKE_FOURCC ('Y', 'V', 'U', '9'):
return GST_VIDEO_FORMAT_YVU9;
case GST_MAKE_FOURCC ('I', 'Y', 'U', '1'):
return GST_VIDEO_FORMAT_IYU1;
case GST_MAKE_FOURCC ('A', 'Y', '6', '4'):
return GST_VIDEO_FORMAT_AYUV64;
default:
return GST_VIDEO_FORMAT_UNKNOWN;
}
}
2011-06-15 14:16:02 +00:00
GstVideoFormat
gst_video_format_from_string (const gchar * format)
{
guint i;
for (i = 0; i < G_N_ELEMENTS (formats); i++) {
if (strcmp (formats[i].fmt, format) == 0)
return formats[i].format;
}
return GST_VIDEO_FORMAT_UNKNOWN;
}
/**
* gst_video_format_to_fourcc:
* @format: a #GstVideoFormat video format
*
* Converts a #GstVideoFormat value into the corresponding FOURCC. Only
* a few YUV formats have corresponding FOURCC values. If @format has
* no corresponding FOURCC value, 0 is returned.
*
* Since: 0.10.16
*
* Returns: the FOURCC corresponding to @format
*/
guint32
gst_video_format_to_fourcc (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
2011-06-15 14:16:02 +00:00
if (format >= G_N_ELEMENTS (formats))
return 0;
2011-06-15 14:16:02 +00:00
return formats[format].fourcc;
}
2011-06-15 14:16:02 +00:00
const gchar *
gst_video_format_to_string (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
2011-06-15 14:16:02 +00:00
if (format >= G_N_ELEMENTS (formats))
return NULL;
2011-06-15 14:16:02 +00:00
return formats[format].fmt;
}
/**
* gst_video_format_is_rgb:
* @format: a #GstVideoFormat
*
* Determine whether the video format is an RGB format.
*
* Since: 0.10.16
*
* Returns: TRUE if @format represents RGB video
*/
gboolean
gst_video_format_is_rgb (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
if (format >= G_N_ELEMENTS (formats))
return FALSE;
return (formats[format].flags & VIDEO_FLAG_RGB) != 0;
}
/**
* gst_video_format_is_yuv:
* @format: a #GstVideoFormat
*
* Determine whether the video format is a YUV format.
*
* Since: 0.10.16
*
* Returns: TRUE if @format represents YUV video
*/
gboolean
gst_video_format_is_yuv (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
if (format >= G_N_ELEMENTS (formats))
return FALSE;
return (formats[format].flags & VIDEO_FLAG_YUV) != 0;
}
/**
* gst_video_format_is_gray:
* @format: a #GstVideoFormat
*
* Determine whether the video format is a grayscale format.
*
* Since: 0.10.29
*
* Returns: TRUE if @format represents grayscale video
*/
gboolean
gst_video_format_is_gray (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
if (format >= G_N_ELEMENTS (formats))
return FALSE;
return (formats[format].flags & VIDEO_FLAG_GRAY) != 0;
}
/**
* gst_video_format_has_alpha:
* @format: a #GstVideoFormat
2011-05-23 10:53:06 +00:00
*
* Returns TRUE or FALSE depending on if the video format provides an
* alpha channel.
*
* Since: 0.10.16
*
* Returns: TRUE if @format has an alpha channel
*/
gboolean
gst_video_format_has_alpha (GstVideoFormat format)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
if (format >= G_N_ELEMENTS (formats))
return FALSE;
return (formats[format].flags & VIDEO_FLAG_ALPHA) != 0;
}
/**
* gst_video_format_get_component_depth:
* @format: a #GstVideoFormat
2011-05-23 21:12:26 +00:00
* @component: the video component (e.g. 0 for 'R' in RGB)
2011-05-23 10:53:06 +00:00
*
* Returns the number of bits used to encode an individual pixel of
2011-05-23 21:12:26 +00:00
* a given @component. Typically this is 8, although higher and lower
* values are possible for some formats.
*
* Since: 0.10.33
*
* Returns: depth of component
*/
int
gst_video_format_get_component_depth (GstVideoFormat format, int component)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component < GST_VIDEO_MAX_PLANES, 0);
if (format >= G_N_ELEMENTS (formats))
return FALSE;
return formats[format].depth[component];
}
/**
* gst_video_format_get_pixel_stride:
* @format: a #GstVideoFormat
* @component: the component index
*
* Calculates the pixel stride (number of bytes from one pixel to the
* pixel to its immediate left) for the video component with an index
* of @component. See @gst_video_format_get_row_stride for a description
* of the component index.
*
* Since: 0.10.16
*
* Returns: pixel stride of component @component
*/
int
gst_video_format_get_pixel_stride (GstVideoFormat format, int component)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
case GST_VIDEO_FORMAT_Y41B:
case GST_VIDEO_FORMAT_Y42B:
case GST_VIDEO_FORMAT_Y444:
case GST_VIDEO_FORMAT_A420:
case GST_VIDEO_FORMAT_YUV9:
case GST_VIDEO_FORMAT_YVU9:
return 1;
case GST_VIDEO_FORMAT_YUY2:
case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
if (component == 0) {
return 2;
} else {
return 4;
}
case GST_VIDEO_FORMAT_IYU1:
/* doesn't make much sense for IYU1 because it's 1 or 3
* for luma depending on position */
return 0;
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_RGBA:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
case GST_VIDEO_FORMAT_r210:
return 4;
case GST_VIDEO_FORMAT_RGB16:
case GST_VIDEO_FORMAT_BGR16:
case GST_VIDEO_FORMAT_RGB15:
case GST_VIDEO_FORMAT_BGR15:
return 2;
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
case GST_VIDEO_FORMAT_v308:
return 3;
case GST_VIDEO_FORMAT_v210:
/* v210 is packed at the bit level, so pixel stride doesn't make sense */
return 0;
case GST_VIDEO_FORMAT_v216:
if (component == 0) {
return 4;
} else {
return 8;
}
case GST_VIDEO_FORMAT_NV12:
case GST_VIDEO_FORMAT_NV21:
if (component == 0) {
return 1;
} else {
return 2;
}
case GST_VIDEO_FORMAT_GRAY8:
case GST_VIDEO_FORMAT_Y800:
return 1;
case GST_VIDEO_FORMAT_GRAY16_BE:
case GST_VIDEO_FORMAT_GRAY16_LE:
case GST_VIDEO_FORMAT_Y16:
return 2;
case GST_VIDEO_FORMAT_UYVP:
/* UYVP is packed at the bit level, so pixel stride doesn't make sense */
return 0;
case GST_VIDEO_FORMAT_RGB8_PALETTED:
return 1;
case GST_VIDEO_FORMAT_ARGB64:
case GST_VIDEO_FORMAT_AYUV64:
return 8;
default:
return 0;
}
}
/**
* gst_video_info_init:
* @info: a #GstVideoInfo
*
* Initialize @info with default values.
*/
void
gst_video_info_init (GstVideoInfo * info)
{
g_return_if_fail (info != NULL);
memset (info, 0, sizeof (GstVideoInfo));
}
/**
* gst_video_info_set_format:
* @info: a #GstVideoInfo
* @format: the format
* @width: a width
* @height: a height
*
* Set the default info for a video frame of @format and @width and @height.
*/
void
gst_video_info_set_format (GstVideoInfo * info, GstVideoFormat format,
guint width, guint height)
{
g_return_if_fail (info != NULL);
g_return_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN);
info->format = format;
info->width = width;
info->heigth = height;
}
/**
* gst_video_info_from_caps:
* @info: a #GstVideoInfo
* @caps: a #GstCaps
*
* Parse @caps and update @info.
*
* Returns: TRUE if @caps could be parsed
*/
gboolean
gst_video_info_from_caps (GstVideoInfo * info, const GstCaps * caps)
{
GstStructure *structure;
const gchar *fmt;
gboolean ok = TRUE;
GstVideoFormat format;
gint width, height g_return_val_if_fail (info != NULL, FALSE);
g_return_val_if_fail (caps != NULL, FALSE);
if (!gst_caps_is_fixed (caps))
return FALSE;
structure = gst_caps_get_structure (caps, 0);
if (!gst_structure_has_name (structure, "video/x-raw"))
goto wrong_name;
if (!(fmt = gst_structure_get_string (structure, "format")))
goto no_format;
format = gst_video_format_from_string (fmt);
if (format == GST_VIDEO_FORMAT_UNKNOWN)
goto unknown_format;
info->format = format;
if (gst_structure_get_int (structure, "width", &width))
info->width = width;
if (gst_structure_get_int (structure, "height", &height))
info->height = height;
return TRUE;
/* ERROR */
wrong_name:
{
return FALSE;
}
no_format:
{
return FALSE;
}
unknown_format:
{
return FALSE;
}
}
GstCaps *
gst_video_info_to_caps (GstVideoInfo * info)
{
return NULL;
}
gboolean
gst_video_info_convert (GstVideoInfo * info,
GstFormat src_format, gint64 src_value, GstFormat dest_format,
gint64 * dest_value)
{
return TRUE;
}
/**
* gst_video_format_get_row_stride:
* @format: a #GstVideoFormat
* @component: the component index
* @width: the width of video
*
* Calculates the row stride (number of bytes from one row of pixels to
* the next) for the video component with an index of @component. For
* YUV video, Y, U, and V have component indices of 0, 1, and 2,
* respectively. For RGB video, R, G, and B have component indicies of
* 0, 1, and 2, respectively. Alpha channels, if present, have a component
* index of 3. The @width parameter always represents the width of the
* video, not the component.
*
* Since: 0.10.16
*
* Returns: row stride of component @component
*/
int
gst_video_format_get_row_stride (GstVideoFormat format, int component,
int width)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
g_return_val_if_fail (width > 0, 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
if (component == 0) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
}
case GST_VIDEO_FORMAT_YUY2:
case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
return GST_ROUND_UP_4 (width * 2);
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_RGBA:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
2011-02-26 03:37:07 +00:00
case GST_VIDEO_FORMAT_r210:
return width * 4;
case GST_VIDEO_FORMAT_RGB16:
case GST_VIDEO_FORMAT_BGR16:
case GST_VIDEO_FORMAT_RGB15:
case GST_VIDEO_FORMAT_BGR15:
return GST_ROUND_UP_4 (width * 2);
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
case GST_VIDEO_FORMAT_v308:
return GST_ROUND_UP_4 (width * 3);
case GST_VIDEO_FORMAT_Y41B:
if (component == 0) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_16 (width) / 4;
}
case GST_VIDEO_FORMAT_Y42B:
if (component == 0) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_8 (width) / 2;
}
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y444:
return GST_ROUND_UP_4 (width);
case GST_VIDEO_FORMAT_v210:
return ((width + 47) / 48) * 128;
case GST_VIDEO_FORMAT_v216:
return GST_ROUND_UP_8 (width * 4);
2009-09-23 08:25:02 +00:00
case GST_VIDEO_FORMAT_NV12:
case GST_VIDEO_FORMAT_NV21:
return GST_ROUND_UP_4 (width);
case GST_VIDEO_FORMAT_GRAY8:
case GST_VIDEO_FORMAT_Y800:
return GST_ROUND_UP_4 (width);
case GST_VIDEO_FORMAT_GRAY16_BE:
case GST_VIDEO_FORMAT_GRAY16_LE:
case GST_VIDEO_FORMAT_Y16:
return GST_ROUND_UP_4 (width * 2);
2010-09-13 02:34:28 +00:00
case GST_VIDEO_FORMAT_UYVP:
return GST_ROUND_UP_4 ((width * 2 * 5 + 3) / 4);
case GST_VIDEO_FORMAT_A420:
if (component == 0 || component == 3) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
}
case GST_VIDEO_FORMAT_RGB8_PALETTED:
return GST_ROUND_UP_4 (width);
case GST_VIDEO_FORMAT_YUV9:
case GST_VIDEO_FORMAT_YVU9:
if (component == 0) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4);
}
case GST_VIDEO_FORMAT_IYU1:
return GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) +
GST_ROUND_UP_4 (width) / 2);
case GST_VIDEO_FORMAT_ARGB64:
case GST_VIDEO_FORMAT_AYUV64:
return width * 8;
default:
return 0;
}
}
/**
* gst_video_format_get_component_width:
* @format: a #GstVideoFormat
* @component: the component index
* @width: the width of video
*
* Calculates the width of the component. See
* @gst_video_format_get_row_stride for a description
* of the component index.
*
* Since: 0.10.16
*
* Returns: width of component @component
*/
int
gst_video_format_get_component_width (GstVideoFormat format,
int component, int width)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
g_return_val_if_fail (width > 0, 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
case GST_VIDEO_FORMAT_YUY2:
case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y42B:
case GST_VIDEO_FORMAT_v210:
case GST_VIDEO_FORMAT_v216:
case GST_VIDEO_FORMAT_NV12:
case GST_VIDEO_FORMAT_NV21:
2010-09-13 02:34:28 +00:00
case GST_VIDEO_FORMAT_UYVP:
if (component == 0) {
return width;
} else {
return GST_ROUND_UP_2 (width) / 2;
}
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y41B:
case GST_VIDEO_FORMAT_YUV9:
case GST_VIDEO_FORMAT_YVU9:
case GST_VIDEO_FORMAT_IYU1:
if (component == 0) {
return width;
} else {
2009-05-15 23:18:18 +00:00
return GST_ROUND_UP_4 (width) / 4;
}
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_RGBA:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
case GST_VIDEO_FORMAT_RGB16:
case GST_VIDEO_FORMAT_BGR16:
case GST_VIDEO_FORMAT_RGB15:
case GST_VIDEO_FORMAT_BGR15:
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y444:
case GST_VIDEO_FORMAT_v308:
case GST_VIDEO_FORMAT_GRAY8:
case GST_VIDEO_FORMAT_GRAY16_BE:
case GST_VIDEO_FORMAT_GRAY16_LE:
case GST_VIDEO_FORMAT_Y800:
case GST_VIDEO_FORMAT_Y16:
case GST_VIDEO_FORMAT_RGB8_PALETTED:
case GST_VIDEO_FORMAT_ARGB64:
case GST_VIDEO_FORMAT_AYUV64:
2011-02-26 03:37:07 +00:00
case GST_VIDEO_FORMAT_r210:
return width;
case GST_VIDEO_FORMAT_A420:
if (component == 0 || component == 3) {
return width;
} else {
return GST_ROUND_UP_2 (width) / 2;
}
default:
return 0;
}
}
/**
* gst_video_format_get_component_height:
* @format: a #GstVideoFormat
* @component: the component index
* @height: the height of video
*
* Calculates the height of the component. See
* @gst_video_format_get_row_stride for a description
* of the component index.
*
* Since: 0.10.16
*
* Returns: height of component @component
*/
int
gst_video_format_get_component_height (GstVideoFormat format,
int component, int height)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
g_return_val_if_fail (height > 0, 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
2009-09-23 08:25:02 +00:00
case GST_VIDEO_FORMAT_NV12:
case GST_VIDEO_FORMAT_NV21:
if (component == 0) {
return height;
} else {
return GST_ROUND_UP_2 (height) / 2;
}
case GST_VIDEO_FORMAT_Y41B:
case GST_VIDEO_FORMAT_Y42B:
case GST_VIDEO_FORMAT_YUY2:
case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_RGBA:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
case GST_VIDEO_FORMAT_RGB16:
case GST_VIDEO_FORMAT_BGR16:
case GST_VIDEO_FORMAT_RGB15:
case GST_VIDEO_FORMAT_BGR15:
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y444:
case GST_VIDEO_FORMAT_v210:
case GST_VIDEO_FORMAT_v216:
case GST_VIDEO_FORMAT_v308:
case GST_VIDEO_FORMAT_GRAY8:
case GST_VIDEO_FORMAT_GRAY16_BE:
case GST_VIDEO_FORMAT_GRAY16_LE:
case GST_VIDEO_FORMAT_Y800:
case GST_VIDEO_FORMAT_Y16:
2010-09-13 02:34:28 +00:00
case GST_VIDEO_FORMAT_UYVP:
case GST_VIDEO_FORMAT_RGB8_PALETTED:
case GST_VIDEO_FORMAT_IYU1:
case GST_VIDEO_FORMAT_ARGB64:
case GST_VIDEO_FORMAT_AYUV64:
2011-02-26 03:37:07 +00:00
case GST_VIDEO_FORMAT_r210:
return height;
case GST_VIDEO_FORMAT_A420:
if (component == 0 || component == 3) {
return height;
} else {
return GST_ROUND_UP_2 (height) / 2;
}
case GST_VIDEO_FORMAT_YUV9:
case GST_VIDEO_FORMAT_YVU9:
if (component == 0) {
return height;
} else {
return GST_ROUND_UP_4 (height) / 4;
}
default:
return 0;
}
}
/**
* gst_video_format_get_component_offset:
* @format: a #GstVideoFormat
* @component: the component index
* @width: the width of video
* @height: the height of video
*
* Calculates the offset (in bytes) of the first pixel of the component
* with index @component. For packed formats, this will typically be a
* small integer (0, 1, 2, 3). For planar formats, this will be a
* (relatively) large offset to the beginning of the second or third
* component planes. See @gst_video_format_get_row_stride for a description
* of the component index.
*
* Since: 0.10.16
*
* Returns: offset of component @component
*/
int
gst_video_format_get_component_offset (GstVideoFormat format,
int component, int width, int height)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
g_return_val_if_fail ((!gst_video_format_is_yuv (format)) || (width > 0
&& height > 0), 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
if (component == 2) {
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) +
GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2);
}
break;
case GST_VIDEO_FORMAT_YV12: /* same as I420, but components 1+2 swapped */
if (component == 0)
return 0;
if (component == 2)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
if (component == 1) {
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) +
GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2);
}
break;
case GST_VIDEO_FORMAT_YUY2:
if (component == 0)
return 0;
if (component == 1)
return 1;
if (component == 2)
return 3;
break;
case GST_VIDEO_FORMAT_YVYU:
if (component == 0)
return 0;
if (component == 1)
return 3;
if (component == 2)
return 1;
break;
case GST_VIDEO_FORMAT_UYVY:
if (component == 0)
return 1;
if (component == 1)
return 0;
if (component == 2)
return 2;
break;
case GST_VIDEO_FORMAT_AYUV:
if (component == 0)
return 1;
if (component == 1)
return 2;
if (component == 2)
return 3;
if (component == 3)
return 0;
break;
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_RGBA:
if (component == 0)
return 0;
if (component == 1)
return 1;
if (component == 2)
return 2;
if (component == 3)
return 3;
break;
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_BGRA:
if (component == 0)
return 2;
if (component == 1)
return 1;
if (component == 2)
return 0;
if (component == 3)
return 3;
break;
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_ARGB:
if (component == 0)
return 1;
if (component == 1)
return 2;
if (component == 2)
return 3;
if (component == 3)
return 0;
break;
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_ABGR:
if (component == 0)
return 3;
if (component == 1)
return 2;
if (component == 2)
return 1;
if (component == 3)
return 0;
break;
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_v308:
if (component == 0)
return 0;
if (component == 1)
return 1;
if (component == 2)
return 2;
break;
case GST_VIDEO_FORMAT_BGR:
if (component == 0)
return 2;
if (component == 1)
return 1;
if (component == 2)
return 0;
break;
case GST_VIDEO_FORMAT_Y41B:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * height;
if (component == 2)
return (GST_ROUND_UP_4 (width) +
(GST_ROUND_UP_16 (width) / 4)) * height;
break;
case GST_VIDEO_FORMAT_Y42B:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * height;
if (component == 2)
return (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_8 (width) / 2)) * height;
break;
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y444:
return GST_ROUND_UP_4 (width) * height * component;
case GST_VIDEO_FORMAT_v210:
2011-02-26 03:37:07 +00:00
case GST_VIDEO_FORMAT_r210:
2009-05-15 23:18:18 +00:00
/* v210 is bit-packed, so this doesn't make sense */
return 0;
case GST_VIDEO_FORMAT_v216:
if (component == 0)
return 0;
if (component == 1)
return 2;
if (component == 2)
return 6;
break;
2009-09-23 08:25:02 +00:00
case GST_VIDEO_FORMAT_NV12:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
if (component == 2)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) + 1;
break;
2009-09-23 08:25:02 +00:00
case GST_VIDEO_FORMAT_NV21:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) + 1;
if (component == 2)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
break;
case GST_VIDEO_FORMAT_GRAY8:
case GST_VIDEO_FORMAT_GRAY16_BE:
case GST_VIDEO_FORMAT_GRAY16_LE:
case GST_VIDEO_FORMAT_Y800:
case GST_VIDEO_FORMAT_Y16:
return 0;
2010-09-13 02:34:28 +00:00
case GST_VIDEO_FORMAT_UYVP:
/* UYVP is bit-packed, so this doesn't make sense */
return 0;
case GST_VIDEO_FORMAT_A420:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
if (component == 2) {
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) +
GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2);
}
if (component == 3) {
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) +
2 * GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2);
}
break;
case GST_VIDEO_FORMAT_RGB8_PALETTED:
return 0;
case GST_VIDEO_FORMAT_YUV9:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * height;
if (component == 2) {
return GST_ROUND_UP_4 (width) * height +
GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
(GST_ROUND_UP_4 (height) / 4);
}
break;
case GST_VIDEO_FORMAT_YVU9:
if (component == 0)
return 0;
if (component == 1) {
2011-04-04 23:00:30 +00:00
return GST_ROUND_UP_4 (width) * height +
GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
(GST_ROUND_UP_4 (height) / 4);
}
if (component == 2)
2011-04-04 23:00:30 +00:00
return GST_ROUND_UP_4 (width) * height;
break;
case GST_VIDEO_FORMAT_IYU1:
if (component == 0)
return 1;
if (component == 1)
return 0;
if (component == 2)
return 4;
break;
case GST_VIDEO_FORMAT_ARGB64:
case GST_VIDEO_FORMAT_AYUV64:
if (component == 0)
return 2;
if (component == 1)
return 4;
if (component == 2)
return 6;
if (component == 3)
return 0;
break;
default:
break;
}
GST_WARNING ("unhandled format %d or component %d", format, component);
return 0;
}
/**
* gst_video_format_get_size:
* @format: a #GstVideoFormat
* @width: the width of video
* @height: the height of video
*
* Calculates the total number of bytes in the raw video format. This
* number should be used when allocating a buffer for raw video.
*
* Since: 0.10.16
*
* Returns: size (in bytes) of raw video format
*/
int
gst_video_format_get_size (GstVideoFormat format, int width, int height)
{
int size;
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (width > 0 && height > 0, 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
size = GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
size += GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2) * 2;
return size;
case GST_VIDEO_FORMAT_IYU1:
return GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) +
GST_ROUND_UP_4 (width) / 2) * height;
case GST_VIDEO_FORMAT_YUY2:
case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
return GST_ROUND_UP_4 (width * 2) * height;
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_RGBA:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
2011-02-26 03:37:07 +00:00
case GST_VIDEO_FORMAT_r210:
return width * 4 * height;
case GST_VIDEO_FORMAT_RGB16:
case GST_VIDEO_FORMAT_BGR16:
case GST_VIDEO_FORMAT_RGB15:
case GST_VIDEO_FORMAT_BGR15:
return GST_ROUND_UP_4 (width * 2) * height;
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
case GST_VIDEO_FORMAT_v308:
return GST_ROUND_UP_4 (width * 3) * height;
case GST_VIDEO_FORMAT_Y41B:
2010-03-09 18:17:04 +00:00
/* simplification of ROUNDUP4(w)*h + 2*((ROUNDUP16(w)/4)*h */
return (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_16 (width) / 2)) * height;
case GST_VIDEO_FORMAT_Y42B:
2010-03-09 18:17:04 +00:00
/* simplification of ROUNDUP4(w)*h + 2*(ROUNDUP8(w)/2)*h */
return (GST_ROUND_UP_4 (width) + GST_ROUND_UP_8 (width)) * height;
2009-05-15 23:18:18 +00:00
case GST_VIDEO_FORMAT_Y444:
return GST_ROUND_UP_4 (width) * height * 3;
case GST_VIDEO_FORMAT_v210:
return ((width + 47) / 48) * 128 * height;
case GST_VIDEO_FORMAT_v216:
return GST_ROUND_UP_8 (width * 4) * height;
2009-09-23 08:25:02 +00:00
case GST_VIDEO_FORMAT_NV12:
case GST_VIDEO_FORMAT_NV21:
return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) * 3 / 2;
case GST_VIDEO_FORMAT_GRAY8:
case GST_VIDEO_FORMAT_Y800:
case GST_VIDEO_FORMAT_RGB8_PALETTED:
return GST_ROUND_UP_4 (width) * height;
case GST_VIDEO_FORMAT_GRAY16_BE:
case GST_VIDEO_FORMAT_GRAY16_LE:
case GST_VIDEO_FORMAT_Y16:
return GST_ROUND_UP_4 (width * 2) * height;
2010-09-13 02:34:28 +00:00
case GST_VIDEO_FORMAT_UYVP:
return GST_ROUND_UP_4 ((width * 2 * 5 + 3) / 4) * height;
case GST_VIDEO_FORMAT_A420:
size = 2 * GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
size += GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2) * 2;
return size;
case GST_VIDEO_FORMAT_YUV9:
case GST_VIDEO_FORMAT_YVU9:
2011-04-04 23:00:30 +00:00
size = GST_ROUND_UP_4 (width) * height;
size += GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
(GST_ROUND_UP_4 (height) / 4) * 2;
return size;
case GST_VIDEO_FORMAT_ARGB64:
case GST_VIDEO_FORMAT_AYUV64:
return width * 8 * height;
default:
return 0;
}
}
/**
* gst_video_get_size_from_caps:
* @caps: a pointer to #GstCaps
* @size: a pointer to a gint that will be assigned the size (in bytes) of a video frame with the given caps
*
* Calculates the total number of bytes in the raw video format for the given
* caps. This number should be used when allocating a buffer for raw video.
*
* Since: 0.10.34
*
* Returns: %TRUE if the size could be calculated from the caps
*/
gboolean
gst_video_get_size_from_caps (const GstCaps * caps, gint * size)
{
GstVideoFormat format = 0;
gint width = 0, height = 0;
g_return_val_if_fail (gst_caps_is_fixed (caps), FALSE);
g_return_val_if_fail (size != NULL, FALSE);
if (gst_video_format_parse_caps (caps, &format, &width, &height) == FALSE) {
GST_WARNING ("Could not parse caps: %" GST_PTR_FORMAT, caps);
return FALSE;
}
*size = gst_video_format_get_size (format, width, height);
return TRUE;
}
/**
* gst_video_format_convert:
* @format: a #GstVideoFormat
* @width: the width of video
* @height: the height of video
* @fps_n: frame rate numerator
* @fps_d: frame rate denominator
* @src_format: #GstFormat of the @src_value
* @src_value: value to convert
* @dest_format: #GstFormat of the @dest_value
* @dest_value: pointer to destination value
*
* Converts among various #GstFormat types. This function handles
* GST_FORMAT_BYTES, GST_FORMAT_TIME, and GST_FORMAT_DEFAULT. For
* raw video, GST_FORMAT_DEFAULT corresponds to video frames. This
* function can be to handle pad queries of the type GST_QUERY_CONVERT.
*
* Since: 0.10.16
*
* Returns: TRUE if the conversion was successful.
*/
gboolean
gst_video_format_convert (GstVideoFormat format, int width, int height,
int fps_n, int fps_d,
GstFormat src_format, gint64 src_value,
GstFormat dest_format, gint64 * dest_value)
{
gboolean ret = FALSE;
int size;
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (width > 0 && height > 0, 0);
size = gst_video_format_get_size (format, width, height);
GST_DEBUG ("converting value %" G_GINT64_FORMAT " from %s to %s",
src_value, gst_format_get_name (src_format),
gst_format_get_name (dest_format));
if (src_format == dest_format) {
*dest_value = src_value;
ret = TRUE;
goto done;
}
if (src_value == -1) {
*dest_value = -1;
ret = TRUE;
goto done;
}
/* bytes to frames */
if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_DEFAULT) {
if (size != 0) {
*dest_value = gst_util_uint64_scale_int (src_value, 1, size);
} else {
GST_ERROR ("blocksize is 0");
*dest_value = 0;
}
ret = TRUE;
goto done;
}
/* frames to bytes */
if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_BYTES) {
*dest_value = gst_util_uint64_scale_int (src_value, size, 1);
ret = TRUE;
goto done;
}
/* time to frames */
if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_DEFAULT) {
if (fps_d != 0) {
*dest_value = gst_util_uint64_scale (src_value,
fps_n, GST_SECOND * fps_d);
} else {
GST_ERROR ("framerate denominator is 0");
*dest_value = 0;
}
ret = TRUE;
goto done;
}
/* frames to time */
if (src_format == GST_FORMAT_DEFAULT && dest_format == GST_FORMAT_TIME) {
if (fps_n != 0) {
*dest_value = gst_util_uint64_scale (src_value,
GST_SECOND * fps_d, fps_n);
} else {
GST_ERROR ("framerate numerator is 0");
*dest_value = 0;
}
ret = TRUE;
goto done;
}
/* time to bytes */
if (src_format == GST_FORMAT_TIME && dest_format == GST_FORMAT_BYTES) {
if (fps_d != 0) {
*dest_value = gst_util_uint64_scale (src_value,
fps_n * size, GST_SECOND * fps_d);
} else {
GST_ERROR ("framerate denominator is 0");
*dest_value = 0;
}
ret = TRUE;
goto done;
}
/* bytes to time */
if (src_format == GST_FORMAT_BYTES && dest_format == GST_FORMAT_TIME) {
if (fps_n != 0 && size != 0) {
*dest_value = gst_util_uint64_scale (src_value,
GST_SECOND * fps_d, fps_n * size);
} else {
GST_ERROR ("framerate denominator and/or blocksize is 0");
*dest_value = 0;
}
ret = TRUE;
}
done:
GST_DEBUG ("ret=%d result %" G_GINT64_FORMAT, ret, *dest_value);
return ret;
}
#define GST_VIDEO_EVENT_STILL_STATE_NAME "GstEventStillFrame"
/**
* gst_video_event_new_still_frame:
* @in_still: boolean value for the still-frame state of the event.
*
* Creates a new Still Frame event. If @in_still is %TRUE, then the event
* represents the start of a still frame sequence. If it is %FALSE, then
* the event ends a still frame sequence.
*
* To parse an event created by gst_video_event_new_still_frame() use
* gst_video_event_parse_still_frame().
*
* Returns: The new GstEvent
* Since: 0.10.26
*/
GstEvent *
gst_video_event_new_still_frame (gboolean in_still)
{
GstEvent *still_event;
GstStructure *s;
s = gst_structure_new (GST_VIDEO_EVENT_STILL_STATE_NAME,
"still-state", G_TYPE_BOOLEAN, in_still, NULL);
still_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s);
return still_event;
}
/**
* gst_video_event_parse_still_frame:
* @event: A #GstEvent to parse
* @in_still: A boolean to receive the still-frame status from the event, or NULL
*
* Parse a #GstEvent, identify if it is a Still Frame event, and
* return the still-frame state from the event if it is.
* If the event represents the start of a still frame, the in_still
* variable will be set to TRUE, otherwise FALSE. It is OK to pass NULL for the
* in_still variable order to just check whether the event is a valid still-frame
* event.
*
* Create a still frame event using gst_video_event_new_still_frame()
*
* Returns: %TRUE if the event is a valid still-frame event. %FALSE if not
* Since: 0.10.26
*/
gboolean
gst_video_event_parse_still_frame (GstEvent * event, gboolean * in_still)
{
const GstStructure *s;
gboolean ev_still_state;
g_return_val_if_fail (event != NULL, FALSE);
if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM)
return FALSE; /* Not a still frame event */
s = gst_event_get_structure (event);
if (s == NULL
|| !gst_structure_has_name (s, GST_VIDEO_EVENT_STILL_STATE_NAME))
return FALSE; /* Not a still frame event */
if (!gst_structure_get_boolean (s, "still-state", &ev_still_state))
return FALSE; /* Not a still frame event */
if (in_still)
*in_still = ev_still_state;
return TRUE;
}
/**
* gst_video_parse_caps_palette:
* @caps: #GstCaps to parse
*
* Returns the palette data from the caps as a #GstBuffer. For
* #GST_VIDEO_FORMAT_RGB8_PALETTED this is containing 256 #guint32
* values, each containing ARGB colors in native endianness.
*
* Returns: a #GstBuffer containing the palette data. Unref after usage.
* Since: 0.10.32
*/
GstBuffer *
gst_video_parse_caps_palette (GstCaps * caps)
{
GstStructure *s;
const GValue *p_v;
GstBuffer *p;
if (!gst_caps_is_fixed (caps))
return NULL;
s = gst_caps_get_structure (caps, 0);
p_v = gst_structure_get_value (s, "palette_data");
if (!p_v || !GST_VALUE_HOLDS_BUFFER (p_v))
return NULL;
p = gst_buffer_ref (gst_value_get_buffer (p_v));
return p;
}