/* GStreamer * Copyright (C) <1999> Erik Walthinsen * This file: * Copyright (c) 2002-2004 Ronald Bultje * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include #include "avcodec.h" #include "gstffmpegcodecmap.h" static GstCaps * gst_ff_vid_caps_new (AVCodecContext * context, const char *mimetype, const char *fieldname, ...) G_GNUC_NULL_TERMINATED; static GstCaps *gst_ff_aud_caps_new (AVCodecContext * context, const char *mimetype, const char *fieldname, ...) G_GNUC_NULL_TERMINATED; /* * Read a palette from a caps. */ static void gst_ffmpeg_get_palette (const GstCaps * caps, AVCodecContext * context) { GstStructure *str = gst_caps_get_structure (caps, 0); const GValue *palette_v; /* do we have a palette? */ if ((palette_v = gst_structure_get_value (str, "palette_data")) && context) { const GstBuffer *palette; palette = gst_value_get_buffer (palette_v); if (palette && GST_BUFFER_SIZE (palette) >= 256 * 4) { if (context->palctrl) av_free (context->palctrl); context->palctrl = av_malloc (sizeof (AVPaletteControl)); context->palctrl->palette_changed = 1; memcpy (context->palctrl->palette, GST_BUFFER_DATA (palette), AVPALETTE_SIZE); } } } static void gst_ffmpeg_set_palette (GstCaps * caps, AVCodecContext * context) { if (context->palctrl) { GstBuffer *palette = gst_buffer_new_and_alloc (256 * 4); memcpy (GST_BUFFER_DATA (palette), context->palctrl->palette, AVPALETTE_SIZE); gst_caps_set_simple (caps, "palette_data", GST_TYPE_BUFFER, palette, NULL); gst_buffer_unref (palette); } } /* this function creates caps with fixed or unfixed width/height * properties depending on whether we've got a context. * * See below for why we use this. * * We should actually do this stuff at the end, like in riff-media.c, * but I'm too lazy today. Maybe later. */ static GstCaps * gst_ff_vid_caps_new (AVCodecContext * context, const char *mimetype, const char *fieldname, ...) { GstStructure *structure = NULL; GstCaps *caps = NULL; va_list var_args; if (context != NULL) { caps = gst_caps_new_simple (mimetype, "width", G_TYPE_INT, context->width, "height", G_TYPE_INT, context->height, "framerate", GST_TYPE_FRACTION, (gint) context->frame_rate, (gint) context->frame_rate_base, NULL); } else { caps = gst_caps_new_simple (mimetype, "width", GST_TYPE_INT_RANGE, 1, G_MAXINT, "height", GST_TYPE_INT_RANGE, 1, G_MAXINT, "framerate", GST_TYPE_FRACTION_RANGE, 0, 1, G_MAXINT, 1, NULL); } structure = gst_caps_get_structure (caps, 0); if (structure) { va_start (var_args, fieldname); gst_structure_set_valist (structure, fieldname, var_args); va_end (var_args); } return caps; } /* same for audio - now with channels/sample rate */ static GstCaps * gst_ff_aud_caps_new (AVCodecContext * context, const char *mimetype, const char *fieldname, ...) { GstCaps *caps = NULL; GstStructure *structure = NULL; va_list var_args; if (context != NULL) { caps = gst_caps_new_simple (mimetype, "rate", G_TYPE_INT, context->sample_rate, "channels", G_TYPE_INT, context->channels, NULL); } else { caps = gst_caps_new_simple (mimetype, NULL); } structure = gst_caps_get_structure (caps, 0); if (structure) { va_start (var_args, fieldname); gst_structure_set_valist (structure, fieldname, var_args); va_end (var_args); } return caps; } /* Convert a FFMPEG Pixel Format and optional AVCodecContext * to a GstCaps. If the context is ommitted, no fixed values * for video/audio size will be included in the GstCaps * * See below for usefulness */ static GstCaps * gst_ffmpeg_pixfmt_to_caps (enum PixelFormat pix_fmt, AVCodecContext * context) { GstCaps *caps = NULL; int bpp = 0, depth = 0, endianness = 0; gulong g_mask = 0, r_mask = 0, b_mask = 0, a_mask = 0; guint32 fmt = 0; switch (pix_fmt) { case PIX_FMT_YUV420P: fmt = GST_MAKE_FOURCC ('I', '4', '2', '0'); break; case PIX_FMT_NV12: fmt = GST_MAKE_FOURCC ('N', 'V', '1', '2'); break; case PIX_FMT_NV21: fmt = GST_MAKE_FOURCC ('N', 'V', '2', '1'); break; case PIX_FMT_YVU420P: fmt = GST_MAKE_FOURCC ('Y', 'V', '1', '2'); break; case PIX_FMT_YUV422: fmt = GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'); break; case PIX_FMT_UYVY422: fmt = GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'); break; case PIX_FMT_YVYU422: fmt = GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'); break; case PIX_FMT_UYVY411: fmt = GST_MAKE_FOURCC ('I', 'Y', 'U', '1'); break; case PIX_FMT_RGB24: bpp = depth = 24; endianness = G_BIG_ENDIAN; r_mask = 0xff0000; g_mask = 0x00ff00; b_mask = 0x0000ff; break; case PIX_FMT_BGR24: bpp = depth = 24; endianness = G_BIG_ENDIAN; r_mask = 0x0000ff; g_mask = 0x00ff00; b_mask = 0xff0000; break; case PIX_FMT_YUV422P: fmt = GST_MAKE_FOURCC ('Y', '4', '2', 'B'); break; case PIX_FMT_YUV444P: fmt = GST_MAKE_FOURCC ('Y', '4', '4', '4'); break; case PIX_FMT_RGB32: bpp = 32; depth = 24; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0x00ff0000; g_mask = 0x0000ff00; b_mask = 0x000000ff; #else r_mask = 0x0000ff00; g_mask = 0x00ff0000; b_mask = 0xff000000; #endif break; case PIX_FMT_BGR32: bpp = 32; depth = 24; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0x0000ff00; g_mask = 0x00ff0000; b_mask = 0xff000000; #else r_mask = 0x00ff0000; g_mask = 0x0000ff00; b_mask = 0x000000ff; #endif break; case PIX_FMT_xRGB32: bpp = 32; depth = 24; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0xff000000; g_mask = 0x00ff0000; b_mask = 0x0000ff00; #else r_mask = 0x000000ff; g_mask = 0x0000ff00; b_mask = 0x00ff0000; #endif break; case PIX_FMT_BGRx32: bpp = 32; depth = 24; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0x000000ff; g_mask = 0x0000ff00; b_mask = 0x00ff0000; #else r_mask = 0xff000000; g_mask = 0x00ff0000; b_mask = 0x0000ff00; #endif break; case PIX_FMT_RGBA32: bpp = 32; depth = 32; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0x00ff0000; g_mask = 0x0000ff00; b_mask = 0x000000ff; a_mask = 0xff000000; #else r_mask = 0x0000ff00; g_mask = 0x00ff0000; b_mask = 0xff000000; a_mask = 0x000000ff; #endif break; case PIX_FMT_BGRA32: bpp = 32; depth = 32; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0x0000ff00; g_mask = 0x00ff0000; b_mask = 0xff000000; a_mask = 0x000000ff; #else r_mask = 0x00ff0000; g_mask = 0x0000ff00; b_mask = 0x000000ff; a_mask = 0xff000000; #endif break; case PIX_FMT_ARGB32: bpp = 32; depth = 32; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0xff000000; g_mask = 0x00ff0000; b_mask = 0x0000ff00; a_mask = 0x000000ff; #else r_mask = 0x000000ff; g_mask = 0x0000ff00; b_mask = 0x00ff0000; a_mask = 0xff000000; #endif break; case PIX_FMT_ABGR32: bpp = 32; depth = 32; endianness = G_BIG_ENDIAN; #if (G_BYTE_ORDER == G_BIG_ENDIAN) r_mask = 0x000000ff; g_mask = 0x0000ff00; b_mask = 0x00ff0000; a_mask = 0xff000000; #else r_mask = 0xff000000; g_mask = 0x00ff0000; b_mask = 0x0000ff00; a_mask = 0x000000ff; #endif break; case PIX_FMT_YUV410P: fmt = GST_MAKE_FOURCC ('Y', 'U', 'V', '9'); break; case PIX_FMT_YVU410P: fmt = GST_MAKE_FOURCC ('Y', 'V', 'U', '9'); break; case PIX_FMT_YUV411P: fmt = GST_MAKE_FOURCC ('Y', '4', '1', 'B'); break; case PIX_FMT_RGB565: bpp = depth = 16; endianness = G_BYTE_ORDER; r_mask = 0xf800; g_mask = 0x07e0; b_mask = 0x001f; break; case PIX_FMT_RGB555: bpp = 16; depth = 15; endianness = G_BYTE_ORDER; r_mask = 0x7c00; g_mask = 0x03e0; b_mask = 0x001f; break; case PIX_FMT_PAL8: bpp = depth = 8; endianness = G_BYTE_ORDER; break; case PIX_FMT_V308: fmt = GST_MAKE_FOURCC ('v', '3', '0', '8'); break; case PIX_FMT_AYUV4444: fmt = GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'); break; case PIX_FMT_GRAY8:{ GstCaps *tmp; bpp = depth = 8; caps = gst_ff_vid_caps_new (context, "video/x-raw-gray", "bpp", G_TYPE_INT, bpp, "depth", G_TYPE_INT, depth, NULL); tmp = gst_ff_vid_caps_new (context, "video/x-raw-yuv", "format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('Y', '8', '0', '0'), NULL); gst_caps_append (caps, tmp); } break; case PIX_FMT_GRAY16_L: bpp = depth = 16; caps = gst_ff_vid_caps_new (context, "video/x-raw-gray", "bpp", G_TYPE_INT, bpp, "depth", G_TYPE_INT, depth, "endianness", G_TYPE_INT, G_LITTLE_ENDIAN, NULL); break; case PIX_FMT_GRAY16_B: bpp = depth = 16; caps = gst_ff_vid_caps_new (context, "video/x-raw-gray", "bpp", G_TYPE_INT, bpp, "depth", G_TYPE_INT, depth, "endianness", G_TYPE_INT, G_BIG_ENDIAN, NULL); break; default: /* give up ... */ break; } if (caps == NULL) { if (bpp != 0) { if (a_mask != 0) { caps = gst_ff_vid_caps_new (context, "video/x-raw-rgb", "bpp", G_TYPE_INT, bpp, "depth", G_TYPE_INT, depth, "red_mask", G_TYPE_INT, r_mask, "green_mask", G_TYPE_INT, g_mask, "blue_mask", G_TYPE_INT, b_mask, "alpha_mask", G_TYPE_INT, a_mask, "endianness", G_TYPE_INT, endianness, NULL); } else if (r_mask != 0) { caps = gst_ff_vid_caps_new (context, "video/x-raw-rgb", "bpp", G_TYPE_INT, bpp, "depth", G_TYPE_INT, depth, "red_mask", G_TYPE_INT, r_mask, "green_mask", G_TYPE_INT, g_mask, "blue_mask", G_TYPE_INT, b_mask, "endianness", G_TYPE_INT, endianness, NULL); } else { caps = gst_ff_vid_caps_new (context, "video/x-raw-rgb", "bpp", G_TYPE_INT, bpp, "depth", G_TYPE_INT, depth, "endianness", G_TYPE_INT, endianness, NULL); if (context && context->pix_fmt == PIX_FMT_PAL8) { gst_ffmpeg_set_palette (caps, context); } } } else if (fmt) { caps = gst_ff_vid_caps_new (context, "video/x-raw-yuv", "format", GST_TYPE_FOURCC, fmt, NULL); } } if (caps != NULL) { GST_DEBUG ("caps for pix_fmt=%d: %" GST_PTR_FORMAT, pix_fmt, caps); } else { GST_LOG ("No caps found for pix_fmt=%d", pix_fmt); } return caps; } /* Convert a FFMPEG Sample Format and optional AVCodecContext * to a GstCaps. If the context is ommitted, no fixed values * for video/audio size will be included in the GstCaps * * See below for usefulness */ static GstCaps * gst_ffmpeg_smpfmt_to_caps (enum SampleFormat sample_fmt, AVCodecContext * context) { GstCaps *caps = NULL; int bpp = 0; gboolean signedness = FALSE; switch (sample_fmt) { case SAMPLE_FMT_S16: signedness = TRUE; bpp = 16; break; default: /* .. */ break; } if (bpp) { caps = gst_ff_aud_caps_new (context, "audio/x-raw-int", "signed", G_TYPE_BOOLEAN, signedness, "endianness", G_TYPE_INT, G_BYTE_ORDER, "width", G_TYPE_INT, bpp, "depth", G_TYPE_INT, bpp, NULL); } if (caps != NULL) { GST_DEBUG ("caps for sample_fmt=%d: %" GST_PTR_FORMAT, sample_fmt, caps); } else { GST_LOG ("No caps found for sample_fmt=%d", sample_fmt); } return caps; } /* Convert a FFMPEG codec Type and optional AVCodecContext * to a GstCaps. If the context is ommitted, no fixed values * for video/audio size will be included in the GstCaps * * CodecType is primarily meant for uncompressed data GstCaps! */ GstCaps * gst_ffmpegcsp_codectype_to_caps (enum CodecType codec_type, AVCodecContext * context) { GstCaps *caps; switch (codec_type) { case CODEC_TYPE_VIDEO: if (context) { caps = gst_ffmpeg_pixfmt_to_caps (context->pix_fmt, context->width == -1 ? NULL : context); } else { GstCaps *temp; enum PixelFormat i; caps = gst_caps_new_empty (); for (i = 0; i < PIX_FMT_NB; i++) { temp = gst_ffmpeg_pixfmt_to_caps (i, NULL); if (temp != NULL) { gst_caps_append (caps, temp); } } } break; case CODEC_TYPE_AUDIO: if (context) { caps = gst_ffmpeg_smpfmt_to_caps (context->sample_fmt, context); } else { GstCaps *temp; enum SampleFormat i; caps = gst_caps_new_empty (); for (i = 0; i <= SAMPLE_FMT_S16; i++) { temp = gst_ffmpeg_smpfmt_to_caps (i, NULL); if (temp != NULL) { gst_caps_append (caps, temp); } } } break; default: /* .. */ caps = NULL; break; } return caps; } /* Convert a GstCaps (audio/raw) to a FFMPEG SampleFmt * and other audio properties in a AVCodecContext. * * For usefulness, see below */ static void gst_ffmpeg_caps_to_smpfmt (const GstCaps * caps, AVCodecContext * context, gboolean raw) { GstStructure *structure; gint depth = 0, width = 0, endianness = 0; gboolean signedness = FALSE; g_return_if_fail (gst_caps_get_size (caps) == 1); structure = gst_caps_get_structure (caps, 0); gst_structure_get_int (structure, "channels", &context->channels); gst_structure_get_int (structure, "rate", &context->sample_rate); if (!raw) return; if (gst_structure_get_int (structure, "width", &width) && gst_structure_get_int (structure, "depth", &depth) && gst_structure_get_boolean (structure, "signed", &signedness) && gst_structure_get_int (structure, "endianness", &endianness)) { if (width == 16 && depth == 16 && endianness == G_BYTE_ORDER && signedness == TRUE) { context->sample_fmt = SAMPLE_FMT_S16; } } } /* Convert a GstCaps (video/raw) to a FFMPEG PixFmt * and other video properties in a AVCodecContext. * * For usefulness, see below */ static void gst_ffmpeg_caps_to_pixfmt (const GstCaps * caps, AVCodecContext * context, gboolean raw) { GstStructure *structure; const GValue *fps; gboolean ret; g_return_if_fail (gst_caps_get_size (caps) == 1); structure = gst_caps_get_structure (caps, 0); ret = gst_structure_get_int (structure, "width", &context->width); ret &= gst_structure_get_int (structure, "height", &context->height); g_return_if_fail (ret == TRUE); fps = gst_structure_get_value (structure, "framerate"); g_return_if_fail (GST_VALUE_HOLDS_FRACTION (fps)); /* framerate does not really matter */ context->frame_rate = gst_value_get_fraction_numerator (fps); context->frame_rate_base = gst_value_get_fraction_denominator (fps); if (!raw) return; if (gst_structure_has_name (structure, "video/x-raw-yuv")) { guint32 fourcc; if (gst_structure_get_fourcc (structure, "format", &fourcc)) { switch (fourcc) { case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'): context->pix_fmt = PIX_FMT_YUV422; break; case GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'): context->pix_fmt = PIX_FMT_UYVY422; break; case GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'): context->pix_fmt = PIX_FMT_YVYU422; break; case GST_MAKE_FOURCC ('I', 'Y', 'U', '1'): context->pix_fmt = PIX_FMT_UYVY411; break; case GST_MAKE_FOURCC ('I', '4', '2', '0'): context->pix_fmt = PIX_FMT_YUV420P; break; case GST_MAKE_FOURCC ('N', 'V', '1', '2'): context->pix_fmt = PIX_FMT_NV12; break; case GST_MAKE_FOURCC ('N', 'V', '2', '1'): context->pix_fmt = PIX_FMT_NV21; break; case GST_MAKE_FOURCC ('Y', 'V', '1', '2'): context->pix_fmt = PIX_FMT_YVU420P; break; case GST_MAKE_FOURCC ('Y', '4', '1', 'B'): context->pix_fmt = PIX_FMT_YUV411P; break; case GST_MAKE_FOURCC ('Y', '4', '2', 'B'): context->pix_fmt = PIX_FMT_YUV422P; break; case GST_MAKE_FOURCC ('Y', 'U', 'V', '9'): context->pix_fmt = PIX_FMT_YUV410P; break; case GST_MAKE_FOURCC ('Y', 'V', 'U', '9'): context->pix_fmt = PIX_FMT_YVU410P; break; case GST_MAKE_FOURCC ('v', '3', '0', '8'): context->pix_fmt = PIX_FMT_V308; break; case GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'): context->pix_fmt = PIX_FMT_AYUV4444; break; case GST_MAKE_FOURCC ('Y', '4', '4', '4'): context->pix_fmt = PIX_FMT_YUV444P; break; case GST_MAKE_FOURCC ('Y', '8', '0', '0'): context->pix_fmt = PIX_FMT_GRAY8; break; } } } else if (gst_structure_has_name (structure, "video/x-raw-rgb")) { gint bpp = 0, rmask = 0, endianness = 0, amask = 0, depth = 0; if (gst_structure_get_int (structure, "bpp", &bpp) && gst_structure_get_int (structure, "endianness", &endianness)) { if (gst_structure_get_int (structure, "red_mask", &rmask)) { switch (bpp) { case 32: if (gst_structure_get_int (structure, "alpha_mask", &amask)) { #if (G_BYTE_ORDER == G_BIG_ENDIAN) if (rmask == 0x0000ff00) context->pix_fmt = PIX_FMT_BGRA32; else if (rmask == 0x00ff0000) context->pix_fmt = PIX_FMT_RGBA32; else if (rmask == 0xff000000) context->pix_fmt = PIX_FMT_ARGB32; else // if (r_mask = 0x000000ff) context->pix_fmt = PIX_FMT_ABGR32; #else if (rmask == 0x00ff0000) context->pix_fmt = PIX_FMT_BGRA32; else if (rmask == 0x0000ff00) context->pix_fmt = PIX_FMT_RGBA32; else if (rmask == 0x000000ff) context->pix_fmt = PIX_FMT_ARGB32; else // if (rmask == 0xff000000) context->pix_fmt = PIX_FMT_ABGR32; #endif } else { #if (G_BYTE_ORDER == G_BIG_ENDIAN) if (rmask == 0x00ff0000) context->pix_fmt = PIX_FMT_RGB32; else if (rmask == 0x0000ff00) context->pix_fmt = PIX_FMT_BGR32; else if (rmask == 0xff000000) context->pix_fmt = PIX_FMT_xRGB32; else // if (rmask == 0x000000ff) context->pix_fmt = PIX_FMT_BGRx32; #else if (rmask == 0x0000ff00) context->pix_fmt = PIX_FMT_RGB32; else if (rmask == 0x00ff0000) context->pix_fmt = PIX_FMT_BGR32; else if (rmask == 0x000000ff) context->pix_fmt = PIX_FMT_xRGB32; else // if (rmask == 0xff000000) context->pix_fmt = PIX_FMT_BGRx32; #endif } break; case 24: if (rmask == 0x0000FF) context->pix_fmt = PIX_FMT_BGR24; else context->pix_fmt = PIX_FMT_RGB24; break; case 16: if (endianness == G_BYTE_ORDER) { context->pix_fmt = PIX_FMT_RGB565; if (gst_structure_get_int (structure, "depth", &depth)) { if (depth == 15) context->pix_fmt = PIX_FMT_RGB555; } } break; case 15: if (endianness == G_BYTE_ORDER) context->pix_fmt = PIX_FMT_RGB555; break; default: /* nothing */ break; } } else { if (bpp == 8) { context->pix_fmt = PIX_FMT_PAL8; gst_ffmpeg_get_palette (caps, context); } } } } else if (gst_structure_has_name (structure, "video/x-raw-gray")) { gint bpp = 0; if (gst_structure_get_int (structure, "bpp", &bpp)) { switch (bpp) { case 8: context->pix_fmt = PIX_FMT_GRAY8; break; case 16:{ gint endianness = 0; if (gst_structure_get_int (structure, "endianness", &endianness)) { if (endianness == G_LITTLE_ENDIAN) context->pix_fmt = PIX_FMT_GRAY16_L; else if (endianness == G_BIG_ENDIAN) context->pix_fmt = PIX_FMT_GRAY16_B; } } break; } } } } /* Convert a GstCaps and a FFMPEG codec Type to a * AVCodecContext. If the context is ommitted, no fixed values * for video/audio size will be included in the context * * CodecType is primarily meant for uncompressed data GstCaps! */ void gst_ffmpegcsp_caps_with_codectype (enum CodecType type, const GstCaps * caps, AVCodecContext * context) { if (context == NULL) return; switch (type) { case CODEC_TYPE_VIDEO: gst_ffmpeg_caps_to_pixfmt (caps, context, TRUE); break; case CODEC_TYPE_AUDIO: gst_ffmpeg_caps_to_smpfmt (caps, context, TRUE); break; default: /* unknown */ break; } } #define GEN_MASK(x) ((1<<(x))-1) #define ROUND_UP_X(v,x) (((v) + GEN_MASK(x)) & ~GEN_MASK(x)) #define DIV_ROUND_UP_X(v,x) (((v) + GEN_MASK(x)) >> (x)) /* * Fill in pointers to memory in a AVPicture, where * everything is aligned by 4 (as required by X). * This is mostly a copy from imgconvert.c with some * small changes. */ int gst_ffmpegcsp_avpicture_fill (AVPicture * picture, uint8_t * ptr, enum PixelFormat pix_fmt, int width, int height) { int size, w2, h2, size2; int stride, stride2; PixFmtInfo *pinfo; pinfo = get_pix_fmt_info (pix_fmt); switch (pix_fmt) { case PIX_FMT_YUV420P: case PIX_FMT_YUV422P: case PIX_FMT_YUV444P: case PIX_FMT_YUV410P: case PIX_FMT_YUV411P: case PIX_FMT_YUVJ420P: case PIX_FMT_YUVJ422P: case PIX_FMT_YUVJ444P: stride = GST_ROUND_UP_4 (width); h2 = ROUND_UP_X (height, pinfo->y_chroma_shift); size = stride * h2; w2 = DIV_ROUND_UP_X (width, pinfo->x_chroma_shift); stride2 = GST_ROUND_UP_4 (w2); h2 = DIV_ROUND_UP_X (height, pinfo->y_chroma_shift); size2 = stride2 * h2; picture->data[0] = ptr; picture->data[1] = picture->data[0] + size; picture->data[2] = picture->data[1] + size2; picture->linesize[0] = stride; picture->linesize[1] = stride2; picture->linesize[2] = stride2; return size + 2 * size2; /* PIX_FMT_YVU420P = YV12: same as PIX_FMT_YUV420P, but * with U and V plane swapped. Strides as in videotestsrc */ case PIX_FMT_YVU410P: case PIX_FMT_YVU420P: stride = GST_ROUND_UP_4 (width); h2 = ROUND_UP_X (height, pinfo->y_chroma_shift); size = stride * h2; w2 = DIV_ROUND_UP_X (width, pinfo->x_chroma_shift); stride2 = GST_ROUND_UP_4 (w2); h2 = DIV_ROUND_UP_X (height, pinfo->y_chroma_shift); size2 = stride2 * h2; picture->data[0] = ptr; picture->data[2] = picture->data[0] + size; picture->data[1] = picture->data[2] + size2; picture->linesize[0] = stride; picture->linesize[1] = stride2; picture->linesize[2] = stride2; return size + 2 * size2; case PIX_FMT_NV12: case PIX_FMT_NV21: stride = GST_ROUND_UP_4 (width); h2 = ROUND_UP_X (height, pinfo->y_chroma_shift); size = stride * h2; w2 = 2 * DIV_ROUND_UP_X (width, pinfo->x_chroma_shift); stride2 = GST_ROUND_UP_4 (w2); h2 = DIV_ROUND_UP_X (height, pinfo->y_chroma_shift); size2 = stride2 * h2; picture->data[0] = ptr; picture->data[1] = picture->data[0] + size; picture->data[2] = NULL; picture->linesize[0] = stride; picture->linesize[1] = stride2; picture->linesize[2] = 0; return size + size2; case PIX_FMT_RGB24: case PIX_FMT_BGR24: stride = GST_ROUND_UP_4 (width * 3); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_AYUV4444: case PIX_FMT_RGB32: case PIX_FMT_RGBA32: case PIX_FMT_ARGB32: case PIX_FMT_BGR32: case PIX_FMT_BGRA32: case PIX_FMT_ABGR32: case PIX_FMT_xRGB32: case PIX_FMT_BGRx32: stride = width * 4; size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_RGB555: case PIX_FMT_RGB565: case PIX_FMT_YUV422: case PIX_FMT_UYVY422: case PIX_FMT_YVYU422: stride = GST_ROUND_UP_4 (width * 2); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_V308: stride = GST_ROUND_UP_4 (width * 3); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_UYVY411: /* FIXME, probably not the right stride */ stride = GST_ROUND_UP_4 (width); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = width + width / 2; return size + size / 2; case PIX_FMT_GRAY8: stride = GST_ROUND_UP_4 (width); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_GRAY16_L: case PIX_FMT_GRAY16_B: stride = GST_ROUND_UP_4 (width * 2); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_MONOWHITE: case PIX_FMT_MONOBLACK: stride = GST_ROUND_UP_4 ((width + 7) >> 3); size = stride * height; picture->data[0] = ptr; picture->data[1] = NULL; picture->data[2] = NULL; picture->linesize[0] = stride; return size; case PIX_FMT_PAL8: /* already forced to be with stride, so same result as other function */ stride = GST_ROUND_UP_4 (width); size = stride * height; picture->data[0] = ptr; picture->data[1] = ptr + size; /* palette is stored here as 256 32 bit words */ picture->data[2] = NULL; picture->linesize[0] = stride; picture->linesize[1] = 4; return size + 256 * 4; default: picture->data[0] = NULL; picture->data[1] = NULL; picture->data[2] = NULL; picture->data[3] = NULL; return -1; } return 0; }