mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-12-22 00:06:36 +00:00
da7a7b47e1
Original commit message from CVS: handle stride correctly
2234 lines
56 KiB
C
2234 lines
56 KiB
C
/*
|
|
* Misc image convertion routines
|
|
* Copyright (c) 2001, 2002, 2003 Fabrice Bellard.
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
/**
|
|
* @file imgconvert.c
|
|
* Misc image convertion routines.
|
|
*/
|
|
|
|
/* TODO:
|
|
* - write 'ffimg' program to test all the image related stuff
|
|
* - move all api to slice based system
|
|
* - integrate deinterlacing, postprocessing and scaling in the conversion process
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include "avcodec.h"
|
|
#include <math.h>
|
|
#include "dsputil.h"
|
|
|
|
#include <string.h>
|
|
|
|
#ifdef HAVE_MMX
|
|
#include "mmx.h"
|
|
#endif
|
|
|
|
#define xglue(x, y) x ## y
|
|
#define glue(x, y) xglue(x, y)
|
|
|
|
#define FF_COLOR_RGB 0 /* RGB color space */
|
|
#define FF_COLOR_GRAY 1 /* gray color space */
|
|
#define FF_COLOR_YUV 2 /* YUV color space. 16 <= Y <= 235, 16 <= U, V <= 240 */
|
|
#define FF_COLOR_YUV_JPEG 3 /* YUV color space. 0 <= Y <= 255, 0 <= U, V <= 255 */
|
|
|
|
#define FF_PIXEL_PLANAR 0 /* each channel has one component in AVPicture */
|
|
#define FF_PIXEL_PACKED 1 /* only one components containing all the channels */
|
|
#define FF_PIXEL_PALETTE 2 /* one components containing indexes for a palette */
|
|
|
|
typedef struct PixFmtInfo
|
|
{
|
|
const char *name;
|
|
uint8_t nb_channels; /* number of channels (including alpha) */
|
|
uint8_t color_type; /* color type (see FF_COLOR_xxx constants) */
|
|
uint8_t pixel_type; /* pixel storage type (see FF_PIXEL_xxx constants) */
|
|
uint8_t is_alpha:1; /* true if alpha can be specified */
|
|
uint8_t x_chroma_shift; /* X chroma subsampling factor is 2 ^ shift */
|
|
uint8_t y_chroma_shift; /* Y chroma subsampling factor is 2 ^ shift */
|
|
uint8_t depth; /* bit depth of the color components */
|
|
}
|
|
PixFmtInfo;
|
|
|
|
/* this table gives more information about formats */
|
|
static PixFmtInfo pix_fmt_info[PIX_FMT_NB] = {
|
|
/* YUV formats */
|
|
/* [PIX_FMT_YUV420P] = */ {
|
|
/* .name = */ "yuv420p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 1,
|
|
/* .y_chroma_shift = */ 1,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUV422] = */ {
|
|
/* .name = */ "yuv422",
|
|
/* .nb_channels = */ 1,
|
|
/* .color_type = */ FF_COLOR_YUV,
|
|
/* .pixel_type = */ FF_PIXEL_PACKED,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 1,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_RGB24] = */ {
|
|
/* .name = */ "rgb24",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_RGB,
|
|
/* .pixel_type = */ FF_PIXEL_PACKED,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_BGR24] = */ {
|
|
/* .name = */ "bgr24",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_RGB,
|
|
/* .pixel_type = */ FF_PIXEL_PACKED,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUV422P] = */ {
|
|
/* .name = */ "yuv422p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 1,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUV444P] = */ {
|
|
/* .name = */ "yuv444p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_RGBA32] = */ {
|
|
/* .name = */ "rgba32",
|
|
/* .nb_channels = */ 4,
|
|
/* .color_type = */ FF_COLOR_RGB,
|
|
/* .pixel_type = */ FF_PIXEL_PACKED,
|
|
/* .is_alpha = */ 1,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUV410P] = */ {
|
|
/* .name = */ "yuv410p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 2,
|
|
/* .y_chroma_shift = */ 2,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUV411P] = */ {
|
|
/* .name = */ "yuv411p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 2,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_RGB565] = */ {
|
|
/* .name = */ "rgb565",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_RGB,
|
|
/* .pixel_type = */ FF_PIXEL_PACKED,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 5,
|
|
},
|
|
/* [PIX_FMT_RGB555] = */ {
|
|
/* .name = */ "rgb555",
|
|
/* .nb_channels = */ 4,
|
|
/* .color_type = */ FF_COLOR_RGB,
|
|
/* .pixel_type = */ FF_PIXEL_PACKED,
|
|
/* .is_alpha = */ 1,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 5,
|
|
},
|
|
/* [PIX_FMT_GRAY8] = */ {
|
|
/* .name = */ "gray",
|
|
/* .nb_channels = */ 1,
|
|
/* .color_type = */ FF_COLOR_GRAY,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_MONOWHITE] = */ {
|
|
/* .name = */ "monow",
|
|
/* .nb_channels = */ 1,
|
|
/* .color_type = */ FF_COLOR_GRAY,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 1,
|
|
},
|
|
/* [PIX_FMT_MONOBLACK] = */ {
|
|
/* .name = */ "monob",
|
|
/* .nb_channels = */ 1,
|
|
/* .color_type = */ FF_COLOR_GRAY,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 1,
|
|
},
|
|
/* [PIX_FMT_PAL8] = */ {
|
|
/* .name = */ "pal8",
|
|
/* .nb_channels = */ 4,
|
|
/* .color_type = */ FF_COLOR_RGB,
|
|
/* .pixel_type = */ FF_PIXEL_PALETTE,
|
|
/* .is_alpha = */ 1,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* JPEG YUV */
|
|
/* [PIX_FMT_YUVJ420P] = */ {
|
|
/* .name = */ "yuvj420p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV_JPEG,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 1,
|
|
/* .y_chroma_shift = */ 1,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUVJ422P] = */ {
|
|
/* .name = */ "yuvj422p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV_JPEG,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 1,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
/* [PIX_FMT_YUVJ444P] = */ {
|
|
/* .name = */ "yuvj444p",
|
|
/* .nb_channels = */ 3,
|
|
/* .color_type = */ FF_COLOR_YUV_JPEG,
|
|
/* .pixel_type = */ FF_PIXEL_PLANAR,
|
|
/* .is_alpha = */ 0,
|
|
/* .x_chroma_shift = */ 0,
|
|
/* .y_chroma_shift = */ 0,
|
|
/* .depth = */ 8,
|
|
},
|
|
};
|
|
|
|
void
|
|
avcodec_get_chroma_sub_sample (int pix_fmt, int *h_shift, int *v_shift)
|
|
{
|
|
*h_shift = pix_fmt_info[pix_fmt].x_chroma_shift;
|
|
*v_shift = pix_fmt_info[pix_fmt].y_chroma_shift;
|
|
}
|
|
|
|
const char *
|
|
avcodec_get_pix_fmt_name (int pix_fmt)
|
|
{
|
|
if (pix_fmt < 0 || pix_fmt >= PIX_FMT_NB)
|
|
return "???";
|
|
else
|
|
return pix_fmt_info[pix_fmt].name;
|
|
}
|
|
|
|
enum PixelFormat
|
|
avcodec_get_pix_fmt (const char *name)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < PIX_FMT_NB; i++)
|
|
if (!strcmp (pix_fmt_info[i].name, name))
|
|
break;
|
|
return i;
|
|
}
|
|
|
|
#ifdef FFMPEG_NOSTRIDE
|
|
/* Picture field are filled with 'ptr' addresses. Also return size */
|
|
int
|
|
avpicture_fill (AVPicture * picture, uint8_t * ptr,
|
|
int pix_fmt, int width, int height)
|
|
{
|
|
int size, w2, h2, size2;
|
|
PixFmtInfo *pinfo;
|
|
|
|
pinfo = &pix_fmt_info[pix_fmt];
|
|
size = width * height;
|
|
switch (pix_fmt) {
|
|
case PIX_FMT_YUV420P:
|
|
case PIX_FMT_YUV422P:
|
|
case PIX_FMT_YUV444P:
|
|
case PIX_FMT_YUV410P:
|
|
case PIX_FMT_YUV411P:
|
|
case PIX_FMT_YUVJ420P:
|
|
case PIX_FMT_YUVJ422P:
|
|
case PIX_FMT_YUVJ444P:
|
|
w2 = (width + (1 << pinfo->x_chroma_shift) - 1) >> pinfo->x_chroma_shift;
|
|
h2 = (height + (1 << pinfo->y_chroma_shift) - 1) >> pinfo->y_chroma_shift;
|
|
size2 = w2 * h2;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = picture->data[0] + size;
|
|
picture->data[2] = picture->data[1] + size2;
|
|
picture->linesize[0] = width;
|
|
picture->linesize[1] = w2;
|
|
picture->linesize[2] = w2;
|
|
return size + 2 * size2;
|
|
case PIX_FMT_RGB24:
|
|
case PIX_FMT_BGR24:
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = width * 3;
|
|
return size * 3;
|
|
case PIX_FMT_RGBA32:
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = width * 4;
|
|
return size * 4;
|
|
case PIX_FMT_RGB555:
|
|
case PIX_FMT_RGB565:
|
|
case PIX_FMT_YUV422:
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = width * 2;
|
|
return size * 2;
|
|
case PIX_FMT_GRAY8:
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = width;
|
|
return size;
|
|
case PIX_FMT_MONOWHITE:
|
|
case PIX_FMT_MONOBLACK:
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = (width + 7) >> 3;
|
|
return picture->linesize[0] * height;
|
|
case PIX_FMT_PAL8:
|
|
size2 = (size + 3) & ~3;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = ptr + size2; /* palette is stored here as 256 32 bit words */
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = width;
|
|
picture->linesize[1] = 4;
|
|
return size2 + 256 * 4;
|
|
default:
|
|
picture->data[0] = NULL;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->data[3] = NULL;
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
#else
|
|
/* Sets bytes per video line and plane pointers, taking stride into account.
|
|
* Stride for each format is the expected image line size rounded up by 4.
|
|
* FIXME: instead of hardcoding here, set the default strides in the picture
|
|
* definition and allow caller to override first. */
|
|
#define ROUND_UP_4(x) (((x) + 3) & ~3)
|
|
int
|
|
avpicture_fill (AVPicture * picture, uint8_t * ptr,
|
|
int pix_fmt, int width, int height)
|
|
{
|
|
int size, w2, h2, size2;
|
|
int stride, stride2;
|
|
PixFmtInfo *pinfo;
|
|
|
|
pinfo = &pix_fmt_info[pix_fmt];
|
|
stride = ROUND_UP_4 (width);
|
|
size = stride * height;
|
|
switch (pix_fmt) {
|
|
case PIX_FMT_YUV420P:
|
|
case PIX_FMT_YUV422P:
|
|
case PIX_FMT_YUV444P:
|
|
case PIX_FMT_YUV410P:
|
|
case PIX_FMT_YUV411P:
|
|
case PIX_FMT_YUVJ420P:
|
|
case PIX_FMT_YUVJ422P:
|
|
case PIX_FMT_YUVJ444P:
|
|
stride = ROUND_UP_4 (width);
|
|
size = stride * height;
|
|
w2 = (width + (1 << pinfo->x_chroma_shift) - 1) >> pinfo->x_chroma_shift;
|
|
stride2 = ROUND_UP_4 (w2);
|
|
h2 = (height + (1 << pinfo->y_chroma_shift) - 1) >> pinfo->y_chroma_shift;
|
|
size2 = stride2 * h2;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = picture->data[0] + size;
|
|
picture->data[2] = picture->data[1] + size2;
|
|
picture->linesize[0] = stride;
|
|
picture->linesize[1] = stride2;
|
|
picture->linesize[2] = stride2;
|
|
return size + 2 * size2;
|
|
case PIX_FMT_RGB24:
|
|
case PIX_FMT_BGR24:
|
|
stride = ROUND_UP_4 (width * 3);
|
|
size = stride * height;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = stride;
|
|
return size;
|
|
case PIX_FMT_RGBA32:
|
|
stride = width * 4;
|
|
size = stride * height;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = stride;
|
|
return size;
|
|
case PIX_FMT_RGB555:
|
|
case PIX_FMT_RGB565:
|
|
case PIX_FMT_YUV422:
|
|
stride = ROUND_UP_4 (width * 2);
|
|
size = stride * height;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = stride;
|
|
return size;
|
|
case PIX_FMT_GRAY8:
|
|
stride = ROUND_UP_4 (width);
|
|
size = stride * height;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = stride;
|
|
return size;
|
|
case PIX_FMT_MONOWHITE:
|
|
case PIX_FMT_MONOBLACK:
|
|
stride = ROUND_UP_4 ((width + 7) >> 3);
|
|
size = stride * height;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = stride;
|
|
return size;
|
|
case PIX_FMT_PAL8:
|
|
/* already forced to be with stride, so same result as other function */
|
|
stride = ROUND_UP_4 (width);
|
|
size = stride * height;
|
|
picture->data[0] = ptr;
|
|
picture->data[1] = ptr + size; /* palette is stored here as 256 32 bit words */
|
|
picture->data[2] = NULL;
|
|
picture->linesize[0] = stride;
|
|
picture->linesize[1] = 4;
|
|
return size + 256 * 4;
|
|
default:
|
|
picture->data[0] = NULL;
|
|
picture->data[1] = NULL;
|
|
picture->data[2] = NULL;
|
|
picture->data[3] = NULL;
|
|
return -1;
|
|
}
|
|
}
|
|
#endif
|
|
|
|
int
|
|
avpicture_layout (const AVPicture * src, int pix_fmt, int width, int height,
|
|
unsigned char *dest, int dest_size)
|
|
{
|
|
PixFmtInfo *pf = &pix_fmt_info[pix_fmt];
|
|
int i, j, w, h, data_planes;
|
|
const unsigned char *s;
|
|
int size = avpicture_get_size (pix_fmt, width, height);
|
|
|
|
if (size > dest_size)
|
|
return -1;
|
|
|
|
if (pf->pixel_type == FF_PIXEL_PACKED || pf->pixel_type == FF_PIXEL_PALETTE) {
|
|
if (pix_fmt == PIX_FMT_YUV422 || pix_fmt == PIX_FMT_RGB565 ||
|
|
pix_fmt == PIX_FMT_RGB555)
|
|
w = width * 2;
|
|
else if (pix_fmt == PIX_FMT_PAL8)
|
|
w = width;
|
|
else
|
|
w = width * (pf->depth * pf->nb_channels / 8);
|
|
|
|
data_planes = 1;
|
|
h = height;
|
|
} else {
|
|
data_planes = pf->nb_channels;
|
|
w = width;
|
|
h = height;
|
|
}
|
|
|
|
for (i = 0; i < data_planes; i++) {
|
|
if (i == 1) {
|
|
w = width >> pf->x_chroma_shift;
|
|
h = height >> pf->y_chroma_shift;
|
|
}
|
|
s = src->data[i];
|
|
for (j = 0; j < h; j++) {
|
|
memcpy (dest, s, w);
|
|
dest += w;
|
|
s += src->linesize[i];
|
|
}
|
|
}
|
|
|
|
if (pf->pixel_type == FF_PIXEL_PALETTE)
|
|
memcpy ((unsigned char *) (((size_t) dest + 3) & ~3), src->data[1],
|
|
256 * 4);
|
|
|
|
return size;
|
|
}
|
|
|
|
int
|
|
avpicture_get_size (int pix_fmt, int width, int height)
|
|
{
|
|
AVPicture dummy_pict;
|
|
|
|
return avpicture_fill (&dummy_pict, NULL, pix_fmt, width, height);
|
|
}
|
|
|
|
/**
|
|
* compute the loss when converting from a pixel format to another
|
|
*/
|
|
int
|
|
avcodec_get_pix_fmt_loss (int dst_pix_fmt, int src_pix_fmt, int has_alpha)
|
|
{
|
|
const PixFmtInfo *pf, *ps;
|
|
int loss;
|
|
|
|
ps = &pix_fmt_info[src_pix_fmt];
|
|
pf = &pix_fmt_info[dst_pix_fmt];
|
|
|
|
/* compute loss */
|
|
loss = 0;
|
|
pf = &pix_fmt_info[dst_pix_fmt];
|
|
if (pf->depth < ps->depth ||
|
|
(dst_pix_fmt == PIX_FMT_RGB555 && src_pix_fmt == PIX_FMT_RGB565))
|
|
loss |= FF_LOSS_DEPTH;
|
|
if (pf->x_chroma_shift > ps->x_chroma_shift ||
|
|
pf->y_chroma_shift > ps->y_chroma_shift)
|
|
loss |= FF_LOSS_RESOLUTION;
|
|
switch (pf->color_type) {
|
|
case FF_COLOR_RGB:
|
|
if (ps->color_type != FF_COLOR_RGB && ps->color_type != FF_COLOR_GRAY)
|
|
loss |= FF_LOSS_COLORSPACE;
|
|
break;
|
|
case FF_COLOR_GRAY:
|
|
if (ps->color_type != FF_COLOR_GRAY)
|
|
loss |= FF_LOSS_COLORSPACE;
|
|
break;
|
|
case FF_COLOR_YUV:
|
|
if (ps->color_type != FF_COLOR_YUV)
|
|
loss |= FF_LOSS_COLORSPACE;
|
|
break;
|
|
case FF_COLOR_YUV_JPEG:
|
|
if (ps->color_type != FF_COLOR_YUV_JPEG &&
|
|
ps->color_type != FF_COLOR_YUV && ps->color_type != FF_COLOR_GRAY)
|
|
loss |= FF_LOSS_COLORSPACE;
|
|
break;
|
|
default:
|
|
/* fail safe test */
|
|
if (ps->color_type != pf->color_type)
|
|
loss |= FF_LOSS_COLORSPACE;
|
|
break;
|
|
}
|
|
if (pf->color_type == FF_COLOR_GRAY && ps->color_type != FF_COLOR_GRAY)
|
|
loss |= FF_LOSS_CHROMA;
|
|
if (!pf->is_alpha && (ps->is_alpha && has_alpha))
|
|
loss |= FF_LOSS_ALPHA;
|
|
if (pf->pixel_type == FF_PIXEL_PALETTE &&
|
|
(ps->pixel_type != FF_PIXEL_PALETTE && ps->color_type != FF_COLOR_GRAY))
|
|
loss |= FF_LOSS_COLORQUANT;
|
|
return loss;
|
|
}
|
|
|
|
static int
|
|
avg_bits_per_pixel (int pix_fmt)
|
|
{
|
|
int bits;
|
|
const PixFmtInfo *pf;
|
|
|
|
pf = &pix_fmt_info[pix_fmt];
|
|
switch (pf->pixel_type) {
|
|
case FF_PIXEL_PACKED:
|
|
switch (pix_fmt) {
|
|
case PIX_FMT_YUV422:
|
|
case PIX_FMT_RGB565:
|
|
case PIX_FMT_RGB555:
|
|
bits = 16;
|
|
break;
|
|
default:
|
|
bits = pf->depth * pf->nb_channels;
|
|
break;
|
|
}
|
|
break;
|
|
case FF_PIXEL_PLANAR:
|
|
if (pf->x_chroma_shift == 0 && pf->y_chroma_shift == 0) {
|
|
bits = pf->depth * pf->nb_channels;
|
|
} else {
|
|
bits = pf->depth + ((2 * pf->depth) >>
|
|
(pf->x_chroma_shift + pf->y_chroma_shift));
|
|
}
|
|
break;
|
|
case FF_PIXEL_PALETTE:
|
|
bits = 8;
|
|
break;
|
|
default:
|
|
bits = -1;
|
|
break;
|
|
}
|
|
return bits;
|
|
}
|
|
|
|
static int
|
|
avcodec_find_best_pix_fmt1 (int pix_fmt_mask,
|
|
int src_pix_fmt, int has_alpha, int loss_mask)
|
|
{
|
|
int dist, i, loss, min_dist, dst_pix_fmt;
|
|
|
|
/* find exact color match with smallest size */
|
|
dst_pix_fmt = -1;
|
|
min_dist = 0x7fffffff;
|
|
for (i = 0; i < PIX_FMT_NB; i++) {
|
|
if (pix_fmt_mask & (1 << i)) {
|
|
loss = avcodec_get_pix_fmt_loss (i, src_pix_fmt, has_alpha) & loss_mask;
|
|
if (loss == 0) {
|
|
dist = avg_bits_per_pixel (i);
|
|
if (dist < min_dist) {
|
|
min_dist = dist;
|
|
dst_pix_fmt = i;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return dst_pix_fmt;
|
|
}
|
|
|
|
/**
|
|
* find best pixel format to convert to. Return -1 if none found
|
|
*/
|
|
int
|
|
avcodec_find_best_pix_fmt (int pix_fmt_mask, int src_pix_fmt,
|
|
int has_alpha, int *loss_ptr)
|
|
{
|
|
int dst_pix_fmt, loss_mask, i;
|
|
static const int loss_mask_order[] = {
|
|
~0, /* no loss first */
|
|
~FF_LOSS_ALPHA,
|
|
~FF_LOSS_RESOLUTION,
|
|
~(FF_LOSS_COLORSPACE | FF_LOSS_RESOLUTION),
|
|
~FF_LOSS_COLORQUANT,
|
|
~FF_LOSS_DEPTH,
|
|
0,
|
|
};
|
|
|
|
/* try with successive loss */
|
|
i = 0;
|
|
for (;;) {
|
|
loss_mask = loss_mask_order[i++];
|
|
dst_pix_fmt = avcodec_find_best_pix_fmt1 (pix_fmt_mask, src_pix_fmt,
|
|
has_alpha, loss_mask);
|
|
if (dst_pix_fmt >= 0)
|
|
goto found;
|
|
if (loss_mask == 0)
|
|
break;
|
|
}
|
|
return -1;
|
|
found:
|
|
if (loss_ptr)
|
|
*loss_ptr = avcodec_get_pix_fmt_loss (dst_pix_fmt, src_pix_fmt, has_alpha);
|
|
return dst_pix_fmt;
|
|
}
|
|
|
|
static void
|
|
img_copy_plane (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
for (; height > 0; height--) {
|
|
memcpy (dst, src, width);
|
|
dst += dst_wrap;
|
|
src += src_wrap;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Copy image 'src' to 'dst'.
|
|
*/
|
|
void
|
|
img_copy (AVPicture * dst, const AVPicture * src,
|
|
int pix_fmt, int width, int height)
|
|
{
|
|
int bwidth, bits, i;
|
|
PixFmtInfo *pf = &pix_fmt_info[pix_fmt];
|
|
|
|
pf = &pix_fmt_info[pix_fmt];
|
|
switch (pf->pixel_type) {
|
|
case FF_PIXEL_PACKED:
|
|
switch (pix_fmt) {
|
|
case PIX_FMT_YUV422:
|
|
case PIX_FMT_RGB565:
|
|
case PIX_FMT_RGB555:
|
|
bits = 16;
|
|
break;
|
|
default:
|
|
bits = pf->depth * pf->nb_channels;
|
|
break;
|
|
}
|
|
bwidth = (width * bits + 7) >> 3;
|
|
img_copy_plane (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0], bwidth, height);
|
|
break;
|
|
case FF_PIXEL_PLANAR:
|
|
for (i = 0; i < pf->nb_channels; i++) {
|
|
int w, h;
|
|
|
|
w = width;
|
|
h = height;
|
|
if (i == 1 || i == 2) {
|
|
w >>= pf->x_chroma_shift;
|
|
h >>= pf->y_chroma_shift;
|
|
}
|
|
bwidth = (w * pf->depth + 7) >> 3;
|
|
img_copy_plane (dst->data[i], dst->linesize[i],
|
|
src->data[i], src->linesize[i], bwidth, h);
|
|
}
|
|
break;
|
|
case FF_PIXEL_PALETTE:
|
|
img_copy_plane (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0], width, height);
|
|
/* copy the palette */
|
|
img_copy_plane (dst->data[1], dst->linesize[1],
|
|
src->data[1], src->linesize[1], 4, 256);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* XXX: totally non optimized */
|
|
|
|
static void
|
|
yuv422_to_yuv420p (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
const uint8_t *p, *p1;
|
|
uint8_t *lum, *cr, *cb, *lum1, *cr1, *cb1;
|
|
int w;
|
|
|
|
p1 = src->data[0];
|
|
lum1 = dst->data[0];
|
|
cb1 = dst->data[1];
|
|
cr1 = dst->data[2];
|
|
|
|
for (; height >= 1; height -= 2) {
|
|
p = p1;
|
|
lum = lum1;
|
|
cb = cb1;
|
|
cr = cr1;
|
|
for (w = width; w >= 2; w -= 2) {
|
|
lum[0] = p[0];
|
|
cb[0] = p[1];
|
|
lum[1] = p[2];
|
|
cr[0] = p[3];
|
|
p += 4;
|
|
lum += 2;
|
|
cb++;
|
|
cr++;
|
|
}
|
|
if (w) {
|
|
lum[0] = p[0];
|
|
cb[0] = p[1];
|
|
cr[0] = p[3];
|
|
cb++;
|
|
cr++;
|
|
}
|
|
p1 += src->linesize[0];
|
|
lum1 += dst->linesize[0];
|
|
if (height > 1) {
|
|
p = p1;
|
|
lum = lum1;
|
|
for (w = width; w >= 2; w -= 2) {
|
|
lum[0] = p[0];
|
|
lum[1] = p[2];
|
|
p += 4;
|
|
lum += 2;
|
|
}
|
|
if (w) {
|
|
lum[0] = p[0];
|
|
}
|
|
p1 += src->linesize[0];
|
|
lum1 += dst->linesize[0];
|
|
}
|
|
cb1 += dst->linesize[1];
|
|
cr1 += dst->linesize[2];
|
|
}
|
|
}
|
|
|
|
static void
|
|
yuv422_to_yuv422p (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
const uint8_t *p, *p1;
|
|
uint8_t *lum, *cr, *cb, *lum1, *cr1, *cb1;
|
|
int w;
|
|
|
|
p1 = src->data[0];
|
|
lum1 = dst->data[0];
|
|
cb1 = dst->data[1];
|
|
cr1 = dst->data[2];
|
|
for (; height > 0; height--) {
|
|
p = p1;
|
|
lum = lum1;
|
|
cb = cb1;
|
|
cr = cr1;
|
|
for (w = width; w >= 2; w -= 2) {
|
|
lum[0] = p[0];
|
|
cb[0] = p[1];
|
|
lum[1] = p[2];
|
|
cr[0] = p[3];
|
|
p += 4;
|
|
lum += 2;
|
|
cb++;
|
|
cr++;
|
|
}
|
|
p1 += src->linesize[0];
|
|
lum1 += dst->linesize[0];
|
|
cb1 += dst->linesize[1];
|
|
cr1 += dst->linesize[2];
|
|
}
|
|
}
|
|
|
|
static void
|
|
yuv422p_to_yuv422 (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
uint8_t *p, *p1;
|
|
const uint8_t *lum, *cr, *cb, *lum1, *cr1, *cb1;
|
|
int w;
|
|
|
|
p1 = dst->data[0];
|
|
lum1 = src->data[0];
|
|
cb1 = src->data[1];
|
|
cr1 = src->data[2];
|
|
for (; height > 0; height--) {
|
|
p = p1;
|
|
lum = lum1;
|
|
cb = cb1;
|
|
cr = cr1;
|
|
for (w = width; w >= 2; w -= 2) {
|
|
p[0] = lum[0];
|
|
p[1] = cb[0];
|
|
p[2] = lum[1];
|
|
p[3] = cr[0];
|
|
p += 4;
|
|
lum += 2;
|
|
cb++;
|
|
cr++;
|
|
}
|
|
p1 += dst->linesize[0];
|
|
lum1 += src->linesize[0];
|
|
cb1 += src->linesize[1];
|
|
cr1 += src->linesize[2];
|
|
}
|
|
}
|
|
|
|
#define SCALEBITS 10
|
|
#define ONE_HALF (1 << (SCALEBITS - 1))
|
|
#define FIX(x) ((int) ((x) * (1<<SCALEBITS) + 0.5))
|
|
|
|
#define YUV_TO_RGB1_CCIR(cb1, cr1)\
|
|
{\
|
|
cb = (cb1) - 128;\
|
|
cr = (cr1) - 128;\
|
|
r_add = FIX(1.40200*255.0/224.0) * cr + ONE_HALF;\
|
|
g_add = - FIX(0.34414*255.0/224.0) * cb - FIX(0.71414*255.0/224.0) * cr + \
|
|
ONE_HALF;\
|
|
b_add = FIX(1.77200*255.0/224.0) * cb + ONE_HALF;\
|
|
}
|
|
|
|
#define YUV_TO_RGB2_CCIR(r, g, b, y1)\
|
|
{\
|
|
y = ((y1) - 16) * FIX(255.0/219.0);\
|
|
r = cm[(y + r_add) >> SCALEBITS];\
|
|
g = cm[(y + g_add) >> SCALEBITS];\
|
|
b = cm[(y + b_add) >> SCALEBITS];\
|
|
}
|
|
|
|
#define YUV_TO_RGB1(cb1, cr1)\
|
|
{\
|
|
cb = (cb1) - 128;\
|
|
cr = (cr1) - 128;\
|
|
r_add = FIX(1.40200) * cr + ONE_HALF;\
|
|
g_add = - FIX(0.34414) * cb - FIX(0.71414) * cr + ONE_HALF;\
|
|
b_add = FIX(1.77200) * cb + ONE_HALF;\
|
|
}
|
|
|
|
#define YUV_TO_RGB2(r, g, b, y1)\
|
|
{\
|
|
y = (y1) << SCALEBITS;\
|
|
r = cm[(y + r_add) >> SCALEBITS];\
|
|
g = cm[(y + g_add) >> SCALEBITS];\
|
|
b = cm[(y + b_add) >> SCALEBITS];\
|
|
}
|
|
|
|
#define Y_CCIR_TO_JPEG(y)\
|
|
cm[((y) * FIX(255.0/219.0) + (ONE_HALF - 16 * FIX(255.0/219.0))) >> SCALEBITS]
|
|
|
|
#define Y_JPEG_TO_CCIR(y)\
|
|
(((y) * FIX(219.0/255.0) + (ONE_HALF + (16 << SCALEBITS))) >> SCALEBITS)
|
|
|
|
#define C_CCIR_TO_JPEG(y)\
|
|
cm[(((y) - 128) * FIX(127.0/112.0) + (ONE_HALF + (128 << SCALEBITS))) >> SCALEBITS]
|
|
|
|
/* NOTE: the clamp is really necessary! */
|
|
static inline int
|
|
C_JPEG_TO_CCIR (int y)
|
|
{
|
|
y = (((y - 128) * FIX (112.0 / 127.0) + (ONE_HALF +
|
|
(128 << SCALEBITS))) >> SCALEBITS);
|
|
if (y < 16)
|
|
y = 16;
|
|
return y;
|
|
}
|
|
|
|
|
|
#define RGB_TO_Y(r, g, b) \
|
|
((FIX(0.29900) * (r) + FIX(0.58700) * (g) + \
|
|
FIX(0.11400) * (b) + ONE_HALF) >> SCALEBITS)
|
|
|
|
#define RGB_TO_U(r1, g1, b1, shift)\
|
|
(((- FIX(0.16874) * r1 - FIX(0.33126) * g1 + \
|
|
FIX(0.50000) * b1 + (ONE_HALF << shift) - 1) >> (SCALEBITS + shift)) + 128)
|
|
|
|
#define RGB_TO_V(r1, g1, b1, shift)\
|
|
(((FIX(0.50000) * r1 - FIX(0.41869) * g1 - \
|
|
FIX(0.08131) * b1 + (ONE_HALF << shift) - 1) >> (SCALEBITS + shift)) + 128)
|
|
|
|
#define RGB_TO_Y_CCIR(r, g, b) \
|
|
((FIX(0.29900*219.0/255.0) * (r) + FIX(0.58700*219.0/255.0) * (g) + \
|
|
FIX(0.11400*219.0/255.0) * (b) + (ONE_HALF + (16 << SCALEBITS))) >> SCALEBITS)
|
|
|
|
#define RGB_TO_U_CCIR(r1, g1, b1, shift)\
|
|
(((- FIX(0.16874*224.0/255.0) * r1 - FIX(0.33126*224.0/255.0) * g1 + \
|
|
FIX(0.50000*224.0/255.0) * b1 + (ONE_HALF << shift) - 1) >> (SCALEBITS + shift)) + 128)
|
|
|
|
#define RGB_TO_V_CCIR(r1, g1, b1, shift)\
|
|
(((FIX(0.50000*224.0/255.0) * r1 - FIX(0.41869*224.0/255.0) * g1 - \
|
|
FIX(0.08131*224.0/255.0) * b1 + (ONE_HALF << shift) - 1) >> (SCALEBITS + shift)) + 128)
|
|
|
|
static uint8_t y_ccir_to_jpeg[256];
|
|
static uint8_t y_jpeg_to_ccir[256];
|
|
static uint8_t c_ccir_to_jpeg[256];
|
|
static uint8_t c_jpeg_to_ccir[256];
|
|
|
|
/* init various conversion tables */
|
|
static void
|
|
img_convert_init (void)
|
|
{
|
|
int i;
|
|
uint8_t *cm = cropTbl + MAX_NEG_CROP;
|
|
|
|
for (i = 0; i < 256; i++) {
|
|
y_ccir_to_jpeg[i] = Y_CCIR_TO_JPEG (i);
|
|
y_jpeg_to_ccir[i] = Y_JPEG_TO_CCIR (i);
|
|
c_ccir_to_jpeg[i] = C_CCIR_TO_JPEG (i);
|
|
c_jpeg_to_ccir[i] = C_JPEG_TO_CCIR (i);
|
|
}
|
|
}
|
|
|
|
/* apply to each pixel the given table */
|
|
static void
|
|
img_apply_table (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap,
|
|
int width, int height, const uint8_t * table1)
|
|
{
|
|
int n;
|
|
const uint8_t *s;
|
|
uint8_t *d;
|
|
const uint8_t *table;
|
|
|
|
table = table1;
|
|
for (; height > 0; height--) {
|
|
s = src;
|
|
d = dst;
|
|
n = width;
|
|
while (n >= 4) {
|
|
d[0] = table[s[0]];
|
|
d[1] = table[s[1]];
|
|
d[2] = table[s[2]];
|
|
d[3] = table[s[3]];
|
|
d += 4;
|
|
s += 4;
|
|
n -= 4;
|
|
}
|
|
while (n > 0) {
|
|
d[0] = table[s[0]];
|
|
d++;
|
|
s++;
|
|
n--;
|
|
}
|
|
dst += dst_wrap;
|
|
src += src_wrap;
|
|
}
|
|
}
|
|
|
|
/* XXX: use generic filter ? */
|
|
/* XXX: in most cases, the sampling position is incorrect */
|
|
|
|
/* 4x1 -> 1x1 */
|
|
static void
|
|
shrink41 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
int w;
|
|
const uint8_t *s;
|
|
uint8_t *d;
|
|
|
|
for (; height > 0; height--) {
|
|
s = src;
|
|
d = dst;
|
|
for (w = width; w > 0; w--) {
|
|
d[0] = (s[0] + s[1] + s[2] + s[3] + 2) >> 2;
|
|
s += 4;
|
|
d++;
|
|
}
|
|
src += src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 2x1 -> 1x1 */
|
|
static void
|
|
shrink21 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
int w;
|
|
const uint8_t *s;
|
|
uint8_t *d;
|
|
|
|
for (; height > 0; height--) {
|
|
s = src;
|
|
d = dst;
|
|
for (w = width; w > 0; w--) {
|
|
d[0] = (s[0] + s[1]) >> 1;
|
|
s += 2;
|
|
d++;
|
|
}
|
|
src += src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 1x2 -> 1x1 */
|
|
static void
|
|
shrink12 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
int w;
|
|
uint8_t *d;
|
|
const uint8_t *s1, *s2;
|
|
|
|
for (; height > 0; height--) {
|
|
s1 = src;
|
|
s2 = s1 + src_wrap;
|
|
d = dst;
|
|
for (w = width; w >= 4; w -= 4) {
|
|
d[0] = (s1[0] + s2[0]) >> 1;
|
|
d[1] = (s1[1] + s2[1]) >> 1;
|
|
d[2] = (s1[2] + s2[2]) >> 1;
|
|
d[3] = (s1[3] + s2[3]) >> 1;
|
|
s1 += 4;
|
|
s2 += 4;
|
|
d += 4;
|
|
}
|
|
for (; w > 0; w--) {
|
|
d[0] = (s1[0] + s2[0]) >> 1;
|
|
s1++;
|
|
s2++;
|
|
d++;
|
|
}
|
|
src += 2 * src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 2x2 -> 1x1 */
|
|
static void
|
|
shrink22 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
int w;
|
|
const uint8_t *s1, *s2;
|
|
uint8_t *d;
|
|
|
|
for (; height > 0; height--) {
|
|
s1 = src;
|
|
s2 = s1 + src_wrap;
|
|
d = dst;
|
|
for (w = width; w >= 4; w -= 4) {
|
|
d[0] = (s1[0] + s1[1] + s2[0] + s2[1] + 2) >> 2;
|
|
d[1] = (s1[2] + s1[3] + s2[2] + s2[3] + 2) >> 2;
|
|
d[2] = (s1[4] + s1[5] + s2[4] + s2[5] + 2) >> 2;
|
|
d[3] = (s1[6] + s1[7] + s2[6] + s2[7] + 2) >> 2;
|
|
s1 += 8;
|
|
s2 += 8;
|
|
d += 4;
|
|
}
|
|
for (; w > 0; w--) {
|
|
d[0] = (s1[0] + s1[1] + s2[0] + s2[1] + 2) >> 2;
|
|
s1 += 2;
|
|
s2 += 2;
|
|
d++;
|
|
}
|
|
src += 2 * src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 4x4 -> 1x1 */
|
|
static void
|
|
shrink44 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
int w;
|
|
const uint8_t *s1, *s2, *s3, *s4;
|
|
uint8_t *d;
|
|
|
|
for (; height > 0; height--) {
|
|
s1 = src;
|
|
s2 = s1 + src_wrap;
|
|
s3 = s2 + src_wrap;
|
|
s4 = s3 + src_wrap;
|
|
d = dst;
|
|
for (w = width; w > 0; w--) {
|
|
d[0] = (s1[0] + s1[1] + s1[2] + s1[3] +
|
|
s2[0] + s2[1] + s2[2] + s2[3] +
|
|
s3[0] + s3[1] + s3[2] + s3[3] +
|
|
s4[0] + s4[1] + s4[2] + s4[3] + 8) >> 4;
|
|
s1 += 4;
|
|
s2 += 4;
|
|
s3 += 4;
|
|
s4 += 4;
|
|
d++;
|
|
}
|
|
src += 4 * src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
static void
|
|
grow21_line (uint8_t * dst, const uint8_t * src, int width)
|
|
{
|
|
int w;
|
|
const uint8_t *s1;
|
|
uint8_t *d;
|
|
|
|
s1 = src;
|
|
d = dst;
|
|
for (w = width; w >= 4; w -= 4) {
|
|
d[1] = d[0] = s1[0];
|
|
d[3] = d[2] = s1[1];
|
|
s1 += 2;
|
|
d += 4;
|
|
}
|
|
for (; w >= 2; w -= 2) {
|
|
d[1] = d[0] = s1[0];
|
|
s1++;
|
|
d += 2;
|
|
}
|
|
/* only needed if width is not a multiple of two */
|
|
/* XXX: veryfy that */
|
|
if (w) {
|
|
d[0] = s1[0];
|
|
}
|
|
}
|
|
|
|
static void
|
|
grow41_line (uint8_t * dst, const uint8_t * src, int width)
|
|
{
|
|
int w, v;
|
|
const uint8_t *s1;
|
|
uint8_t *d;
|
|
|
|
s1 = src;
|
|
d = dst;
|
|
for (w = width; w >= 4; w -= 4) {
|
|
v = s1[0];
|
|
d[0] = v;
|
|
d[1] = v;
|
|
d[2] = v;
|
|
d[3] = v;
|
|
s1++;
|
|
d += 4;
|
|
}
|
|
}
|
|
|
|
/* 1x1 -> 2x1 */
|
|
static void
|
|
grow21 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
for (; height > 0; height--) {
|
|
grow21_line (dst, src, width);
|
|
src += src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 1x1 -> 2x2 */
|
|
static void
|
|
grow22 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
for (; height > 0; height--) {
|
|
grow21_line (dst, src, width);
|
|
if (height % 2)
|
|
src += src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 1x1 -> 4x1 */
|
|
static void
|
|
grow41 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
for (; height > 0; height--) {
|
|
grow41_line (dst, src, width);
|
|
src += src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 1x1 -> 4x4 */
|
|
static void
|
|
grow44 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
for (; height > 0; height--) {
|
|
grow41_line (dst, src, width);
|
|
if ((height & 3) == 1)
|
|
src += src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* 1x2 -> 2x1 */
|
|
static void
|
|
conv411 (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height)
|
|
{
|
|
int w, c;
|
|
const uint8_t *s1, *s2;
|
|
uint8_t *d;
|
|
|
|
width >>= 1;
|
|
|
|
for (; height > 0; height--) {
|
|
s1 = src;
|
|
s2 = src + src_wrap;
|
|
d = dst;
|
|
for (w = width; w > 0; w--) {
|
|
c = (s1[0] + s2[0]) >> 1;
|
|
d[0] = c;
|
|
d[1] = c;
|
|
s1++;
|
|
s2++;
|
|
d += 2;
|
|
}
|
|
src += src_wrap * 2;
|
|
dst += dst_wrap;
|
|
}
|
|
}
|
|
|
|
/* XXX: add jpeg quantize code */
|
|
|
|
#define TRANSP_INDEX (6*6*6)
|
|
|
|
/* this is maybe slow, but allows for extensions */
|
|
static inline unsigned char
|
|
gif_clut_index (uint8_t r, uint8_t g, uint8_t b)
|
|
{
|
|
return ((((r) / 47) % 6) * 6 * 6 + (((g) / 47) % 6) * 6 + (((b) / 47) % 6));
|
|
}
|
|
|
|
static void
|
|
build_rgb_palette (uint8_t * palette, int has_alpha)
|
|
{
|
|
uint32_t *pal;
|
|
static const uint8_t pal_value[6] = { 0x00, 0x33, 0x66, 0x99, 0xcc, 0xff };
|
|
int i, r, g, b;
|
|
|
|
pal = (uint32_t *) palette;
|
|
i = 0;
|
|
for (r = 0; r < 6; r++) {
|
|
for (g = 0; g < 6; g++) {
|
|
for (b = 0; b < 6; b++) {
|
|
pal[i++] = (0xffU << 24) | (pal_value[r] << 16) |
|
|
(pal_value[g] << 8) | pal_value[b];
|
|
}
|
|
}
|
|
}
|
|
if (has_alpha)
|
|
pal[i++] = 0;
|
|
while (i < 256)
|
|
pal[i++] = 0xff000000;
|
|
}
|
|
|
|
/* copy bit n to bits 0 ... n - 1 */
|
|
static inline unsigned int
|
|
bitcopy_n (unsigned int a, int n)
|
|
{
|
|
int mask;
|
|
|
|
mask = (1 << n) - 1;
|
|
return (a & (0xff & ~mask)) | ((-((a >> n) & 1)) & mask);
|
|
}
|
|
|
|
/* rgb555 handling */
|
|
|
|
#define RGB_NAME rgb555
|
|
|
|
#define RGB_IN(r, g, b, s)\
|
|
{\
|
|
unsigned int v = ((const uint16_t *)(s))[0];\
|
|
r = bitcopy_n(v >> (10 - 3), 3);\
|
|
g = bitcopy_n(v >> (5 - 3), 3);\
|
|
b = bitcopy_n(v << 3, 3);\
|
|
}
|
|
|
|
#define RGBA_IN(r, g, b, a, s)\
|
|
{\
|
|
unsigned int v = ((const uint16_t *)(s))[0];\
|
|
r = bitcopy_n(v >> (10 - 3), 3);\
|
|
g = bitcopy_n(v >> (5 - 3), 3);\
|
|
b = bitcopy_n(v << 3, 3);\
|
|
a = (-(v >> 15)) & 0xff;\
|
|
}
|
|
|
|
#define RGBA_OUT(d, r, g, b, a)\
|
|
{\
|
|
((uint16_t *)(d))[0] = ((r >> 3) << 10) | ((g >> 3) << 5) | (b >> 3) | \
|
|
((a << 8) & 0x8000);\
|
|
}
|
|
|
|
#define BPP 2
|
|
|
|
#include "imgconvert_template.h"
|
|
|
|
/* rgb565 handling */
|
|
|
|
#define RGB_NAME rgb565
|
|
|
|
#define RGB_IN(r, g, b, s)\
|
|
{\
|
|
unsigned int v = ((const uint16_t *)(s))[0];\
|
|
r = bitcopy_n(v >> (11 - 3), 3);\
|
|
g = bitcopy_n(v >> (5 - 2), 2);\
|
|
b = bitcopy_n(v << 3, 3);\
|
|
}
|
|
|
|
#define RGB_OUT(d, r, g, b)\
|
|
{\
|
|
((uint16_t *)(d))[0] = ((r >> 3) << 11) | ((g >> 2) << 5) | (b >> 3);\
|
|
}
|
|
|
|
#define BPP 2
|
|
|
|
#include "imgconvert_template.h"
|
|
|
|
/* bgr24 handling */
|
|
|
|
#define RGB_NAME bgr24
|
|
|
|
#define RGB_IN(r, g, b, s)\
|
|
{\
|
|
b = (s)[0];\
|
|
g = (s)[1];\
|
|
r = (s)[2];\
|
|
}
|
|
|
|
#define RGB_OUT(d, r, g, b)\
|
|
{\
|
|
(d)[0] = b;\
|
|
(d)[1] = g;\
|
|
(d)[2] = r;\
|
|
}
|
|
|
|
#define BPP 3
|
|
|
|
#include "imgconvert_template.h"
|
|
|
|
#undef RGB_IN
|
|
#undef RGB_OUT
|
|
#undef BPP
|
|
|
|
/* rgb24 handling */
|
|
|
|
#define RGB_NAME rgb24
|
|
#define FMT_RGB24
|
|
|
|
#define RGB_IN(r, g, b, s)\
|
|
{\
|
|
r = (s)[0];\
|
|
g = (s)[1];\
|
|
b = (s)[2];\
|
|
}
|
|
|
|
#define RGB_OUT(d, r, g, b)\
|
|
{\
|
|
(d)[0] = r;\
|
|
(d)[1] = g;\
|
|
(d)[2] = b;\
|
|
}
|
|
|
|
#define BPP 3
|
|
|
|
#include "imgconvert_template.h"
|
|
|
|
/* rgba32 handling */
|
|
|
|
#define RGB_NAME rgba32
|
|
#define FMT_RGBA32
|
|
|
|
#define RGB_IN(r, g, b, s)\
|
|
{\
|
|
unsigned int v = ((const uint32_t *)(s))[0];\
|
|
r = (v >> 16) & 0xff;\
|
|
g = (v >> 8) & 0xff;\
|
|
b = v & 0xff;\
|
|
}
|
|
|
|
#define RGBA_IN(r, g, b, a, s)\
|
|
{\
|
|
unsigned int v = ((const uint32_t *)(s))[0];\
|
|
a = (v >> 24) & 0xff;\
|
|
r = (v >> 16) & 0xff;\
|
|
g = (v >> 8) & 0xff;\
|
|
b = v & 0xff;\
|
|
}
|
|
|
|
#define RGBA_OUT(d, r, g, b, a)\
|
|
{\
|
|
((uint32_t *)(d))[0] = (a << 24) | (r << 16) | (g << 8) | b;\
|
|
}
|
|
|
|
#define BPP 4
|
|
|
|
#include "imgconvert_template.h"
|
|
|
|
static void
|
|
mono_to_gray (AVPicture * dst, const AVPicture * src,
|
|
int width, int height, int xor_mask)
|
|
{
|
|
const unsigned char *p;
|
|
unsigned char *q;
|
|
int v, dst_wrap, src_wrap;
|
|
int y, w;
|
|
|
|
p = src->data[0];
|
|
src_wrap = src->linesize[0] - ((width + 7) >> 3);
|
|
|
|
q = dst->data[0];
|
|
dst_wrap = dst->linesize[0] - width;
|
|
for (y = 0; y < height; y++) {
|
|
w = width;
|
|
while (w >= 8) {
|
|
v = *p++ ^ xor_mask;
|
|
q[0] = -(v >> 7);
|
|
q[1] = -((v >> 6) & 1);
|
|
q[2] = -((v >> 5) & 1);
|
|
q[3] = -((v >> 4) & 1);
|
|
q[4] = -((v >> 3) & 1);
|
|
q[5] = -((v >> 2) & 1);
|
|
q[6] = -((v >> 1) & 1);
|
|
q[7] = -((v >> 0) & 1);
|
|
w -= 8;
|
|
q += 8;
|
|
}
|
|
if (w > 0) {
|
|
v = *p++ ^ xor_mask;
|
|
do {
|
|
q[0] = -((v >> 7) & 1);
|
|
q++;
|
|
v <<= 1;
|
|
} while (--w);
|
|
}
|
|
p += src_wrap;
|
|
q += dst_wrap;
|
|
}
|
|
}
|
|
|
|
static void
|
|
monowhite_to_gray (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
mono_to_gray (dst, src, width, height, 0xff);
|
|
}
|
|
|
|
static void
|
|
monoblack_to_gray (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
mono_to_gray (dst, src, width, height, 0x00);
|
|
}
|
|
|
|
static void
|
|
gray_to_mono (AVPicture * dst, const AVPicture * src,
|
|
int width, int height, int xor_mask)
|
|
{
|
|
int n;
|
|
const uint8_t *s;
|
|
uint8_t *d;
|
|
int j, b, v, n1, src_wrap, dst_wrap, y;
|
|
|
|
s = src->data[0];
|
|
src_wrap = src->linesize[0] - width;
|
|
|
|
d = dst->data[0];
|
|
dst_wrap = dst->linesize[0] - ((width + 7) >> 3);
|
|
|
|
for (y = 0; y < height; y++) {
|
|
n = width;
|
|
while (n >= 8) {
|
|
v = 0;
|
|
for (j = 0; j < 8; j++) {
|
|
b = s[0];
|
|
s++;
|
|
v = (v << 1) | (b >> 7);
|
|
}
|
|
d[0] = v ^ xor_mask;
|
|
d++;
|
|
n -= 8;
|
|
}
|
|
if (n > 0) {
|
|
n1 = n;
|
|
v = 0;
|
|
while (n > 0) {
|
|
b = s[0];
|
|
s++;
|
|
v = (v << 1) | (b >> 7);
|
|
n--;
|
|
}
|
|
d[0] = (v << (8 - (n1 & 7))) ^ xor_mask;
|
|
d++;
|
|
}
|
|
s += src_wrap;
|
|
d += dst_wrap;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gray_to_monowhite (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
gray_to_mono (dst, src, width, height, 0xff);
|
|
}
|
|
|
|
static void
|
|
gray_to_monoblack (AVPicture * dst, const AVPicture * src,
|
|
int width, int height)
|
|
{
|
|
gray_to_mono (dst, src, width, height, 0x00);
|
|
}
|
|
|
|
typedef void (*convert_func) (AVPicture * dst, const AVPicture * src, int width,
|
|
int height);
|
|
|
|
typedef struct ConvertEntry
|
|
{
|
|
int from;
|
|
int to;
|
|
convert_func convert;
|
|
}
|
|
ConvertEntry;
|
|
|
|
/* Add each new convertion function in this table. In order to be able
|
|
to convert from any format to any format, the following constraints
|
|
must be satisfied:
|
|
|
|
- all FF_COLOR_RGB formats must convert to and from PIX_FMT_RGB24
|
|
|
|
- all FF_COLOR_GRAY formats must convert to and from PIX_FMT_GRAY8
|
|
|
|
- all FF_COLOR_RGB formats with alpha must convert to and from PIX_FMT_RGBA32
|
|
|
|
- PIX_FMT_YUV444P and PIX_FMT_YUVJ444P must convert to and from
|
|
PIX_FMT_RGB24.
|
|
|
|
- PIX_FMT_422 must convert to and from PIX_FMT_422P.
|
|
|
|
The other conversion functions are just optimisations for common cases.
|
|
*/
|
|
static ConvertEntry convert_table[] = {
|
|
{PIX_FMT_YUV420P, PIX_FMT_RGB24, yuv420p_to_rgb24},
|
|
{PIX_FMT_YUV420P, PIX_FMT_BGR24, yuv420p_to_bgr24},
|
|
{PIX_FMT_YUV420P, PIX_FMT_RGBA32, yuv420p_to_rgba32},
|
|
{PIX_FMT_YUV420P, PIX_FMT_RGB565, yuv420p_to_rgb565},
|
|
{PIX_FMT_YUV420P, PIX_FMT_RGB555, yuv420p_to_rgb555},
|
|
|
|
{PIX_FMT_YUV422P, PIX_FMT_YUV422, yuv422p_to_yuv422},
|
|
{PIX_FMT_YUV444P, PIX_FMT_RGB24, yuv444p_to_rgb24},
|
|
{PIX_FMT_YUV422, PIX_FMT_YUV420P, yuv422_to_yuv420p},
|
|
{PIX_FMT_YUV422, PIX_FMT_YUV422P, yuv422_to_yuv422p},
|
|
|
|
{PIX_FMT_YUVJ420P, PIX_FMT_RGB555, yuvj420p_to_rgb555},
|
|
{PIX_FMT_YUVJ420P, PIX_FMT_RGB565, yuvj420p_to_rgb565},
|
|
{PIX_FMT_YUVJ420P, PIX_FMT_BGR24, yuvj420p_to_bgr24},
|
|
{PIX_FMT_YUVJ420P, PIX_FMT_RGB24, yuvj420p_to_rgb24},
|
|
{PIX_FMT_YUVJ420P, PIX_FMT_RGBA32, yuvj420p_to_rgba32},
|
|
{PIX_FMT_YUVJ444P, PIX_FMT_RGB24, yuvj444p_to_rgb24},
|
|
|
|
{PIX_FMT_RGB24, PIX_FMT_YUV420P, rgb24_to_yuv420p},
|
|
{PIX_FMT_RGB24, PIX_FMT_RGB565, rgb24_to_rgb565},
|
|
{PIX_FMT_RGB24, PIX_FMT_RGB555, rgb24_to_rgb555},
|
|
{PIX_FMT_RGB24, PIX_FMT_RGBA32, rgb24_to_rgba32},
|
|
{PIX_FMT_RGB24, PIX_FMT_BGR24, rgb24_to_bgr24},
|
|
{PIX_FMT_RGB24, PIX_FMT_GRAY8, rgb24_to_gray},
|
|
{PIX_FMT_RGB24, PIX_FMT_PAL8, rgb24_to_pal8},
|
|
{PIX_FMT_RGB24, PIX_FMT_YUV444P, rgb24_to_yuv444p},
|
|
{PIX_FMT_RGB24, PIX_FMT_YUVJ420P, rgb24_to_yuvj420p},
|
|
{PIX_FMT_RGB24, PIX_FMT_YUVJ444P, rgb24_to_yuvj444p},
|
|
|
|
{PIX_FMT_RGBA32, PIX_FMT_RGB24, rgba32_to_rgb24},
|
|
{PIX_FMT_RGBA32, PIX_FMT_RGB555, rgba32_to_rgb555},
|
|
{PIX_FMT_RGBA32, PIX_FMT_PAL8, rgba32_to_pal8},
|
|
{PIX_FMT_RGBA32, PIX_FMT_YUV420P, rgba32_to_yuv420p},
|
|
{PIX_FMT_RGBA32, PIX_FMT_GRAY8, rgba32_to_gray},
|
|
|
|
{PIX_FMT_BGR24, PIX_FMT_RGB24, bgr24_to_rgb24},
|
|
{PIX_FMT_BGR24, PIX_FMT_YUV420P, bgr24_to_yuv420p},
|
|
{PIX_FMT_BGR24, PIX_FMT_GRAY8, bgr24_to_gray},
|
|
|
|
{PIX_FMT_RGB555, PIX_FMT_RGB24, rgb555_to_rgb24},
|
|
{PIX_FMT_RGB555, PIX_FMT_RGBA32, rgb555_to_rgba32},
|
|
{PIX_FMT_RGB555, PIX_FMT_YUV420P, rgb555_to_yuv420p},
|
|
{PIX_FMT_RGB555, PIX_FMT_GRAY8, rgb555_to_gray},
|
|
|
|
{PIX_FMT_RGB565, PIX_FMT_RGB24, rgb565_to_rgb24},
|
|
{PIX_FMT_RGB565, PIX_FMT_YUV420P, rgb565_to_yuv420p},
|
|
{PIX_FMT_RGB565, PIX_FMT_GRAY8, rgb565_to_gray},
|
|
|
|
{PIX_FMT_GRAY8, PIX_FMT_RGB555, gray_to_rgb555},
|
|
{PIX_FMT_GRAY8, PIX_FMT_RGB565, gray_to_rgb565},
|
|
{PIX_FMT_GRAY8, PIX_FMT_RGB24, gray_to_rgb24},
|
|
{PIX_FMT_GRAY8, PIX_FMT_BGR24, gray_to_bgr24},
|
|
{PIX_FMT_GRAY8, PIX_FMT_RGBA32, gray_to_rgba32},
|
|
{PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, gray_to_monowhite},
|
|
{PIX_FMT_GRAY8, PIX_FMT_MONOBLACK, gray_to_monoblack},
|
|
|
|
{PIX_FMT_MONOWHITE, PIX_FMT_GRAY8, monowhite_to_gray},
|
|
|
|
{PIX_FMT_MONOBLACK, PIX_FMT_GRAY8, monoblack_to_gray},
|
|
|
|
{PIX_FMT_PAL8, PIX_FMT_RGB555, pal8_to_rgb555},
|
|
{PIX_FMT_PAL8, PIX_FMT_RGB565, pal8_to_rgb565},
|
|
{PIX_FMT_PAL8, PIX_FMT_BGR24, pal8_to_bgr24},
|
|
{PIX_FMT_PAL8, PIX_FMT_RGB24, pal8_to_rgb24},
|
|
{PIX_FMT_PAL8, PIX_FMT_RGBA32, pal8_to_rgba32},
|
|
};
|
|
|
|
static convert_func
|
|
convert_table_lookup (int src_pix_fmt, int dst_pix_fmt)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < sizeof (convert_table) / sizeof (convert_table[0]); i++) {
|
|
if (convert_table[i].from == src_pix_fmt &&
|
|
convert_table[i].to == dst_pix_fmt) {
|
|
return convert_table[i].convert;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
int
|
|
avpicture_alloc (AVPicture * picture, int pix_fmt, int width, int height)
|
|
{
|
|
unsigned int size;
|
|
void *ptr;
|
|
|
|
size = avpicture_get_size (pix_fmt, width, height);
|
|
ptr = av_malloc (size);
|
|
if (!ptr)
|
|
goto fail;
|
|
avpicture_fill (picture, ptr, pix_fmt, width, height);
|
|
return 0;
|
|
fail:
|
|
memset (picture, 0, sizeof (AVPicture));
|
|
return -1;
|
|
}
|
|
|
|
void
|
|
avpicture_free (AVPicture * picture)
|
|
{
|
|
av_free (picture->data[0]);
|
|
}
|
|
|
|
/* return true if yuv planar */
|
|
static inline int
|
|
is_yuv_planar (PixFmtInfo * ps)
|
|
{
|
|
return (ps->color_type == FF_COLOR_YUV ||
|
|
ps->color_type == FF_COLOR_YUV_JPEG) && ps->pixel_type == FF_PIXEL_PLANAR;
|
|
}
|
|
|
|
/* XXX: always use linesize. Return -1 if not supported */
|
|
int
|
|
img_convert (AVPicture * dst, int dst_pix_fmt,
|
|
const AVPicture * src, int src_pix_fmt, int src_width, int src_height)
|
|
{
|
|
static int inited;
|
|
int i, ret, dst_width, dst_height, int_pix_fmt;
|
|
PixFmtInfo *src_pix, *dst_pix;
|
|
convert_func convert;
|
|
AVPicture tmp1, *tmp = &tmp1;
|
|
|
|
if (src_pix_fmt < 0 || src_pix_fmt >= PIX_FMT_NB ||
|
|
dst_pix_fmt < 0 || dst_pix_fmt >= PIX_FMT_NB)
|
|
return -1;
|
|
if (src_width <= 0 || src_height <= 0)
|
|
return 0;
|
|
|
|
if (!inited) {
|
|
inited = 1;
|
|
img_convert_init ();
|
|
}
|
|
|
|
dst_width = src_width;
|
|
dst_height = src_height;
|
|
|
|
dst_pix = &pix_fmt_info[dst_pix_fmt];
|
|
src_pix = &pix_fmt_info[src_pix_fmt];
|
|
|
|
if (src_pix_fmt == dst_pix_fmt) {
|
|
/* no conversion needed: just copy */
|
|
img_copy (dst, src, dst_pix_fmt, dst_width, dst_height);
|
|
return 0;
|
|
}
|
|
|
|
convert = convert_table_lookup (src_pix_fmt, dst_pix_fmt);
|
|
if (convert) {
|
|
/* specific convertion routine */
|
|
convert (dst, src, dst_width, dst_height);
|
|
return 0;
|
|
}
|
|
|
|
/* gray to YUV */
|
|
if (is_yuv_planar (dst_pix) && src_pix_fmt == PIX_FMT_GRAY8) {
|
|
int w, h, y;
|
|
uint8_t *d;
|
|
|
|
if (dst_pix->color_type == FF_COLOR_YUV_JPEG) {
|
|
img_copy_plane (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0], dst_width, dst_height);
|
|
} else {
|
|
img_apply_table (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0],
|
|
dst_width, dst_height, y_jpeg_to_ccir);
|
|
}
|
|
/* fill U and V with 128 */
|
|
w = dst_width;
|
|
h = dst_height;
|
|
w >>= dst_pix->x_chroma_shift;
|
|
h >>= dst_pix->y_chroma_shift;
|
|
for (i = 1; i <= 2; i++) {
|
|
d = dst->data[i];
|
|
for (y = 0; y < h; y++) {
|
|
memset (d, 128, w);
|
|
d += dst->linesize[i];
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* YUV to gray */
|
|
if (is_yuv_planar (src_pix) && dst_pix_fmt == PIX_FMT_GRAY8) {
|
|
if (src_pix->color_type == FF_COLOR_YUV_JPEG) {
|
|
img_copy_plane (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0], dst_width, dst_height);
|
|
} else {
|
|
img_apply_table (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0],
|
|
dst_width, dst_height, y_ccir_to_jpeg);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* YUV to YUV planar */
|
|
if (is_yuv_planar (dst_pix) && is_yuv_planar (src_pix)) {
|
|
int x_shift, y_shift, w, h, xy_shift;
|
|
void (*resize_func) (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src, int src_wrap, int width, int height);
|
|
|
|
/* compute chroma size of the smallest dimensions */
|
|
w = dst_width;
|
|
h = dst_height;
|
|
if (dst_pix->x_chroma_shift >= src_pix->x_chroma_shift)
|
|
w >>= dst_pix->x_chroma_shift;
|
|
else
|
|
w >>= src_pix->x_chroma_shift;
|
|
if (dst_pix->y_chroma_shift >= src_pix->y_chroma_shift)
|
|
h >>= dst_pix->y_chroma_shift;
|
|
else
|
|
h >>= src_pix->y_chroma_shift;
|
|
|
|
x_shift = (dst_pix->x_chroma_shift - src_pix->x_chroma_shift);
|
|
y_shift = (dst_pix->y_chroma_shift - src_pix->y_chroma_shift);
|
|
xy_shift = ((x_shift & 0xf) << 4) | (y_shift & 0xf);
|
|
/* there must be filters for conversion at least from and to
|
|
YUV444 format */
|
|
switch (xy_shift) {
|
|
case 0x00:
|
|
resize_func = img_copy_plane;
|
|
break;
|
|
case 0x10:
|
|
resize_func = shrink21;
|
|
break;
|
|
case 0x20:
|
|
resize_func = shrink41;
|
|
break;
|
|
case 0x01:
|
|
resize_func = shrink12;
|
|
break;
|
|
case 0x11:
|
|
resize_func = shrink22;
|
|
break;
|
|
case 0x22:
|
|
resize_func = shrink44;
|
|
break;
|
|
case 0xf0:
|
|
resize_func = grow21;
|
|
break;
|
|
case 0xe0:
|
|
resize_func = grow41;
|
|
break;
|
|
case 0xff:
|
|
resize_func = grow22;
|
|
break;
|
|
case 0xee:
|
|
resize_func = grow44;
|
|
break;
|
|
case 0xf1:
|
|
resize_func = conv411;
|
|
break;
|
|
default:
|
|
/* currently not handled */
|
|
goto no_chroma_filter;
|
|
}
|
|
|
|
img_copy_plane (dst->data[0], dst->linesize[0],
|
|
src->data[0], src->linesize[0], dst_width, dst_height);
|
|
|
|
for (i = 1; i <= 2; i++)
|
|
resize_func (dst->data[i], dst->linesize[i],
|
|
src->data[i], src->linesize[i],
|
|
dst_width >> dst_pix->x_chroma_shift,
|
|
dst_height >> dst_pix->y_chroma_shift);
|
|
/* if yuv color space conversion is needed, we do it here on
|
|
the destination image */
|
|
if (dst_pix->color_type != src_pix->color_type) {
|
|
const uint8_t *y_table, *c_table;
|
|
|
|
if (dst_pix->color_type == FF_COLOR_YUV) {
|
|
y_table = y_jpeg_to_ccir;
|
|
c_table = c_jpeg_to_ccir;
|
|
} else {
|
|
y_table = y_ccir_to_jpeg;
|
|
c_table = c_ccir_to_jpeg;
|
|
}
|
|
img_apply_table (dst->data[0], dst->linesize[0],
|
|
dst->data[0], dst->linesize[0], dst_width, dst_height, y_table);
|
|
|
|
for (i = 1; i <= 2; i++)
|
|
img_apply_table (dst->data[i], dst->linesize[i],
|
|
dst->data[i], dst->linesize[i],
|
|
dst_width >> dst_pix->x_chroma_shift,
|
|
dst_height >> dst_pix->y_chroma_shift, c_table);
|
|
}
|
|
return 0;
|
|
}
|
|
no_chroma_filter:
|
|
|
|
/* try to use an intermediate format */
|
|
if (src_pix_fmt == PIX_FMT_YUV422 || dst_pix_fmt == PIX_FMT_YUV422) {
|
|
/* specific case: convert to YUV422P first */
|
|
int_pix_fmt = PIX_FMT_YUV422P;
|
|
} else if ((src_pix->color_type == FF_COLOR_GRAY &&
|
|
src_pix_fmt != PIX_FMT_GRAY8) ||
|
|
(dst_pix->color_type == FF_COLOR_GRAY && dst_pix_fmt != PIX_FMT_GRAY8)) {
|
|
/* gray8 is the normalized format */
|
|
int_pix_fmt = PIX_FMT_GRAY8;
|
|
} else if ((is_yuv_planar (src_pix) &&
|
|
src_pix_fmt != PIX_FMT_YUV444P && src_pix_fmt != PIX_FMT_YUVJ444P)) {
|
|
/* yuv444 is the normalized format */
|
|
if (src_pix->color_type == FF_COLOR_YUV_JPEG)
|
|
int_pix_fmt = PIX_FMT_YUVJ444P;
|
|
else
|
|
int_pix_fmt = PIX_FMT_YUV444P;
|
|
} else if ((is_yuv_planar (dst_pix) &&
|
|
dst_pix_fmt != PIX_FMT_YUV444P && dst_pix_fmt != PIX_FMT_YUVJ444P)) {
|
|
/* yuv444 is the normalized format */
|
|
if (dst_pix->color_type == FF_COLOR_YUV_JPEG)
|
|
int_pix_fmt = PIX_FMT_YUVJ444P;
|
|
else
|
|
int_pix_fmt = PIX_FMT_YUV444P;
|
|
} else {
|
|
/* the two formats are rgb or gray8 or yuv[j]444p */
|
|
if (src_pix->is_alpha && dst_pix->is_alpha)
|
|
int_pix_fmt = PIX_FMT_RGBA32;
|
|
else
|
|
int_pix_fmt = PIX_FMT_RGB24;
|
|
}
|
|
if (avpicture_alloc (tmp, int_pix_fmt, dst_width, dst_height) < 0)
|
|
return -1;
|
|
ret = -1;
|
|
if (img_convert (tmp, int_pix_fmt,
|
|
src, src_pix_fmt, src_width, src_height) < 0)
|
|
goto fail1;
|
|
if (img_convert (dst, dst_pix_fmt,
|
|
tmp, int_pix_fmt, dst_width, dst_height) < 0)
|
|
goto fail1;
|
|
ret = 0;
|
|
fail1:
|
|
avpicture_free (tmp);
|
|
return ret;
|
|
}
|
|
|
|
/* NOTE: we scan all the pixels to have an exact information */
|
|
static int
|
|
get_alpha_info_pal8 (const AVPicture * src, int width, int height)
|
|
{
|
|
const unsigned char *p;
|
|
int src_wrap, ret, x, y;
|
|
unsigned int a;
|
|
uint32_t *palette = (uint32_t *) src->data[1];
|
|
|
|
p = src->data[0];
|
|
src_wrap = src->linesize[0] - width;
|
|
ret = 0;
|
|
for (y = 0; y < height; y++) {
|
|
for (x = 0; x < width; x++) {
|
|
a = palette[p[0]] >> 24;
|
|
if (a == 0x00) {
|
|
ret |= FF_ALPHA_TRANSP;
|
|
} else if (a != 0xff) {
|
|
ret |= FF_ALPHA_SEMI_TRANSP;
|
|
}
|
|
p++;
|
|
}
|
|
p += src_wrap;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Tell if an image really has transparent alpha values.
|
|
* @return ored mask of FF_ALPHA_xxx constants
|
|
*/
|
|
int
|
|
img_get_alpha_info (const AVPicture * src, int pix_fmt, int width, int height)
|
|
{
|
|
PixFmtInfo *pf = &pix_fmt_info[pix_fmt];
|
|
int ret;
|
|
|
|
pf = &pix_fmt_info[pix_fmt];
|
|
/* no alpha can be represented in format */
|
|
if (!pf->is_alpha)
|
|
return 0;
|
|
switch (pix_fmt) {
|
|
case PIX_FMT_RGBA32:
|
|
ret = get_alpha_info_rgba32 (src, width, height);
|
|
break;
|
|
case PIX_FMT_RGB555:
|
|
ret = get_alpha_info_rgb555 (src, width, height);
|
|
break;
|
|
case PIX_FMT_PAL8:
|
|
ret = get_alpha_info_pal8 (src, width, height);
|
|
break;
|
|
default:
|
|
/* we do not know, so everything is indicated */
|
|
ret = FF_ALPHA_TRANSP | FF_ALPHA_SEMI_TRANSP;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
#ifdef HAVE_MMX
|
|
#define DEINT_INPLACE_LINE_LUM \
|
|
movd_m2r(lum_m4[0],mm0);\
|
|
movd_m2r(lum_m3[0],mm1);\
|
|
movd_m2r(lum_m2[0],mm2);\
|
|
movd_m2r(lum_m1[0],mm3);\
|
|
movd_m2r(lum[0],mm4);\
|
|
punpcklbw_r2r(mm7,mm0);\
|
|
movd_r2m(mm2,lum_m4[0]);\
|
|
punpcklbw_r2r(mm7,mm1);\
|
|
punpcklbw_r2r(mm7,mm2);\
|
|
punpcklbw_r2r(mm7,mm3);\
|
|
punpcklbw_r2r(mm7,mm4);\
|
|
paddw_r2r(mm3,mm1);\
|
|
psllw_i2r(1,mm2);\
|
|
paddw_r2r(mm4,mm0);\
|
|
psllw_i2r(2,mm1);\
|
|
paddw_r2r(mm6,mm2);\
|
|
paddw_r2r(mm2,mm1);\
|
|
psubusw_r2r(mm0,mm1);\
|
|
psrlw_i2r(3,mm1);\
|
|
packuswb_r2r(mm7,mm1);\
|
|
movd_r2m(mm1,lum_m2[0]);
|
|
|
|
#define DEINT_LINE_LUM \
|
|
movd_m2r(lum_m4[0],mm0);\
|
|
movd_m2r(lum_m3[0],mm1);\
|
|
movd_m2r(lum_m2[0],mm2);\
|
|
movd_m2r(lum_m1[0],mm3);\
|
|
movd_m2r(lum[0],mm4);\
|
|
punpcklbw_r2r(mm7,mm0);\
|
|
punpcklbw_r2r(mm7,mm1);\
|
|
punpcklbw_r2r(mm7,mm2);\
|
|
punpcklbw_r2r(mm7,mm3);\
|
|
punpcklbw_r2r(mm7,mm4);\
|
|
paddw_r2r(mm3,mm1);\
|
|
psllw_i2r(1,mm2);\
|
|
paddw_r2r(mm4,mm0);\
|
|
psllw_i2r(2,mm1);\
|
|
paddw_r2r(mm6,mm2);\
|
|
paddw_r2r(mm2,mm1);\
|
|
psubusw_r2r(mm0,mm1);\
|
|
psrlw_i2r(3,mm1);\
|
|
packuswb_r2r(mm7,mm1);\
|
|
movd_r2m(mm1,dst[0]);
|
|
#endif
|
|
|
|
/* filter parameters: [-1 4 2 4 -1] // 8 */
|
|
static void
|
|
deinterlace_line (uint8_t * dst,
|
|
const uint8_t * lum_m4, const uint8_t * lum_m3,
|
|
const uint8_t * lum_m2, const uint8_t * lum_m1,
|
|
const uint8_t * lum, int size)
|
|
{
|
|
#ifndef HAVE_MMX
|
|
uint8_t *cm = cropTbl + MAX_NEG_CROP;
|
|
int sum;
|
|
|
|
for (; size > 0; size--) {
|
|
sum = -lum_m4[0];
|
|
sum += lum_m3[0] << 2;
|
|
sum += lum_m2[0] << 1;
|
|
sum += lum_m1[0] << 2;
|
|
sum += -lum[0];
|
|
dst[0] = cm[(sum + 4) >> 3];
|
|
lum_m4++;
|
|
lum_m3++;
|
|
lum_m2++;
|
|
lum_m1++;
|
|
lum++;
|
|
dst++;
|
|
}
|
|
#else
|
|
|
|
{
|
|
mmx_t rounder;
|
|
|
|
rounder.uw[0] = 4;
|
|
rounder.uw[1] = 4;
|
|
rounder.uw[2] = 4;
|
|
rounder.uw[3] = 4;
|
|
pxor_r2r (mm7, mm7);
|
|
movq_m2r (rounder, mm6);
|
|
}
|
|
for (; size > 3; size -= 4) {
|
|
DEINT_LINE_LUM lum_m4 += 4;
|
|
|
|
lum_m3 += 4;
|
|
lum_m2 += 4;
|
|
lum_m1 += 4;
|
|
lum += 4;
|
|
dst += 4;
|
|
}
|
|
#endif
|
|
}
|
|
static void
|
|
deinterlace_line_inplace (uint8_t * lum_m4, uint8_t * lum_m3, uint8_t * lum_m2,
|
|
uint8_t * lum_m1, uint8_t * lum, int size)
|
|
{
|
|
#ifndef HAVE_MMX
|
|
uint8_t *cm = cropTbl + MAX_NEG_CROP;
|
|
int sum;
|
|
|
|
for (; size > 0; size--) {
|
|
sum = -lum_m4[0];
|
|
sum += lum_m3[0] << 2;
|
|
sum += lum_m2[0] << 1;
|
|
lum_m4[0] = lum_m2[0];
|
|
sum += lum_m1[0] << 2;
|
|
sum += -lum[0];
|
|
lum_m2[0] = cm[(sum + 4) >> 3];
|
|
lum_m4++;
|
|
lum_m3++;
|
|
lum_m2++;
|
|
lum_m1++;
|
|
lum++;
|
|
}
|
|
#else
|
|
|
|
{
|
|
mmx_t rounder;
|
|
|
|
rounder.uw[0] = 4;
|
|
rounder.uw[1] = 4;
|
|
rounder.uw[2] = 4;
|
|
rounder.uw[3] = 4;
|
|
pxor_r2r (mm7, mm7);
|
|
movq_m2r (rounder, mm6);
|
|
}
|
|
for (; size > 3; size -= 4) {
|
|
DEINT_INPLACE_LINE_LUM lum_m4 += 4;
|
|
|
|
lum_m3 += 4;
|
|
lum_m2 += 4;
|
|
lum_m1 += 4;
|
|
lum += 4;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
/* deinterlacing : 2 temporal taps, 3 spatial taps linear filter. The
|
|
top field is copied as is, but the bottom field is deinterlaced
|
|
against the top field. */
|
|
static void
|
|
deinterlace_bottom_field (uint8_t * dst, int dst_wrap,
|
|
const uint8_t * src1, int src_wrap, int width, int height)
|
|
{
|
|
const uint8_t *src_m2, *src_m1, *src_0, *src_p1, *src_p2;
|
|
int y;
|
|
|
|
src_m2 = src1;
|
|
src_m1 = src1;
|
|
src_0 = &src_m1[src_wrap];
|
|
src_p1 = &src_0[src_wrap];
|
|
src_p2 = &src_p1[src_wrap];
|
|
for (y = 0; y < (height - 2); y += 2) {
|
|
memcpy (dst, src_m1, width);
|
|
dst += dst_wrap;
|
|
deinterlace_line (dst, src_m2, src_m1, src_0, src_p1, src_p2, width);
|
|
src_m2 = src_0;
|
|
src_m1 = src_p1;
|
|
src_0 = src_p2;
|
|
src_p1 += 2 * src_wrap;
|
|
src_p2 += 2 * src_wrap;
|
|
dst += dst_wrap;
|
|
}
|
|
memcpy (dst, src_m1, width);
|
|
dst += dst_wrap;
|
|
/* do last line */
|
|
deinterlace_line (dst, src_m2, src_m1, src_0, src_0, src_0, width);
|
|
}
|
|
|
|
static void
|
|
deinterlace_bottom_field_inplace (uint8_t * src1, int src_wrap,
|
|
int width, int height)
|
|
{
|
|
uint8_t *src_m1, *src_0, *src_p1, *src_p2;
|
|
int y;
|
|
uint8_t *buf;
|
|
|
|
buf = (uint8_t *) av_malloc (width);
|
|
|
|
src_m1 = src1;
|
|
memcpy (buf, src_m1, width);
|
|
src_0 = &src_m1[src_wrap];
|
|
src_p1 = &src_0[src_wrap];
|
|
src_p2 = &src_p1[src_wrap];
|
|
for (y = 0; y < (height - 2); y += 2) {
|
|
deinterlace_line_inplace (buf, src_m1, src_0, src_p1, src_p2, width);
|
|
src_m1 = src_p1;
|
|
src_0 = src_p2;
|
|
src_p1 += 2 * src_wrap;
|
|
src_p2 += 2 * src_wrap;
|
|
}
|
|
/* do last line */
|
|
deinterlace_line_inplace (buf, src_m1, src_0, src_0, src_0, width);
|
|
av_free (buf);
|
|
}
|
|
|
|
|
|
/* deinterlace - if not supported return -1 */
|
|
int
|
|
avpicture_deinterlace (AVPicture * dst, const AVPicture * src,
|
|
int pix_fmt, int width, int height)
|
|
{
|
|
int i;
|
|
|
|
if (pix_fmt != PIX_FMT_YUV420P &&
|
|
pix_fmt != PIX_FMT_YUV422P &&
|
|
pix_fmt != PIX_FMT_YUV444P && pix_fmt != PIX_FMT_YUV411P)
|
|
return -1;
|
|
if ((width & 3) != 0 || (height & 3) != 0)
|
|
return -1;
|
|
|
|
for (i = 0; i < 3; i++) {
|
|
if (i == 1) {
|
|
switch (pix_fmt) {
|
|
case PIX_FMT_YUV420P:
|
|
width >>= 1;
|
|
height >>= 1;
|
|
break;
|
|
case PIX_FMT_YUV422P:
|
|
width >>= 1;
|
|
break;
|
|
case PIX_FMT_YUV411P:
|
|
width >>= 2;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
if (src == dst) {
|
|
deinterlace_bottom_field_inplace (dst->data[i], dst->linesize[i],
|
|
width, height);
|
|
} else {
|
|
deinterlace_bottom_field (dst->data[i], dst->linesize[i],
|
|
src->data[i], src->linesize[i], width, height);
|
|
}
|
|
}
|
|
#ifdef HAVE_MMX
|
|
emms ();
|
|
#endif
|
|
return 0;
|
|
}
|
|
|
|
#undef FIX
|