/* GStreamer * Copyright (C) 1999 Erik Walthinsen * Copyright (C) 2006 Tim-Philipp Müller * Copyright (C) 2010 Sebastian Dröge * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ /** * SECTION:element-videobox * @see_also: #GstVideoCrop * * This plugin crops or enlarges the image. It takes 4 values as input, a * top, bottom, left and right offset. Positive values will crop that much * pixels from the respective border of the image, negative values will add * that much pixels. When pixels are added, you can specify their color. * Some predefined colors are usable with an enum property. * * The plugin is alpha channel aware and will try to negotiate with a format * that supports alpha channels first. When alpha channel is active two * other properties, alpha and border_alpha can be used to set the alpha * values of the inner picture and the border respectively. an alpha value of * 0.0 means total transparency, 1.0 is opaque. * * The videobox plugin has many uses such as doing a mosaic of pictures, * letterboxing video, cutting out pieces of video, picture in picture, etc.. * * Setting autocrop to true changes the behavior of the plugin so that * caps determine crop properties rather than the other way around: given * input and output dimensions, the crop values are selected so that the * smaller frame is effectively centered in the larger frame. This * involves either cropping or padding. * * If you use autocrop there is little point in setting the other * properties manually because they will be overriden if the caps change, * but nothing stops you from doing so. * * Sample pipeline: * |[ * gst-launch videotestsrc ! videobox autocrop=true ! \ * "video/x-raw-yuv, width=600, height=400" ! ffmpegcolorspace ! ximagesink * ]| */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstvideobox.h" #include "gstvideoboxorc.h" #include #include #include #include "gst/glib-compat-private.h" GST_DEBUG_CATEGORY_STATIC (videobox_debug); #define GST_CAT_DEFAULT videobox_debug /* From videotestsrc.c */ static const guint8 yuv_sdtv_colors_Y[VIDEO_BOX_FILL_LAST] = { 16, 145, 41, 81, 210, 235 }; static const guint8 yuv_sdtv_colors_U[VIDEO_BOX_FILL_LAST] = { 128, 54, 240, 90, 16, 128 }; static const guint8 yuv_sdtv_colors_V[VIDEO_BOX_FILL_LAST] = { 128, 34, 110, 240, 146, 128 }; static const guint8 yuv_hdtv_colors_Y[VIDEO_BOX_FILL_LAST] = { 16, 173, 32, 63, 219, 235 }; static const guint8 yuv_hdtv_colors_U[VIDEO_BOX_FILL_LAST] = { 128, 42, 240, 102, 16, 128 }; static const guint8 yuv_hdtv_colors_V[VIDEO_BOX_FILL_LAST] = { 128, 26, 118, 240, 138, 128 }; static const guint8 rgb_colors_R[VIDEO_BOX_FILL_LAST] = { 0, 0, 0, 255, 255, 255 }; static const guint8 rgb_colors_G[VIDEO_BOX_FILL_LAST] = { 0, 255, 0, 0, 255, 255 }; static const guint8 rgb_colors_B[VIDEO_BOX_FILL_LAST] = { 0, 0, 255, 0, 0, 255 }; /* Generated by -bad/ext/cog/generate_tables */ static const int cog_ycbcr_to_rgb_matrix_8bit_hdtv[] = { 298, 0, 459, -63514, 298, -55, -136, 19681, 298, 541, 0, -73988, }; static const int cog_ycbcr_to_rgb_matrix_8bit_sdtv[] = { 298, 0, 409, -57068, 298, -100, -208, 34707, 298, 516, 0, -70870, }; static const gint cog_rgb_to_ycbcr_matrix_8bit_hdtv[] = { 47, 157, 16, 4096, -26, -87, 112, 32768, 112, -102, -10, 32768, }; static const gint cog_rgb_to_ycbcr_matrix_8bit_sdtv[] = { 66, 129, 25, 4096, -38, -74, 112, 32768, 112, -94, -18, 32768, }; static const gint cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit[] = { 256, -30, -53, 10600, 0, 261, 29, -4367, 0, 19, 262, -3289, }; static const gint cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit[] = { 256, 25, 49, -9536, 0, 253, -28, 3958, 0, -19, 252, 2918, }; static const gint cog_identity_matrix_8bit[] = { 256, 0, 0, 0, 0, 256, 0, 0, 0, 0, 256, 0, }; #define APPLY_MATRIX(m,o,v1,v2,v3) ((m[o*4] * v1 + m[o*4+1] * v2 + m[o*4+2] * v3 + m[o*4+3]) >> 8) static void fill_ayuv (GstVideoBoxFill fill_type, guint b_alpha, GstVideoFormat format, guint8 * dest, gboolean sdtv, gint width, gint height) { guint32 empty_pixel; b_alpha = CLAMP (b_alpha, 0, 255); if (sdtv) empty_pixel = GUINT32_FROM_BE ((b_alpha << 24) | (yuv_sdtv_colors_Y[fill_type] << 16) | (yuv_sdtv_colors_U[fill_type] << 8) | yuv_sdtv_colors_V[fill_type]); else empty_pixel = GUINT32_FROM_BE ((b_alpha << 24) | (yuv_hdtv_colors_Y[fill_type] << 16) | (yuv_hdtv_colors_U[fill_type] << 8) | yuv_hdtv_colors_V[fill_type]); orc_splat_u32 ((guint32 *) dest, empty_pixel, width * height); } static void copy_ayuv_ayuv (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; gint src_stride = 4 * src_width; gint dest_stride = 4 * dest_width; dest = dest + dest_y * dest_width * 4 + dest_x * 4; src = src + src_y * src_width * 4 + src_x * 4; w *= 4; if (dest_sdtv != src_sdtv) { gint matrix[12]; gint y, u, v; memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { /* ORC FIXME */ dest[j] = (src[j] * i_alpha) >> 8; y = src[j + 1]; u = src[j + 2]; v = src[j + 3]; dest[j + 1] = APPLY_MATRIX (matrix, 0, y, u, v); dest[j + 2] = APPLY_MATRIX (matrix, 1, y, u, v); dest[j + 3] = APPLY_MATRIX (matrix, 2, y, u, v); } dest += dest_stride; src += src_stride; } } else { for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { /* ORC FIXME */ dest[j] = (src[j] * i_alpha) >> 8; dest[j + 1] = src[j + 1]; dest[j + 2] = src[j + 2]; dest[j + 3] = src[j + 3]; } dest += dest_stride; src += src_stride; } } } static void copy_ayuv_i420 (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; guint8 *destY, *destY2, *destU, *destV; gint dest_strideY, dest_strideUV; const guint8 *src2; gint src_stride; gint y_idx, uv_idx; gint y1, y2, y3, y4; gint u1, u2, u3, u4; gint v1, v2, v3, v4; gint matrix[12]; dest_strideY = gst_video_format_get_row_stride (dest_format, 0, dest_width); dest_strideUV = gst_video_format_get_row_stride (dest_format, 1, dest_width); src_stride = 4 * src_width; destY = dest + gst_video_format_get_component_offset (dest_format, 0, dest_width, dest_height); destU = dest + gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); destV = dest + gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); destY = destY + dest_y * dest_strideY + dest_x; destY2 = (dest_y < dest_height) ? destY + dest_strideY : destY; destU = destU + (dest_y / 2) * dest_strideUV + dest_x / 2; destV = destV + (dest_y / 2) * dest_strideUV + dest_x / 2; src = src + src_y * src_stride + src_x * 4; src2 = (src_y < src_height) ? src + src_stride : src; h = dest_y + h; w = dest_x + w; if (src_sdtv != dest_sdtv) memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); else memcpy (matrix, cog_identity_matrix_8bit, 12 * sizeof (gint)); /* 1. Handle the first destination scanline specially if it * doesn't start at the macro pixel boundary, i.e. blend * with the background! */ if (dest_y % 2 == 1) { /* 1.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = src[4 * 0 + 1]; u1 = src[4 * 0 + 2]; v1 = src[4 * 0 + 3]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ((3 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ((3 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); j = dest_x + 1; y_idx = uv_idx = 1; } else { j = dest_x; y_idx = uv_idx = 0; } /* 1.2. Copy all macro pixels from the source to the destination * but blend with the background because we're only filling * the lower part of the macro pixels. */ for (; j < w - 1; j += 2) { y1 = src[4 * y_idx + 1]; y2 = src[4 * y_idx + 4 + 1]; u1 = src[4 * y_idx + 2]; u2 = src[4 * y_idx + 4 + 2]; v1 = src[4 * y_idx + 3]; v2 = src[4 * y_idx + 4 + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[uv_idx] = CLAMP ( (2 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[uv_idx] = CLAMP ( (2 * destV[uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); y_idx += 2; uv_idx++; } /* 1.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = src[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[uv_idx] = CLAMP ( (destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[uv_idx] = CLAMP ( (destV[uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 1) { y1 = src[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[uv_idx] = CLAMP ( (3 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[uv_idx] = CLAMP ((3 * destV[uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); } destY += dest_strideY; destY2 += dest_strideY; destU += dest_strideUV; destV += dest_strideUV; src += src_stride; src2 += src_stride; i = dest_y + 1; } else { i = dest_y; } /* 2. Copy all macro pixel scanlines, the destination scanline * now starts at macro pixel boundary. */ for (; i < h - 1; i += 2) { /* 2.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = src[4 * 0 + 1]; y2 = src2[4 * 0 + 1]; u1 = src[4 * 0 + 2]; u2 = src2[4 * 0 + 2]; v1 = src[4 * 0 + 3]; v2 = src2[4 * 0 + 3]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY2[0] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[0] = CLAMP ( (2 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[0] = CLAMP ( (2 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); j = dest_x + 1; y_idx = uv_idx = 1; } else { j = dest_x; y_idx = uv_idx = 0; } /* 2.2. Copy all macro pixels from the source to the destination. * All pixels now start at macro pixel boundary, i.e. no * blending with the background is necessary. */ for (; j < w - 1; j += 2) { y1 = src[4 * y_idx + 1]; y2 = src[4 * y_idx + 4 + 1]; y3 = src2[4 * y_idx + 1]; y4 = src2[4 * y_idx + 4 + 1]; u1 = src[4 * y_idx + 2]; u2 = src[4 * y_idx + 4 + 2]; u3 = src2[4 * y_idx + 2]; u4 = src2[4 * y_idx + 4 + 2]; v1 = src[4 * y_idx + 3]; v2 = src[4 * y_idx + 4 + 3]; v3 = src2[4 * y_idx + 3]; v4 = src2[4 * y_idx + 4 + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destY2[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y3, u3, v3), 0, 255); destY2[y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y4, u4, v4), 0, 255); destU[uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2) + APPLY_MATRIX (matrix, 1, y3, u3, v3) + APPLY_MATRIX (matrix, 1, y4, u4, v4)) / 4, 0, 255); destV[uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2) + APPLY_MATRIX (matrix, 2, y3, u3, v3) + APPLY_MATRIX (matrix, 2, y4, u4, v4)) / 4, 0, 255); y_idx += 2; uv_idx++; } /* 2.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = src[4 * y_idx + 1]; y2 = src2[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; u2 = src2[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; v2 = src2[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY2[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 2, 0, 255); destV[uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 2, 0, 255); } else if (j == w - 1) { y1 = src[4 * y_idx + 1]; y2 = src2[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; u2 = src2[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; v2 = src2[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY2[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[uv_idx] = CLAMP ( (2 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); destV[uv_idx] = CLAMP ( (2 * destV[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); } destY += 2 * dest_strideY; destY2 += 2 * dest_strideY; destU += dest_strideUV; destV += dest_strideUV; src += 2 * src_stride; src2 += 2 * src_stride; } /* 3. Handle the last scanline if one exists. This again * doesn't start at macro pixel boundary but should * only fill the upper part of the macro pixels. */ if (i == h - 1 && i == dest_height - 1) { /* 3.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = src[4 * 0 + 1]; u1 = src[4 * 0 + 2]; v1 = src[4 * 0 + 3]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ((destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[0] = CLAMP ((destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 2, 0, 255); j = dest_x + 1; y_idx = uv_idx = 1; } else { j = dest_x; y_idx = uv_idx = 0; } /* 3.2. Copy all macro pixels from the source to the destination * but blend with the background because we're only filling * the upper part of the macro pixels. */ for (; j < w - 1; j += 2) { y1 = src[4 * y_idx + 1]; y2 = src[4 * y_idx + 4 + 1]; u1 = src[4 * y_idx + 2]; u2 = src[4 * y_idx + 4 + 2]; v1 = src[4 * y_idx + 3]; v2 = src[4 * y_idx + 4 + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[uv_idx] = CLAMP ( (2 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[uv_idx] = CLAMP ( (2 * destV[uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); y_idx += 2; uv_idx++; } /* 3.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = src[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[uv_idx] = CLAMP ( (destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[uv_idx] = CLAMP ( (destV[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 1) { y1 = src[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[uv_idx] = CLAMP ( (3 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[uv_idx] = CLAMP ((3 * destV[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); } } else if (i == h - 1) { /* 3.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = src[4 * 0 + 1]; u1 = src[4 * 0 + 2]; v1 = src[4 * 0 + 3]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ((3 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ((3 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); j = dest_x + 1; y_idx = uv_idx = 1; } else { j = dest_x; y_idx = uv_idx = 0; } /* 3.2. Copy all macro pixels from the source to the destination * but blend with the background because we're only filling * the upper part of the macro pixels. */ for (; j < w - 1; j += 2) { y1 = src[4 * y_idx + 1]; y2 = src[4 * y_idx + 4 + 1]; u1 = src[4 * y_idx + 2]; u2 = src[4 * y_idx + 4 + 2]; v1 = src[4 * y_idx + 3]; v2 = src[4 * y_idx + 4 + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[uv_idx] = CLAMP ( (2 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[uv_idx] = CLAMP ( (2 * destV[uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); y_idx += 2; uv_idx++; } /* 3.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = src[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[uv_idx] = CLAMP ( (destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[uv_idx] = CLAMP ( (destV[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 1) { y1 = src[4 * y_idx + 1]; u1 = src[4 * y_idx + 2]; v1 = src[4 * y_idx + 3]; destY[y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[uv_idx] = CLAMP ( (3 * destU[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[uv_idx] = CLAMP ((3 * destV[uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); } } } static void fill_planar_yuv (GstVideoBoxFill fill_type, guint b_alpha, GstVideoFormat format, guint8 * dest, gboolean sdtv, gint width, gint height) { guint8 empty_pixel[3]; guint8 *destY, *destU, *destV; gint strideY, strideUV; gint heightY, heightUV; if (sdtv) { empty_pixel[0] = yuv_sdtv_colors_Y[fill_type]; empty_pixel[1] = yuv_sdtv_colors_U[fill_type]; empty_pixel[2] = yuv_sdtv_colors_V[fill_type]; } else { empty_pixel[0] = yuv_hdtv_colors_Y[fill_type]; empty_pixel[1] = yuv_hdtv_colors_U[fill_type]; empty_pixel[2] = yuv_hdtv_colors_V[fill_type]; } strideY = gst_video_format_get_row_stride (format, 0, width); strideUV = gst_video_format_get_row_stride (format, 1, width); destY = dest + gst_video_format_get_component_offset (format, 0, width, height); destU = dest + gst_video_format_get_component_offset (format, 1, width, height); destV = dest + gst_video_format_get_component_offset (format, 2, width, height); heightY = gst_video_format_get_component_height (format, 0, height); heightUV = gst_video_format_get_component_height (format, 1, height); memset (destY, empty_pixel[0], strideY * heightY); memset (destU, empty_pixel[1], strideUV * heightUV); memset (destV, empty_pixel[2], strideUV * heightUV); } static void copy_y444_y444 (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; guint8 *destY, *destU, *destV; const guint8 *srcY, *srcU, *srcV; gint dest_stride; gint src_stride; dest_stride = gst_video_format_get_row_stride (dest_format, 0, dest_width); src_stride = gst_video_format_get_row_stride (src_format, 0, src_width); destY = dest + gst_video_format_get_component_offset (dest_format, 0, dest_width, dest_height); destU = dest + gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); destV = dest + gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); srcY = src + gst_video_format_get_component_offset (src_format, 0, src_width, src_height); srcU = src + gst_video_format_get_component_offset (src_format, 1, src_width, src_height); srcV = src + gst_video_format_get_component_offset (src_format, 2, src_width, src_height); destY = destY + dest_y * dest_stride + dest_x; destU = destU + dest_y * dest_stride + dest_x; destV = destV + dest_y * dest_stride + dest_x; srcY = srcY + src_y * src_stride + src_x; srcU = srcU + src_y * src_stride + src_x; srcV = srcV + src_y * src_stride + src_x; if (src_sdtv != dest_sdtv) { gint matrix[12]; gint y, u, v; memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); for (i = 0; i < h; i++) { for (j = 0; j < w; j++) { y = APPLY_MATRIX (matrix, 0, srcY[j], srcU[j], srcV[j]); u = APPLY_MATRIX (matrix, 1, srcY[j], srcU[j], srcV[j]); v = APPLY_MATRIX (matrix, 2, srcY[j], srcU[j], srcV[j]); destY[j] = y; destU[j] = u; destV[j] = v; } destY += dest_stride; destU += dest_stride; destV += dest_stride; srcY += src_stride; srcU += src_stride; srcV += src_stride; } } else { for (i = 0; i < h; i++) { memcpy (destY, srcY, w); memcpy (destU, srcU, w); memcpy (destV, srcV, w); destY += dest_stride; destU += dest_stride; destV += dest_stride; srcY += src_stride; srcU += src_stride; srcV += src_stride; } } } static void copy_y42b_y42b (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; guint8 *destY, *destU, *destV; const guint8 *srcY, *srcU, *srcV; gint dest_strideY, dest_strideUV; gint src_strideY, src_strideUV; gint src_y_idx, src_uv_idx; gint dest_y_idx, dest_uv_idx; gint matrix[12]; gint y1, y2; gint u1, u2; gint v1, v2; dest_strideY = gst_video_format_get_row_stride (dest_format, 0, dest_width); dest_strideUV = gst_video_format_get_row_stride (dest_format, 1, dest_width); src_strideY = gst_video_format_get_row_stride (src_format, 0, src_width); src_strideUV = gst_video_format_get_row_stride (src_format, 1, src_width); destY = dest + gst_video_format_get_component_offset (dest_format, 0, dest_width, dest_height); destU = dest + gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); destV = dest + gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); srcY = src + gst_video_format_get_component_offset (src_format, 0, src_width, src_height); srcU = src + gst_video_format_get_component_offset (src_format, 1, src_width, src_height); srcV = src + gst_video_format_get_component_offset (src_format, 2, src_width, src_height); destY = destY + dest_y * dest_strideY + dest_x; destU = destU + dest_y * dest_strideUV + dest_x / 2; destV = destV + dest_y * dest_strideUV + dest_x / 2; srcY = srcY + src_y * src_strideY + src_x; srcU = srcU + src_y * src_strideUV + src_x / 2; srcV = srcV + src_y * src_strideUV + src_x / 2; h = dest_y + h; w = dest_x + w; if (src_sdtv != dest_sdtv) memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); else memcpy (matrix, cog_identity_matrix_8bit, 12 * sizeof (gint)); /* 1. Copy all macro pixel scanlines, the destination scanline * now starts at macro pixel boundary. */ for (i = dest_y; i < h; i++) { /* 1.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = srcY[0]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ( (destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[0] = CLAMP ( (destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 2, 0, 255); j = dest_x + 1; src_y_idx = dest_y_idx = dest_uv_idx = 1; src_uv_idx = (src_x % 2) + 1; } else { j = dest_x; src_y_idx = dest_y_idx = dest_uv_idx = 0; src_uv_idx = (src_x % 2); } /* 1.2. Copy all macro pixels from the source to the destination. * All pixels now start at macro pixel boundary, i.e. no * blending with the background is necessary. */ for (; j < w - 1; j += 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; src_uv_idx++; u2 = srcU[src_uv_idx / 2]; v2 = srcV[src_uv_idx / 2]; src_uv_idx++; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 2, 0, 255); dest_y_idx += 2; src_y_idx += 2; dest_uv_idx++; } /* 1.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); destV[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); } else if (j == w - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ( (destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); } destY += dest_strideY; destU += dest_strideUV; destV += dest_strideUV; srcY += src_strideY; srcU += src_strideUV; srcV += src_strideUV; } } static void copy_y41b_y41b (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; guint8 *destY, *destU, *destV; const guint8 *srcY, *srcU, *srcV; gint dest_strideY, dest_strideUV; gint src_strideY, src_strideUV; gint src_y_idx, src_uv_idx; gint dest_y_idx, dest_uv_idx; gint matrix[12]; gint y1, y2, y3, y4; gint u1, u2, u3, u4; gint v1, v2, v3, v4; dest_strideY = gst_video_format_get_row_stride (dest_format, 0, dest_width); dest_strideUV = gst_video_format_get_row_stride (dest_format, 1, dest_width); src_strideY = gst_video_format_get_row_stride (src_format, 0, src_width); src_strideUV = gst_video_format_get_row_stride (src_format, 1, src_width); destY = dest + gst_video_format_get_component_offset (dest_format, 0, dest_width, dest_height); destU = dest + gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); destV = dest + gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); srcY = src + gst_video_format_get_component_offset (src_format, 0, src_width, src_height); srcU = src + gst_video_format_get_component_offset (src_format, 1, src_width, src_height); srcV = src + gst_video_format_get_component_offset (src_format, 2, src_width, src_height); destY = destY + dest_y * dest_strideY + dest_x; destU = destU + dest_y * dest_strideUV + dest_x / 4; destV = destV + dest_y * dest_strideUV + dest_x / 4; srcY = srcY + src_y * src_strideY + src_x; srcU = srcU + src_y * src_strideUV + src_x / 4; srcV = srcV + src_y * src_strideUV + src_x / 4; h = dest_y + h; w = dest_x + w; if (src_sdtv != dest_sdtv) memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); else memcpy (matrix, cog_identity_matrix_8bit, 12 * sizeof (gint)); /* 1. Copy all macro pixel scanlines, the destination scanline * now starts at macro pixel boundary. */ for (i = dest_y; i < h; i++) { /* 1.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 4 == 1) { y1 = srcY[0]; y2 = srcY[1]; y3 = srcY[2]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u1, v1), 0, 255); destY[2] = CLAMP (APPLY_MATRIX (matrix, 0, y3, u1, v1), 0, 255); destU[0] = CLAMP ( (destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u1, v1) + APPLY_MATRIX (matrix, 1, y3, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ((destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u1, v1) + APPLY_MATRIX (matrix, 2, y3, u1, v1)) / 4, 0, 255); j = dest_x + 3; src_y_idx = dest_y_idx = 3; dest_uv_idx = 1; src_uv_idx = (src_x % 4) + 3; } else if (dest_x % 4 == 2) { y1 = srcY[0]; y2 = srcY[1]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u1, v1), 0, 255); destU[0] = CLAMP ( (2 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ((2 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u1, v1)) / 4, 0, 255); j = dest_x + 2; src_y_idx = dest_y_idx = 2; dest_uv_idx = 1; src_uv_idx = (src_x % 4) + 2; } else if (dest_x % 4 == 3) { y1 = srcY[0]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ( (3 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ( (3 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); j = dest_x + 1; src_y_idx = dest_y_idx = 1; dest_uv_idx = 1; src_uv_idx = (src_x % 4) + 1; } else { j = dest_x; src_y_idx = dest_y_idx = dest_uv_idx = 0; src_uv_idx = (src_x % 4); } /* 1.2. Copy all macro pixels from the source to the destination. * All pixels now start at macro pixel boundary, i.e. no * blending with the background is necessary. */ for (; j < w - 3; j += 4) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; y3 = srcY[src_y_idx + 2]; y4 = srcY[src_y_idx + 3]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; src_uv_idx++; u2 = srcU[src_uv_idx / 4]; v2 = srcV[src_uv_idx / 4]; src_uv_idx++; u3 = srcU[src_uv_idx / 4]; v3 = srcV[src_uv_idx / 4]; src_uv_idx++; u4 = srcU[src_uv_idx / 4]; v4 = srcV[src_uv_idx / 4]; src_uv_idx++; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destY[dest_y_idx + 2] = CLAMP (APPLY_MATRIX (matrix, 0, y3, u3, v3), 0, 255); destY[dest_y_idx + 3] = CLAMP (APPLY_MATRIX (matrix, 0, y4, u4, v4), 0, 255); destU[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2) + APPLY_MATRIX (matrix, 1, y3, u3, v3) + APPLY_MATRIX (matrix, 1, y4, u4, v4)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ((APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2) + APPLY_MATRIX (matrix, 2, y3, u3, v3) + APPLY_MATRIX (matrix, 2, y4, u4, v4)) / 4, 0, 255); dest_y_idx += 4; src_y_idx += 4; dest_uv_idx++; } /* 1.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); destV[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); } else if (j == w - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (destU[dest_uv_idx] + 3 * APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ( (destV[dest_uv_idx] + 3 * APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); } else if (j == w - 2 && j == dest_width - 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); destV[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); } else if (j == w - 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ((destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ((destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 3 && j == dest_width - 3) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; y3 = srcY[src_y_idx + 2]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u1, v1), 0, 255); destY[dest_y_idx + 2] = CLAMP (APPLY_MATRIX (matrix, 0, y3, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); destV[dest_uv_idx] = CLAMP (APPLY_MATRIX (matrix, 1, y1, u1, v1), 0, 255); } else if (j == w - 3) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; y3 = srcY[src_y_idx + 2]; u1 = srcU[src_uv_idx / 4]; v1 = srcV[src_uv_idx / 4]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u1, v1), 0, 255); destY[dest_y_idx + 2] = CLAMP (APPLY_MATRIX (matrix, 0, y3, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ((3 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ((3 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); } destY += dest_strideY; destU += dest_strideUV; destV += dest_strideUV; srcY += src_strideY; srcU += src_strideUV; srcV += src_strideUV; } } static void copy_i420_i420 (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; guint8 *destY, *destU, *destV; const guint8 *srcY, *srcU, *srcV; guint8 *destY2; const guint8 *srcY2, *srcU2, *srcV2; gint dest_strideY, dest_strideUV; gint src_strideY, src_strideUV; gint src_y_idx, src_uv_idx; gint dest_y_idx, dest_uv_idx; gint matrix[12]; gint y1, y2, y3, y4; gint u1, u2, u3, u4; gint v1, v2, v3, v4; dest_strideY = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 0, dest_width); dest_strideUV = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 1, dest_width); src_strideY = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 0, src_width); src_strideUV = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 1, src_width); destY = dest + gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 0, dest_width, dest_height); destU = dest + gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); destV = dest + gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); srcY = src + gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 0, src_width, src_height); srcU = src + gst_video_format_get_component_offset (src_format, 1, src_width, src_height); srcV = src + gst_video_format_get_component_offset (src_format, 2, src_width, src_height); destY = destY + dest_y * dest_strideY + dest_x; destU = destU + (dest_y / 2) * dest_strideUV + dest_x / 2; destV = destV + (dest_y / 2) * dest_strideUV + dest_x / 2; srcY = srcY + src_y * src_strideY + src_x; srcU = srcU + (src_y / 2) * src_strideUV + src_x / 2; srcV = srcV + (src_y / 2) * src_strideUV + src_x / 2; destY2 = destY + dest_strideY; srcY2 = srcY + src_strideY; h = dest_y + h; w = dest_x + w; if (src_sdtv != dest_sdtv) memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); else memcpy (matrix, cog_identity_matrix_8bit, 12 * sizeof (gint)); /* 1. Handle the first destination scanline specially if it * doesn't start at the macro pixel boundary, i.e. blend * with the background! */ if (dest_y % 2 == 1) { /* 1.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = srcY[0]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ((3 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ((3 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); j = dest_x + 1; src_y_idx = dest_y_idx = dest_uv_idx = 1; src_uv_idx = (src_x % 2) + 1; } else { j = dest_x; src_y_idx = dest_y_idx = dest_uv_idx = 0; src_uv_idx = (src_x % 2); } /* 1.2. Copy all macro pixels from the source to the destination * but blend with the background because we're only filling * the lower part of the macro pixels. */ for (; j < w - 1; j += 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; src_uv_idx++; u2 = srcU[src_uv_idx / 2]; v2 = srcV[src_uv_idx / 2]; src_uv_idx++; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[dest_uv_idx] = CLAMP ((2 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ((2 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); dest_y_idx += 2; src_y_idx += 2; dest_uv_idx++; } /* 1.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ((destV[dest_uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (3 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ((3 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); } destY += dest_strideY; destY2 += dest_strideY; destU += dest_strideUV; destV += dest_strideUV; srcY += src_strideY; srcY2 += src_strideY; src_y++; if (src_y % 2 == 0) { srcU += src_strideUV; srcV += src_strideUV; } i = dest_y + 1; } else { i = dest_y; } /* 2. Copy all macro pixel scanlines, the destination scanline * now starts at macro pixel boundary. */ for (; i < h - 1; i += 2) { /* 2.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ srcU2 = srcU; srcV2 = srcV; if (src_y % 2 == 1) { srcU2 += src_strideUV; srcV2 += src_strideUV; } if (dest_x % 2 == 1) { y1 = srcY[0]; y2 = srcY2[0]; u1 = srcU[0]; v1 = srcV[0]; u2 = srcU2[0]; v2 = srcV2[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY2[0] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[0] = CLAMP ( (2 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[0] = CLAMP ( (2 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); j = dest_x + 1; src_y_idx = dest_y_idx = dest_uv_idx = 1; src_uv_idx = (src_x % 2) + 1; } else { j = dest_x; src_y_idx = dest_y_idx = dest_uv_idx = 0; src_uv_idx = (src_x % 2); } /* 2.2. Copy all macro pixels from the source to the destination. * All pixels now start at macro pixel boundary, i.e. no * blending with the background is necessary. */ for (; j < w - 1; j += 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; y3 = srcY2[src_y_idx]; y4 = srcY2[src_y_idx + 1]; u1 = srcU[src_uv_idx / 2]; u3 = srcU2[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; v3 = srcV2[src_uv_idx / 2]; src_uv_idx++; u2 = srcU[src_uv_idx / 2]; u4 = srcU2[src_uv_idx / 2]; v2 = srcV[src_uv_idx / 2]; v4 = srcV2[src_uv_idx / 2]; src_uv_idx++; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destY2[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y3, u3, v3), 0, 255); destY2[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y4, u4, v4), 0, 255); destU[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2) + APPLY_MATRIX (matrix, 1, y3, u3, v3) + APPLY_MATRIX (matrix, 1, y4, u4, v4)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2) + APPLY_MATRIX (matrix, 2, y3, u3, v3) + APPLY_MATRIX (matrix, 2, y4, u4, v4)) / 4, 0, 255); dest_y_idx += 2; src_y_idx += 2; dest_uv_idx++; } /* 2.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = srcY[src_y_idx]; y2 = srcY2[src_y_idx]; u1 = srcU[src_uv_idx / 2]; u2 = srcU2[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; v2 = srcV2[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY2[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ( (APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 2, 0, 255); } else if (j == w - 1) { y1 = srcY[src_y_idx]; y2 = srcY2[src_y_idx]; u1 = srcU[src_uv_idx / 2]; u2 = srcU2[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; v2 = srcV2[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY2[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[dest_uv_idx] = CLAMP ( (2 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ( (2 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); } destY += 2 * dest_strideY; destY2 += 2 * dest_strideY; destU += dest_strideUV; destV += dest_strideUV; srcY += 2 * src_strideY; srcY2 += 2 * src_strideY; src_y += 2; srcU += src_strideUV; srcV += src_strideUV; } /* 3. Handle the last scanline if one exists. This again * doesn't start at macro pixel boundary but should * only fill the upper part of the macro pixels. */ if (i == h - 1 && i == dest_height - 1) { /* 3.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = srcY[0]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ((destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[0] = CLAMP ((destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 2, 0, 255); j = dest_x + 1; src_y_idx = dest_y_idx = dest_uv_idx = 1; src_uv_idx = (src_x % 2) + 1; } else { j = dest_x; src_y_idx = dest_y_idx = dest_uv_idx = 0; src_uv_idx = (src_x % 2); } /* 3.2. Copy all macro pixels from the source to the destination * but blend with the background because we're only filling * the upper part of the macro pixels. */ for (; j < w - 1; j += 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; src_uv_idx++; u2 = srcU[src_uv_idx / 2]; v2 = srcV[src_uv_idx / 2]; src_uv_idx++; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[dest_uv_idx] = CLAMP ( (2 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ( (2 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); dest_y_idx += 2; src_y_idx += 2; dest_uv_idx++; } /* 3.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ((destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (3 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ((3 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); } } else if (i == h - 1) { /* 3.1. Handle the first destination pixel if it doesn't * start at the macro pixel boundary, i.e. blend with * the background! */ if (dest_x % 2 == 1) { y1 = srcY[0]; u1 = srcU[0]; v1 = srcV[0]; destY[0] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[0] = CLAMP ((3 * destU[0] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[0] = CLAMP ((3 * destV[0] + APPLY_MATRIX (matrix, 2, y1, u1, v1)) / 4, 0, 255); j = dest_x + 1; src_y_idx = dest_y_idx = dest_uv_idx = 1; src_uv_idx = (src_x % 2) + 1; } else { j = dest_x; src_y_idx = dest_y_idx = dest_uv_idx = 0; src_uv_idx = (src_x % 2); } /* 3.2. Copy all macro pixels from the source to the destination * but blend with the background because we're only filling * the upper part of the macro pixels. */ for (; j < w - 1; j += 2) { y1 = srcY[src_y_idx]; y2 = srcY[src_y_idx + 1]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; src_uv_idx++; u2 = srcU[src_uv_idx / 2]; v2 = srcV[src_uv_idx / 2]; src_uv_idx++; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destY[dest_y_idx + 1] = CLAMP (APPLY_MATRIX (matrix, 0, y2, u2, v2), 0, 255); destU[dest_uv_idx] = CLAMP ( (2 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1) + APPLY_MATRIX (matrix, 1, y2, u2, v2)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ( (2 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 2, y1, u1, v1) + APPLY_MATRIX (matrix, 2, y2, u2, v2)) / 4, 0, 255); dest_y_idx += 2; src_y_idx += 2; dest_uv_idx++; } /* 3.3. Now copy the last pixel if one exists and blend it * with the background because we only fill part of * the macro pixel. In case this is the last pixel of * the destination we will a larger part. */ if (j == w - 1 && j == dest_width - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); destV[dest_uv_idx] = CLAMP ((destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 2, 0, 255); } else if (j == w - 1) { y1 = srcY[src_y_idx]; u1 = srcU[src_uv_idx / 2]; v1 = srcV[src_uv_idx / 2]; destY[dest_y_idx] = CLAMP (APPLY_MATRIX (matrix, 0, y1, u1, v1), 0, 255); destU[dest_uv_idx] = CLAMP ( (3 * destU[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); destV[dest_uv_idx] = CLAMP ((3 * destV[dest_uv_idx] + APPLY_MATRIX (matrix, 1, y1, u1, v1)) / 4, 0, 255); } } } static void copy_i420_ayuv (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { const guint8 *srcY, *srcU, *srcV; gint src_strideY, src_strideUV; gint dest_stride; src_strideY = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 0, src_width); src_strideUV = gst_video_format_get_row_stride (GST_VIDEO_FORMAT_I420, 1, src_width); srcY = src + gst_video_format_get_component_offset (GST_VIDEO_FORMAT_I420, 0, src_width, src_height); srcU = src + gst_video_format_get_component_offset (src_format, 1, src_width, src_height); srcV = src + gst_video_format_get_component_offset (src_format, 2, src_width, src_height); dest_stride = dest_width * 4; dest = dest + dest_y * dest_stride + dest_x * 4; srcY = srcY + src_y * src_strideY + src_x; srcU = srcU + (src_y / 2) * src_strideUV + src_x / 2; srcV = srcV + (src_y / 2) * src_strideUV + src_x / 2; i_alpha = CLAMP (i_alpha, 0, 255); if (src_sdtv != dest_sdtv) { gint i, j, uv_idx; gint y, u, v; gint y1, u1, v1; gint matrix[12]; memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); for (i = 0; i < h; i++) { for (j = 0, uv_idx = src_x % 2; j < w; j++, uv_idx++) { y = srcY[j]; u = srcU[uv_idx / 2]; v = srcV[uv_idx / 2]; y1 = APPLY_MATRIX (matrix, 0, y, u, v); u1 = APPLY_MATRIX (matrix, 1, y, u, v); v1 = APPLY_MATRIX (matrix, 2, y, u, v); dest[4 * j + 0] = i_alpha; dest[4 * j + 1] = y1; dest[4 * j + 2] = u1; dest[4 * j + 3] = v1; } dest += dest_stride; src_y++; srcY += src_strideY; if (src_y % 2 == 0) { srcU += src_strideUV; srcV += src_strideUV; } } } else { gint i, j, uv_idx; gint y, u, v; for (i = 0; i < h; i++) { for (j = 0, uv_idx = src_x % 2; j < w; j++, uv_idx++) { y = srcY[j]; u = srcU[uv_idx / 2]; v = srcV[uv_idx / 2]; dest[4 * j + 0] = i_alpha; dest[4 * j + 1] = y; dest[4 * j + 2] = u; dest[4 * j + 3] = v; } dest += dest_stride; src_y++; srcY += src_strideY; if (src_y % 2 == 0) { srcU += src_strideUV; srcV += src_strideUV; } } } } static void fill_rgb32 (GstVideoBoxFill fill_type, guint b_alpha, GstVideoFormat format, guint8 * dest, gboolean sdtv, gint width, gint height) { guint32 empty_pixel; gint p[4]; p[0] = gst_video_format_get_component_offset (format, 3, width, height); p[1] = gst_video_format_get_component_offset (format, 0, width, height); p[2] = gst_video_format_get_component_offset (format, 1, width, height); p[3] = gst_video_format_get_component_offset (format, 2, width, height); b_alpha = CLAMP (b_alpha, 0, 255); empty_pixel = GUINT32_FROM_LE ((b_alpha << (p[0] * 8)) | (rgb_colors_R[fill_type] << (p[1] * 8)) | (rgb_colors_G[fill_type] << (p[2] * 8)) | (rgb_colors_B[fill_type] << (p[3] * 8))); orc_splat_u32 ((guint32 *) dest, empty_pixel, width * height); } static void fill_rgb24 (GstVideoBoxFill fill_type, guint b_alpha, GstVideoFormat format, guint8 * dest, gboolean sdtv, gint width, gint height) { gint dest_stride = GST_ROUND_UP_4 (width * 3); gint p[4]; gint i, j; p[0] = gst_video_format_get_component_offset (format, 3, width, height); p[1] = gst_video_format_get_component_offset (format, 0, width, height); p[2] = gst_video_format_get_component_offset (format, 1, width, height); p[3] = gst_video_format_get_component_offset (format, 2, width, height); for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { dest[3 * j + p[1]] = rgb_colors_R[fill_type]; dest[3 * j + p[2]] = rgb_colors_G[fill_type]; dest[3 * j + p[3]] = rgb_colors_B[fill_type]; } dest += dest_stride; } } static void copy_rgb32 (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; gint src_stride, dest_stride; gboolean in_alpha, out_alpha; gint in_bpp, out_bpp; gint p_out[4]; gint p_in[4]; gboolean packed_out = (dest_format == GST_VIDEO_FORMAT_RGB || dest_format == GST_VIDEO_FORMAT_BGR); gboolean packed_in = (src_format == GST_VIDEO_FORMAT_RGB || src_format == GST_VIDEO_FORMAT_BGR); src_stride = (packed_in) ? GST_ROUND_UP_4 (3 * src_width) : 4 * src_width; dest_stride = (packed_out) ? GST_ROUND_UP_4 (3 * dest_width) : 4 * dest_width; in_bpp = (packed_in) ? 3 : 4; out_bpp = (packed_out) ? 3 : 4; out_alpha = gst_video_format_has_alpha (dest_format); p_out[0] = gst_video_format_get_component_offset (dest_format, 3, dest_width, dest_height); p_out[1] = gst_video_format_get_component_offset (dest_format, 0, dest_width, dest_height); p_out[2] = gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); p_out[3] = gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); in_alpha = gst_video_format_has_alpha (src_format); p_in[0] = gst_video_format_get_component_offset (src_format, 3, src_width, src_height); p_in[1] = gst_video_format_get_component_offset (src_format, 0, src_width, src_height); p_in[2] = gst_video_format_get_component_offset (src_format, 1, src_width, src_height); p_in[3] = gst_video_format_get_component_offset (src_format, 2, src_width, src_height); dest = dest + dest_y * dest_stride + dest_x * out_bpp; src = src + src_y * src_stride + src_x * in_bpp; if (in_alpha && out_alpha) { w *= 4; for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { dest[j + p_out[0]] = (src[j + p_in[0]] * i_alpha) >> 8; dest[j + p_out[1]] = src[j + p_in[1]]; dest[j + p_out[2]] = src[j + p_in[2]]; dest[j + p_out[3]] = src[j + p_in[3]]; } dest += dest_stride; src += src_stride; } } else if (out_alpha && !packed_in) { w *= 4; i_alpha = CLAMP (i_alpha, 0, 255); for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { dest[j + p_out[0]] = i_alpha; dest[j + p_out[1]] = src[j + p_in[1]]; dest[j + p_out[2]] = src[j + p_in[2]]; dest[j + p_out[3]] = src[j + p_in[3]]; } dest += dest_stride; src += src_stride; } } else if (out_alpha && packed_in) { i_alpha = CLAMP (i_alpha, 0, 255); for (i = 0; i < h; i++) { for (j = 0; j < w; j++) { dest[4 * j + p_out[0]] = i_alpha; dest[4 * j + p_out[1]] = src[in_bpp * j + p_in[1]]; dest[4 * j + p_out[2]] = src[in_bpp * j + p_in[2]]; dest[4 * j + p_out[3]] = src[in_bpp * j + p_in[3]]; } dest += dest_stride; src += src_stride; } } else if (!packed_out && !packed_in) { w *= 4; for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { dest[j + p_out[1]] = src[j + p_in[1]]; dest[j + p_out[2]] = src[j + p_in[2]]; dest[j + p_out[3]] = src[j + p_in[3]]; } dest += dest_stride; src += src_stride; } } else { for (i = 0; i < h; i++) { for (j = 0; j < w; j++) { dest[out_bpp * j + p_out[1]] = src[in_bpp * j + p_in[1]]; dest[out_bpp * j + p_out[2]] = src[in_bpp * j + p_in[2]]; dest[out_bpp * j + p_out[3]] = src[in_bpp * j + p_in[3]]; } dest += dest_stride; src += src_stride; } } } static void copy_rgb32_ayuv (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; gint src_stride, dest_stride; gboolean in_alpha; gint in_bpp; gint p_in[4]; gboolean packed_in = (src_format == GST_VIDEO_FORMAT_RGB || src_format == GST_VIDEO_FORMAT_BGR); gint matrix[12]; gint a; gint y, u, v; gint r, g, b; src_stride = (packed_in) ? GST_ROUND_UP_4 (3 * src_width) : 4 * src_width; dest_stride = 4 * dest_width; in_bpp = (packed_in) ? 3 : 4; in_alpha = gst_video_format_has_alpha (src_format); p_in[0] = gst_video_format_get_component_offset (src_format, 3, src_width, src_height); p_in[1] = gst_video_format_get_component_offset (src_format, 0, src_width, src_height); p_in[2] = gst_video_format_get_component_offset (src_format, 1, src_width, src_height); p_in[3] = gst_video_format_get_component_offset (src_format, 2, src_width, src_height); memcpy (matrix, (dest_sdtv) ? cog_rgb_to_ycbcr_matrix_8bit_sdtv : cog_rgb_to_ycbcr_matrix_8bit_hdtv, 12 * sizeof (gint)); dest = dest + dest_y * dest_stride + dest_x * 4; src = src + src_y * src_stride + src_x * in_bpp; if (in_alpha) { w *= 4; for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { a = (src[j + p_in[0]] * i_alpha) >> 8; r = src[j + p_in[1]]; g = src[j + p_in[2]]; b = src[j + p_in[3]]; y = APPLY_MATRIX (matrix, 0, r, g, b); u = APPLY_MATRIX (matrix, 1, r, g, b); v = APPLY_MATRIX (matrix, 2, r, g, b); dest[j + 0] = a; dest[j + 1] = CLAMP (y, 0, 255); dest[j + 2] = CLAMP (u, 0, 255); dest[j + 3] = CLAMP (v, 0, 255); } dest += dest_stride; src += src_stride; } } else if (!packed_in) { w *= 4; i_alpha = CLAMP (i_alpha, 0, 255); for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { a = i_alpha; r = src[j + p_in[1]]; g = src[j + p_in[2]]; b = src[j + p_in[3]]; y = APPLY_MATRIX (matrix, 0, r, g, b); u = APPLY_MATRIX (matrix, 1, r, g, b); v = APPLY_MATRIX (matrix, 2, r, g, b); dest[j + 0] = a; dest[j + 1] = CLAMP (y, 0, 255); dest[j + 2] = CLAMP (u, 0, 255); dest[j + 3] = CLAMP (v, 0, 255); } dest += dest_stride; src += src_stride; } } else { i_alpha = CLAMP (i_alpha, 0, 255); for (i = 0; i < h; i++) { for (j = 0; j < w; j++) { a = i_alpha; r = src[in_bpp * j + p_in[1]]; g = src[in_bpp * j + p_in[2]]; b = src[in_bpp * j + p_in[3]]; y = APPLY_MATRIX (matrix, 0, r, g, b); u = APPLY_MATRIX (matrix, 1, r, g, b); v = APPLY_MATRIX (matrix, 2, r, g, b); dest[4 * j + 0] = a; dest[4 * j + 1] = CLAMP (y, 0, 255); dest[4 * j + 2] = CLAMP (u, 0, 255); dest[4 * j + 3] = CLAMP (v, 0, 255); } dest += dest_stride; src += src_stride; } } } static void copy_ayuv_rgb32 (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; gint src_stride, dest_stride; gboolean out_alpha; gint out_bpp; gint p_out[4]; gboolean packed_out = (dest_format == GST_VIDEO_FORMAT_RGB || dest_format == GST_VIDEO_FORMAT_BGR); gint matrix[12]; gint a; gint y, u, v; gint r, g, b; dest_stride = (packed_out) ? GST_ROUND_UP_4 (3 * dest_width) : 4 * dest_width; src_stride = 4 * src_width; out_bpp = (packed_out) ? 3 : 4; out_alpha = gst_video_format_has_alpha (dest_format); p_out[0] = gst_video_format_get_component_offset (dest_format, 3, dest_width, dest_height); p_out[1] = gst_video_format_get_component_offset (dest_format, 0, dest_width, dest_height); p_out[2] = gst_video_format_get_component_offset (dest_format, 1, dest_width, dest_height); p_out[3] = gst_video_format_get_component_offset (dest_format, 2, dest_width, dest_height); memcpy (matrix, (src_sdtv) ? cog_ycbcr_to_rgb_matrix_8bit_sdtv : cog_ycbcr_to_rgb_matrix_8bit_hdtv, 12 * sizeof (gint)); dest = dest + dest_y * dest_stride + dest_x * out_bpp; src = src + src_y * src_stride + src_x * 4; if (out_alpha) { w *= 4; for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { a = (src[j + 0] * i_alpha) >> 8; y = src[j + 1]; u = src[j + 2]; v = src[j + 3]; r = APPLY_MATRIX (matrix, 0, y, u, v); g = APPLY_MATRIX (matrix, 1, y, u, v); b = APPLY_MATRIX (matrix, 2, y, u, v); dest[j + p_out[0]] = a; dest[j + p_out[1]] = CLAMP (r, 0, 255); dest[j + p_out[2]] = CLAMP (g, 0, 255); dest[j + p_out[3]] = CLAMP (b, 0, 255); } dest += dest_stride; src += src_stride; } } else if (!packed_out) { w *= 4; for (i = 0; i < h; i++) { for (j = 0; j < w; j += 4) { y = src[j + 1]; u = src[j + 2]; v = src[j + 3]; r = APPLY_MATRIX (matrix, 0, y, u, v); g = APPLY_MATRIX (matrix, 1, y, u, v); b = APPLY_MATRIX (matrix, 2, y, u, v); dest[j + p_out[1]] = CLAMP (r, 0, 255); dest[j + p_out[2]] = CLAMP (g, 0, 255); dest[j + p_out[3]] = CLAMP (b, 0, 255); } dest += dest_stride; src += src_stride; } } else { for (i = 0; i < h; i++) { for (j = 0; j < w; j++) { y = src[4 * j + 1]; u = src[4 * j + 2]; v = src[4 * j + 3]; r = APPLY_MATRIX (matrix, 0, y, u, v); g = APPLY_MATRIX (matrix, 1, y, u, v); b = APPLY_MATRIX (matrix, 2, y, u, v); dest[out_bpp * j + p_out[1]] = CLAMP (r, 0, 255); dest[out_bpp * j + p_out[2]] = CLAMP (g, 0, 255); dest[out_bpp * j + p_out[3]] = CLAMP (b, 0, 255); } dest += dest_stride; src += src_stride; } } } static void fill_gray (GstVideoBoxFill fill_type, guint b_alpha, GstVideoFormat format, guint8 * dest, gboolean sdtv, gint width, gint height) { gint i, j; gint dest_stride; if (format == GST_VIDEO_FORMAT_GRAY8) { guint8 val = yuv_sdtv_colors_Y[fill_type]; dest_stride = GST_ROUND_UP_4 (width); for (i = 0; i < height; i++) { memset (dest, val, width); dest += dest_stride; } } else { guint16 val = yuv_sdtv_colors_Y[fill_type] << 8; dest_stride = GST_ROUND_UP_4 (width * 2); if (format == GST_VIDEO_FORMAT_GRAY16_BE) { for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { GST_WRITE_UINT16_BE (dest + 2 * j, val); } dest += dest_stride; } } else { for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { GST_WRITE_UINT16_LE (dest + 2 * j, val); } dest += dest_stride; } } } } static void copy_packed_simple (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i; gint src_stride, dest_stride; gint pixel_stride, row_size; src_stride = gst_video_format_get_row_stride (src_format, 0, src_width); dest_stride = gst_video_format_get_row_stride (dest_format, 0, dest_width); pixel_stride = gst_video_format_get_pixel_stride (dest_format, 0); row_size = w * pixel_stride; dest = dest + dest_y * dest_stride + dest_x * pixel_stride; src = src + src_y * src_stride + src_x * pixel_stride; for (i = 0; i < h; i++) { memcpy (dest, src, row_size); dest += dest_stride; src += src_stride; } } static void fill_yuy2 (GstVideoBoxFill fill_type, guint b_alpha, GstVideoFormat format, guint8 * dest, gboolean sdtv, gint width, gint height) { guint8 y, u, v; gint i, j; gint stride = gst_video_format_get_row_stride (format, 0, width); y = (sdtv) ? yuv_sdtv_colors_Y[fill_type] : yuv_hdtv_colors_Y[fill_type]; u = (sdtv) ? yuv_sdtv_colors_U[fill_type] : yuv_hdtv_colors_U[fill_type]; v = (sdtv) ? yuv_sdtv_colors_V[fill_type] : yuv_hdtv_colors_V[fill_type]; width = width + (width % 2); if (format == GST_VIDEO_FORMAT_YUY2) { for (i = 0; i < height; i++) { for (j = 0; j < width; j += 2) { dest[j * 2 + 0] = y; dest[j * 2 + 1] = u; dest[j * 2 + 2] = y; dest[j * 2 + 3] = v; } dest += stride; } } else if (format == GST_VIDEO_FORMAT_YVYU) { for (i = 0; i < height; i++) { for (j = 0; j < width; j += 2) { dest[j * 2 + 0] = y; dest[j * 2 + 1] = v; dest[j * 2 + 2] = y; dest[j * 2 + 3] = u; } dest += stride; } } else { for (i = 0; i < height; i++) { for (j = 0; j < width; j += 2) { dest[j * 2 + 0] = u; dest[j * 2 + 1] = y; dest[j * 2 + 2] = v; dest[j * 2 + 3] = y; } dest += stride; } } } static void copy_yuy2_yuy2 (guint i_alpha, GstVideoFormat dest_format, guint8 * dest, gboolean dest_sdtv, gint dest_width, gint dest_height, gint dest_x, gint dest_y, GstVideoFormat src_format, const guint8 * src, gboolean src_sdtv, gint src_width, gint src_height, gint src_x, gint src_y, gint w, gint h) { gint i, j; gint src_stride, dest_stride; src_stride = gst_video_format_get_row_stride (src_format, 0, src_width); dest_stride = gst_video_format_get_row_stride (dest_format, 0, dest_width); dest_x = (dest_x & ~1); src_x = (src_x & ~1); w = w + (w % 2); dest = dest + dest_y * dest_stride + dest_x * 2; src = src + src_y * src_stride + src_x * 2; if (src_sdtv != dest_sdtv) { gint y1, u1, v1; gint y2, u2, v2; gint matrix[12]; memcpy (matrix, dest_sdtv ? cog_ycbcr_hdtv_to_ycbcr_sdtv_matrix_8bit : cog_ycbcr_sdtv_to_ycbcr_hdtv_matrix_8bit, 12 * sizeof (gint)); if (src_format == GST_VIDEO_FORMAT_YUY2) { for (i = 0; i < h; i++) { for (j = 0; j < w; j += 2) { y1 = src[j * 2 + 0]; y2 = src[j * 2 + 2]; u1 = u2 = src[j * 2 + 1]; v1 = v2 = src[j * 2 + 3]; dest[j * 2 + 0] = APPLY_MATRIX (matrix, 0, y1, u1, v1); dest[j * 2 + 1] = APPLY_MATRIX (matrix, 1, y1, u1, v1); dest[j * 2 + 2] = APPLY_MATRIX (matrix, 0, y1, u2, v2); dest[j * 2 + 3] = APPLY_MATRIX (matrix, 2, y2, u2, v2); } dest += dest_stride; src += src_stride; } } else if (src_format == GST_VIDEO_FORMAT_YVYU) { for (i = 0; i < h; i++) { for (j = 0; j < w; j += 2) { y1 = src[j * 2 + 0]; y2 = src[j * 2 + 2]; v1 = v2 = src[j * 2 + 1]; u1 = u2 = src[j * 2 + 3]; dest[j * 2 + 0] = APPLY_MATRIX (matrix, 0, y1, u1, v1); dest[j * 2 + 1] = APPLY_MATRIX (matrix, 2, y1, u1, v1); dest[j * 2 + 2] = APPLY_MATRIX (matrix, 0, y1, u2, v2); dest[j * 2 + 3] = APPLY_MATRIX (matrix, 1, y2, u2, v2); } dest += dest_stride; src += src_stride; } } else { for (i = 0; i < h; i++) { for (j = 0; j < w; j += 2) { u1 = u2 = src[j * 2 + 0]; v1 = v2 = src[j * 2 + 2]; y1 = src[j * 2 + 1]; y2 = src[j * 2 + 3]; dest[j * 2 + 1] = APPLY_MATRIX (matrix, 0, y1, u1, v1); dest[j * 2 + 0] = APPLY_MATRIX (matrix, 1, y1, u1, v1); dest[j * 2 + 3] = APPLY_MATRIX (matrix, 0, y1, u2, v2); dest[j * 2 + 2] = APPLY_MATRIX (matrix, 2, y2, u2, v2); } dest += dest_stride; src += src_stride; } } } else { for (i = 0; i < h; i++) { memcpy (dest, src, w * 2); dest += dest_stride; src += src_stride; } } } #define DEFAULT_LEFT 0 #define DEFAULT_RIGHT 0 #define DEFAULT_TOP 0 #define DEFAULT_BOTTOM 0 #define DEFAULT_FILL_TYPE VIDEO_BOX_FILL_BLACK #define DEFAULT_ALPHA 1.0 #define DEFAULT_BORDER_ALPHA 1.0 enum { PROP_0, PROP_LEFT, PROP_RIGHT, PROP_TOP, PROP_BOTTOM, PROP_FILL_TYPE, PROP_ALPHA, PROP_BORDER_ALPHA, PROP_AUTOCROP /* FILL ME */ }; static GstStaticPadTemplate gst_video_box_src_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("AYUV") ";" GST_VIDEO_CAPS_ARGB ";" GST_VIDEO_CAPS_BGRA ";" GST_VIDEO_CAPS_ABGR ";" GST_VIDEO_CAPS_RGBA ";" GST_VIDEO_CAPS_xRGB ";" GST_VIDEO_CAPS_BGRx ";" GST_VIDEO_CAPS_xBGR ";" GST_VIDEO_CAPS_RGBx ";" GST_VIDEO_CAPS_RGB ";" GST_VIDEO_CAPS_BGR ";" GST_VIDEO_CAPS_YUV ("Y444") ";" GST_VIDEO_CAPS_YUV ("Y42B") ";" GST_VIDEO_CAPS_YUV ("YUY2") ";" GST_VIDEO_CAPS_YUV ("YVYU") ";" GST_VIDEO_CAPS_YUV ("UYVY") ";" GST_VIDEO_CAPS_YUV ("I420") ";" GST_VIDEO_CAPS_YUV ("YV12") ";" GST_VIDEO_CAPS_YUV ("Y41B") ";" GST_VIDEO_CAPS_GRAY8 ";" GST_VIDEO_CAPS_GRAY16 ("BIG_ENDIAN") ";" GST_VIDEO_CAPS_GRAY16 ("LITTLE_ENDIAN")) ); static GstStaticPadTemplate gst_video_box_sink_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("AYUV") ";" GST_VIDEO_CAPS_ARGB ";" GST_VIDEO_CAPS_BGRA ";" GST_VIDEO_CAPS_ABGR ";" GST_VIDEO_CAPS_RGBA ";" GST_VIDEO_CAPS_xRGB ";" GST_VIDEO_CAPS_BGRx ";" GST_VIDEO_CAPS_xBGR ";" GST_VIDEO_CAPS_RGBx ";" GST_VIDEO_CAPS_RGB ";" GST_VIDEO_CAPS_BGR ";" GST_VIDEO_CAPS_YUV ("Y444") ";" GST_VIDEO_CAPS_YUV ("Y42B") ";" GST_VIDEO_CAPS_YUV ("YUY2") ";" GST_VIDEO_CAPS_YUV ("YVYU") ";" GST_VIDEO_CAPS_YUV ("UYVY") ";" GST_VIDEO_CAPS_YUV ("I420") ";" GST_VIDEO_CAPS_YUV ("YV12") ";" GST_VIDEO_CAPS_YUV ("Y41B") ";" GST_VIDEO_CAPS_GRAY8 ";" GST_VIDEO_CAPS_GRAY16 ("BIG_ENDIAN") ";" GST_VIDEO_CAPS_GRAY16 ("LITTLE_ENDIAN")) ); GST_BOILERPLATE (GstVideoBox, gst_video_box, GstBaseTransform, GST_TYPE_BASE_TRANSFORM); static void gst_video_box_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_video_box_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static gboolean gst_video_box_recalc_transform (GstVideoBox * video_box); static GstCaps *gst_video_box_transform_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * from); static gboolean gst_video_box_set_caps (GstBaseTransform * trans, GstCaps * in, GstCaps * out); static gboolean gst_video_box_get_unit_size (GstBaseTransform * trans, GstCaps * caps, guint * size); static GstFlowReturn gst_video_box_transform (GstBaseTransform * trans, GstBuffer * in, GstBuffer * out); static void gst_video_box_before_transform (GstBaseTransform * trans, GstBuffer * in); static void gst_video_box_fixate_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * caps, GstCaps * othercaps); static gboolean gst_video_box_src_event (GstBaseTransform * trans, GstEvent * event); #define GST_TYPE_VIDEO_BOX_FILL (gst_video_box_fill_get_type()) static GType gst_video_box_fill_get_type (void) { static GType video_box_fill_type = 0; static const GEnumValue video_box_fill[] = { {VIDEO_BOX_FILL_BLACK, "Black", "black"}, {VIDEO_BOX_FILL_GREEN, "Green", "green"}, {VIDEO_BOX_FILL_BLUE, "Blue", "blue"}, {VIDEO_BOX_FILL_RED, "Red", "red"}, {VIDEO_BOX_FILL_YELLOW, "Yellow", "yellow"}, {VIDEO_BOX_FILL_WHITE, "White", "white"}, {0, NULL, NULL}, }; if (!video_box_fill_type) { video_box_fill_type = g_enum_register_static ("GstVideoBoxFill", video_box_fill); } return video_box_fill_type; } static void gst_video_box_base_init (gpointer g_class) { GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); gst_element_class_set_details_simple (element_class, "Video box filter", "Filter/Effect/Video", "Resizes a video by adding borders or cropping", "Wim Taymans "); gst_element_class_add_static_pad_template (element_class, &gst_video_box_sink_template); gst_element_class_add_static_pad_template (element_class, &gst_video_box_src_template); } static void gst_video_box_finalize (GObject * object) { GstVideoBox *video_box = GST_VIDEO_BOX (object); if (video_box->mutex) { g_mutex_free (video_box->mutex); video_box->mutex = NULL; } G_OBJECT_CLASS (parent_class)->finalize (object); } static void gst_video_box_class_init (GstVideoBoxClass * klass) { GObjectClass *gobject_class = (GObjectClass *) klass; GstBaseTransformClass *trans_class = (GstBaseTransformClass *) klass; gobject_class->set_property = gst_video_box_set_property; gobject_class->get_property = gst_video_box_get_property; gobject_class->finalize = gst_video_box_finalize; g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_FILL_TYPE, g_param_spec_enum ("fill", "Fill", "How to fill the borders", GST_TYPE_VIDEO_BOX_FILL, DEFAULT_FILL_TYPE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_LEFT, g_param_spec_int ("left", "Left", "Pixels to box at left (<0 = add a border)", G_MININT, G_MAXINT, DEFAULT_LEFT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_RIGHT, g_param_spec_int ("right", "Right", "Pixels to box at right (<0 = add a border)", G_MININT, G_MAXINT, DEFAULT_RIGHT, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_TOP, g_param_spec_int ("top", "Top", "Pixels to box at top (<0 = add a border)", G_MININT, G_MAXINT, DEFAULT_TOP, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_BOTTOM, g_param_spec_int ("bottom", "Bottom", "Pixels to box at bottom (<0 = add a border)", G_MININT, G_MAXINT, DEFAULT_BOTTOM, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_ALPHA, g_param_spec_double ("alpha", "Alpha", "Alpha value picture", 0.0, 1.0, DEFAULT_ALPHA, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_BORDER_ALPHA, g_param_spec_double ("border-alpha", "Border Alpha", "Alpha value of the border", 0.0, 1.0, DEFAULT_BORDER_ALPHA, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | GST_PARAM_CONTROLLABLE)); /** * GstVideoBox:autocrop * * If set to %TRUE videobox will automatically crop/pad the input * video to be centered in the output. * * Since: 0.10.16 **/ g_object_class_install_property (G_OBJECT_CLASS (klass), PROP_AUTOCROP, g_param_spec_boolean ("autocrop", "Auto crop", "Auto crop", FALSE, G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS)); trans_class->transform = GST_DEBUG_FUNCPTR (gst_video_box_transform); trans_class->before_transform = GST_DEBUG_FUNCPTR (gst_video_box_before_transform); trans_class->transform_caps = GST_DEBUG_FUNCPTR (gst_video_box_transform_caps); trans_class->set_caps = GST_DEBUG_FUNCPTR (gst_video_box_set_caps); trans_class->get_unit_size = GST_DEBUG_FUNCPTR (gst_video_box_get_unit_size); trans_class->fixate_caps = GST_DEBUG_FUNCPTR (gst_video_box_fixate_caps); trans_class->src_event = GST_DEBUG_FUNCPTR (gst_video_box_src_event); } static void gst_video_box_init (GstVideoBox * video_box, GstVideoBoxClass * g_class) { video_box->box_right = DEFAULT_RIGHT; video_box->box_left = DEFAULT_LEFT; video_box->box_top = DEFAULT_TOP; video_box->box_bottom = DEFAULT_BOTTOM; video_box->crop_right = 0; video_box->crop_left = 0; video_box->crop_top = 0; video_box->crop_bottom = 0; video_box->fill_type = DEFAULT_FILL_TYPE; video_box->alpha = DEFAULT_ALPHA; video_box->border_alpha = DEFAULT_BORDER_ALPHA; video_box->autocrop = FALSE; video_box->mutex = g_mutex_new (); } static void gst_video_box_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { GstVideoBox *video_box = GST_VIDEO_BOX (object); g_mutex_lock (video_box->mutex); switch (prop_id) { case PROP_LEFT: video_box->box_left = g_value_get_int (value); if (video_box->box_left < 0) { video_box->border_left = -video_box->box_left; video_box->crop_left = 0; } else { video_box->border_left = 0; video_box->crop_left = video_box->box_left; } break; case PROP_RIGHT: video_box->box_right = g_value_get_int (value); if (video_box->box_right < 0) { video_box->border_right = -video_box->box_right; video_box->crop_right = 0; } else { video_box->border_right = 0; video_box->crop_right = video_box->box_right; } break; case PROP_TOP: video_box->box_top = g_value_get_int (value); if (video_box->box_top < 0) { video_box->border_top = -video_box->box_top; video_box->crop_top = 0; } else { video_box->border_top = 0; video_box->crop_top = video_box->box_top; } break; case PROP_BOTTOM: video_box->box_bottom = g_value_get_int (value); if (video_box->box_bottom < 0) { video_box->border_bottom = -video_box->box_bottom; video_box->crop_bottom = 0; } else { video_box->border_bottom = 0; video_box->crop_bottom = video_box->box_bottom; } break; case PROP_FILL_TYPE: video_box->fill_type = g_value_get_enum (value); break; case PROP_ALPHA: video_box->alpha = g_value_get_double (value); break; case PROP_BORDER_ALPHA: video_box->border_alpha = g_value_get_double (value); break; case PROP_AUTOCROP: video_box->autocrop = g_value_get_boolean (value); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } gst_video_box_recalc_transform (video_box); GST_DEBUG_OBJECT (video_box, "Calling reconfigure"); gst_base_transform_reconfigure (GST_BASE_TRANSFORM_CAST (video_box)); g_mutex_unlock (video_box->mutex); } static void gst_video_box_autocrop (GstVideoBox * video_box) { gint crop_w = video_box->in_width - video_box->out_width; gint crop_h = video_box->in_height - video_box->out_height; video_box->box_left = crop_w / 2; if (video_box->box_left < 0) { video_box->border_left = -video_box->box_left; video_box->crop_left = 0; } else { video_box->border_left = 0; video_box->crop_left = video_box->box_left; } /* Round down/up for odd width differences */ if (crop_w < 0) crop_w -= 1; else crop_w += 1; video_box->box_right = crop_w / 2; if (video_box->box_right < 0) { video_box->border_right = -video_box->box_right; video_box->crop_right = 0; } else { video_box->border_right = 0; video_box->crop_right = video_box->box_right; } video_box->box_top = crop_h / 2; if (video_box->box_top < 0) { video_box->border_top = -video_box->box_top; video_box->crop_top = 0; } else { video_box->border_top = 0; video_box->crop_top = video_box->box_top; } /* Round down/up for odd height differences */ if (crop_h < 0) crop_h -= 1; else crop_h += 1; video_box->box_bottom = crop_h / 2; if (video_box->box_bottom < 0) { video_box->border_bottom = -video_box->box_bottom; video_box->crop_bottom = 0; } else { video_box->border_bottom = 0; video_box->crop_bottom = video_box->box_bottom; } } static void gst_video_box_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { GstVideoBox *video_box = GST_VIDEO_BOX (object); switch (prop_id) { case PROP_LEFT: g_value_set_int (value, video_box->box_left); break; case PROP_RIGHT: g_value_set_int (value, video_box->box_right); break; case PROP_TOP: g_value_set_int (value, video_box->box_top); break; case PROP_BOTTOM: g_value_set_int (value, video_box->box_bottom); break; case PROP_FILL_TYPE: g_value_set_enum (value, video_box->fill_type); break; case PROP_ALPHA: g_value_set_double (value, video_box->alpha); break; case PROP_BORDER_ALPHA: g_value_set_double (value, video_box->border_alpha); break; case PROP_AUTOCROP: g_value_set_boolean (value, video_box->autocrop); break; default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static inline gint gst_video_box_transform_dimension (gint val, gint delta) { gint64 new_val = (gint64) val + (gint64) delta; new_val = CLAMP (new_val, 1, G_MAXINT); return (gint) new_val; } static gboolean gst_video_box_transform_dimension_value (const GValue * src_val, gint delta, GValue * dest_val) { gboolean ret = TRUE; g_value_init (dest_val, G_VALUE_TYPE (src_val)); if (G_VALUE_HOLDS_INT (src_val)) { gint ival = g_value_get_int (src_val); ival = gst_video_box_transform_dimension (ival, delta); g_value_set_int (dest_val, ival); } else if (GST_VALUE_HOLDS_INT_RANGE (src_val)) { gint min = gst_value_get_int_range_min (src_val); gint max = gst_value_get_int_range_max (src_val); min = gst_video_box_transform_dimension (min, delta); max = gst_video_box_transform_dimension (max, delta); if (min > max) { ret = FALSE; g_value_unset (dest_val); } else { gst_value_set_int_range (dest_val, min, max); } } else if (GST_VALUE_HOLDS_LIST (src_val)) { gint i; for (i = 0; i < gst_value_list_get_size (src_val); ++i) { const GValue *list_val; GValue newval = { 0, }; list_val = gst_value_list_get_value (src_val, i); if (gst_video_box_transform_dimension_value (list_val, delta, &newval)) gst_value_list_append_value (dest_val, &newval); g_value_unset (&newval); } if (gst_value_list_get_size (dest_val) == 0) { g_value_unset (dest_val); ret = FALSE; } } else { g_value_unset (dest_val); ret = FALSE; } return ret; } static GstCaps * gst_video_box_transform_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * from) { GstVideoBox *video_box = GST_VIDEO_BOX (trans); GstCaps *to, *ret; const GstCaps *templ; const gchar *name; GstStructure *structure; GstPad *other; to = gst_caps_copy (from); /* Just to be sure... */ gst_caps_truncate (to); structure = gst_caps_get_structure (to, 0); /* Transform width/height */ if (video_box->autocrop) { gst_structure_remove_field (structure, "width"); gst_structure_remove_field (structure, "height"); } else { gint dw = 0, dh = 0; const GValue *v; GValue w_val = { 0, }; GValue h_val = { 0, }; /* calculate width and height */ if (direction == GST_PAD_SINK) { dw -= video_box->box_left; dw -= video_box->box_right; } else { dw += video_box->box_left; dw += video_box->box_right; } if (direction == GST_PAD_SINK) { dh -= video_box->box_top; dh -= video_box->box_bottom; } else { dh += video_box->box_top; dh += video_box->box_bottom; } v = gst_structure_get_value (structure, "width"); if (!gst_video_box_transform_dimension_value (v, dw, &w_val)) { GST_WARNING_OBJECT (video_box, "could not tranform width value with dw=%d" ", caps structure=%" GST_PTR_FORMAT, dw, structure); gst_caps_unref (to); to = gst_caps_new_empty (); return to; } gst_structure_set_value (structure, "width", &w_val); v = gst_structure_get_value (structure, "height"); if (!gst_video_box_transform_dimension_value (v, dh, &h_val)) { g_value_unset (&w_val); GST_WARNING_OBJECT (video_box, "could not tranform height value with dh=%d" ", caps structure=%" GST_PTR_FORMAT, dh, structure); gst_caps_unref (to); to = gst_caps_new_empty (); return to; } gst_structure_set_value (structure, "height", &h_val); g_value_unset (&w_val); g_value_unset (&h_val); } /* Supported conversions: * I420->AYUV * I420->YV12 * YV12->AYUV * YV12->I420 * AYUV->I420 * AYUV->YV12 * AYUV->xRGB (24bpp, 32bpp, incl. alpha) * xRGB->xRGB (24bpp, 32bpp, from/to all variants, incl. alpha) * xRGB->AYUV (24bpp, 32bpp, incl. alpha) * * Passthrough only for everything else. */ name = gst_structure_get_name (structure); if (g_str_equal (name, "video/x-raw-yuv")) { guint32 fourcc; if (gst_structure_get_fourcc (structure, "format", &fourcc) && (fourcc == GST_STR_FOURCC ("AYUV") || fourcc == GST_STR_FOURCC ("I420") || fourcc == GST_STR_FOURCC ("YV12"))) { GValue list = { 0, }; GValue val = { 0, }; GstStructure *s2; /* get rid of format */ gst_structure_remove_field (structure, "format"); gst_structure_remove_field (structure, "color-matrix"); gst_structure_remove_field (structure, "chroma-site"); s2 = gst_structure_copy (structure); g_value_init (&list, GST_TYPE_LIST); g_value_init (&val, GST_TYPE_FOURCC); gst_value_set_fourcc (&val, GST_STR_FOURCC ("AYUV")); gst_value_list_append_value (&list, &val); g_value_reset (&val); gst_value_set_fourcc (&val, GST_STR_FOURCC ("I420")); gst_value_list_append_value (&list, &val); g_value_reset (&val); gst_value_set_fourcc (&val, GST_STR_FOURCC ("YV12")); gst_value_list_append_value (&list, &val); g_value_unset (&val); gst_structure_set_value (structure, "format", &list); g_value_unset (&list); /* We can only convert to RGB if input is AYUV */ if (fourcc == GST_STR_FOURCC ("AYUV")) { gst_structure_set_name (s2, "video/x-raw-rgb"); g_value_init (&list, GST_TYPE_LIST); g_value_init (&val, G_TYPE_INT); g_value_set_int (&val, 32); gst_value_list_append_value (&list, &val); g_value_reset (&val); g_value_set_int (&val, 24); gst_value_list_append_value (&list, &val); g_value_unset (&val); gst_structure_set_value (s2, "depth", &list); gst_structure_set_value (s2, "bpp", &list); g_value_unset (&list); gst_caps_append_structure (to, s2); } else { gst_structure_free (s2); } } } else if (g_str_equal (name, "video/x-raw-rgb")) { gint bpp; if (gst_structure_get_int (structure, "bpp", &bpp) && (bpp == 32 || bpp == 24)) { GValue list = { 0, }; GValue val = { 0, }; GstStructure *s2; /* get rid of format */ gst_structure_remove_field (structure, "depth"); gst_structure_remove_field (structure, "bpp"); gst_structure_remove_field (structure, "red_mask"); gst_structure_remove_field (structure, "green_mask"); gst_structure_remove_field (structure, "blue_mask"); gst_structure_remove_field (structure, "alpha_mask"); s2 = gst_structure_copy (structure); g_value_init (&list, GST_TYPE_LIST); g_value_init (&val, G_TYPE_INT); g_value_set_int (&val, 32); gst_value_list_append_value (&list, &val); g_value_reset (&val); g_value_set_int (&val, 24); gst_value_list_append_value (&list, &val); g_value_unset (&val); gst_structure_set_value (structure, "depth", &list); gst_structure_set_value (structure, "bpp", &list); g_value_unset (&list); gst_structure_set_name (s2, "video/x-raw-yuv"); gst_structure_set (s2, "format", GST_TYPE_FOURCC, GST_STR_FOURCC ("AYUV"), NULL); gst_caps_append_structure (to, s2); } } /* filter against set allowed caps on the pad */ other = (direction == GST_PAD_SINK) ? trans->srcpad : trans->sinkpad; templ = gst_pad_get_pad_template_caps (other); ret = gst_caps_intersect (to, templ); gst_caps_unref (to); GST_DEBUG_OBJECT (video_box, "direction %d, transformed %" GST_PTR_FORMAT " to %" GST_PTR_FORMAT, direction, from, ret); return ret; } static gboolean gst_video_box_recalc_transform (GstVideoBox * video_box) { gboolean res = TRUE; /* if we have the same format in and out and we don't need to perform any * cropping at all, we can just operate in passthrough mode */ if (video_box->in_format == video_box->out_format && video_box->box_left == 0 && video_box->box_right == 0 && video_box->box_top == 0 && video_box->box_bottom == 0 && video_box->in_sdtv == video_box->out_sdtv) { GST_LOG_OBJECT (video_box, "we are using passthrough"); gst_base_transform_set_passthrough (GST_BASE_TRANSFORM_CAST (video_box), TRUE); } else { GST_LOG_OBJECT (video_box, "we are not using passthrough"); gst_base_transform_set_passthrough (GST_BASE_TRANSFORM_CAST (video_box), FALSE); } return res; } static gboolean gst_video_box_select_processing_functions (GstVideoBox * video_box) { switch (video_box->out_format) { case GST_VIDEO_FORMAT_AYUV: video_box->fill = fill_ayuv; switch (video_box->in_format) { case GST_VIDEO_FORMAT_AYUV: video_box->copy = copy_ayuv_ayuv; break; case GST_VIDEO_FORMAT_I420: case GST_VIDEO_FORMAT_YV12: video_box->copy = copy_i420_ayuv; break; case GST_VIDEO_FORMAT_ARGB: case GST_VIDEO_FORMAT_ABGR: case GST_VIDEO_FORMAT_RGBA: case GST_VIDEO_FORMAT_BGRA: case GST_VIDEO_FORMAT_xRGB: case GST_VIDEO_FORMAT_xBGR: case GST_VIDEO_FORMAT_RGBx: case GST_VIDEO_FORMAT_BGRx: case GST_VIDEO_FORMAT_RGB: case GST_VIDEO_FORMAT_BGR: video_box->copy = copy_rgb32_ayuv; break; default: break; } break; case GST_VIDEO_FORMAT_I420: case GST_VIDEO_FORMAT_YV12: video_box->fill = fill_planar_yuv; switch (video_box->in_format) { case GST_VIDEO_FORMAT_AYUV: video_box->copy = copy_ayuv_i420; break; case GST_VIDEO_FORMAT_I420: case GST_VIDEO_FORMAT_YV12: video_box->copy = copy_i420_i420; break; default: break; } break; case GST_VIDEO_FORMAT_ARGB: case GST_VIDEO_FORMAT_ABGR: case GST_VIDEO_FORMAT_RGBA: case GST_VIDEO_FORMAT_BGRA: case GST_VIDEO_FORMAT_xRGB: case GST_VIDEO_FORMAT_xBGR: case GST_VIDEO_FORMAT_RGBx: case GST_VIDEO_FORMAT_BGRx: case GST_VIDEO_FORMAT_RGB: case GST_VIDEO_FORMAT_BGR: video_box->fill = (video_box->out_format == GST_VIDEO_FORMAT_BGR || video_box->out_format == GST_VIDEO_FORMAT_RGB) ? fill_rgb24 : fill_rgb32; switch (video_box->in_format) { case GST_VIDEO_FORMAT_ARGB: case GST_VIDEO_FORMAT_ABGR: case GST_VIDEO_FORMAT_RGBA: case GST_VIDEO_FORMAT_BGRA: case GST_VIDEO_FORMAT_xRGB: case GST_VIDEO_FORMAT_xBGR: case GST_VIDEO_FORMAT_RGBx: case GST_VIDEO_FORMAT_BGRx: case GST_VIDEO_FORMAT_RGB: case GST_VIDEO_FORMAT_BGR: video_box->copy = copy_rgb32; break; case GST_VIDEO_FORMAT_AYUV: video_box->copy = copy_ayuv_rgb32; default: break; } break; case GST_VIDEO_FORMAT_GRAY8: case GST_VIDEO_FORMAT_GRAY16_BE: case GST_VIDEO_FORMAT_GRAY16_LE: video_box->fill = fill_gray; switch (video_box->in_format) { case GST_VIDEO_FORMAT_GRAY8: case GST_VIDEO_FORMAT_GRAY16_BE: case GST_VIDEO_FORMAT_GRAY16_LE: video_box->copy = copy_packed_simple; break; default: break; } break; case GST_VIDEO_FORMAT_YUY2: case GST_VIDEO_FORMAT_YVYU: case GST_VIDEO_FORMAT_UYVY: video_box->fill = fill_yuy2; switch (video_box->in_format) { case GST_VIDEO_FORMAT_YUY2: case GST_VIDEO_FORMAT_YVYU: case GST_VIDEO_FORMAT_UYVY: video_box->copy = copy_yuy2_yuy2; break; default: break; } break; case GST_VIDEO_FORMAT_Y444: case GST_VIDEO_FORMAT_Y42B: case GST_VIDEO_FORMAT_Y41B: video_box->fill = fill_planar_yuv; switch (video_box->in_format) { case GST_VIDEO_FORMAT_Y444: video_box->copy = copy_y444_y444; break; case GST_VIDEO_FORMAT_Y42B: video_box->copy = copy_y42b_y42b; break; case GST_VIDEO_FORMAT_Y41B: video_box->copy = copy_y41b_y41b; break; default: break; } break; default: break; } return video_box->fill != NULL && video_box->copy != NULL; } static gboolean gst_video_box_set_caps (GstBaseTransform * trans, GstCaps * in, GstCaps * out) { GstVideoBox *video_box = GST_VIDEO_BOX (trans); gboolean ret; const gchar *matrix; g_mutex_lock (video_box->mutex); ret = gst_video_format_parse_caps (in, &video_box->in_format, &video_box->in_width, &video_box->in_height); ret &= gst_video_format_parse_caps (out, &video_box->out_format, &video_box->out_width, &video_box->out_height); matrix = gst_video_parse_caps_color_matrix (in); video_box->in_sdtv = matrix ? g_str_equal (matrix, "sdtv") : TRUE; matrix = gst_video_parse_caps_color_matrix (out); video_box->out_sdtv = matrix ? g_str_equal (matrix, "sdtv") : TRUE; /* something wrong getting the caps */ if (!ret) goto no_caps; GST_DEBUG_OBJECT (trans, "Input w: %d h: %d", video_box->in_width, video_box->in_height); GST_DEBUG_OBJECT (trans, "Output w: %d h: %d", video_box->out_width, video_box->out_height); if (video_box->autocrop) gst_video_box_autocrop (video_box); /* recalc the transformation strategy */ ret = gst_video_box_recalc_transform (video_box); if (ret) ret = gst_video_box_select_processing_functions (video_box); g_mutex_unlock (video_box->mutex); return ret; /* ERRORS */ no_caps: { GST_DEBUG_OBJECT (video_box, "Invalid caps: %" GST_PTR_FORMAT " -> %" GST_PTR_FORMAT, in, out); g_mutex_unlock (video_box->mutex); return FALSE; } } static gboolean gst_video_box_get_unit_size (GstBaseTransform * trans, GstCaps * caps, guint * size) { GstVideoFormat format; gint width, height; gboolean ret; g_assert (size); ret = gst_video_format_parse_caps (caps, &format, &width, &height); if (!ret) { GST_ERROR_OBJECT (trans, "Invalid caps: %" GST_PTR_FORMAT, caps); return FALSE; } *size = gst_video_format_get_size (format, width, height); GST_LOG_OBJECT (trans, "Returning from _unit_size %d", *size); return TRUE; } static void gst_video_box_fixate_caps (GstBaseTransform * trans, GstPadDirection direction, GstCaps * caps, GstCaps * othercaps) { gint width, height; GstStructure *s; gboolean ret; ret = gst_video_format_parse_caps (caps, NULL, &width, &height); if (!ret) return; s = gst_caps_get_structure (othercaps, 0); gst_structure_fixate_field_nearest_int (s, "width", width); gst_structure_fixate_field_nearest_int (s, "height", height); } static gboolean gst_video_box_src_event (GstBaseTransform * trans, GstEvent * event) { GstVideoBox *video_box = GST_VIDEO_BOX (trans); GstStructure *new_structure; const GstStructure *structure; const gchar *event_name; gdouble pointer_x; gdouble pointer_y; GST_OBJECT_LOCK (video_box); if (GST_EVENT_TYPE (event) == GST_EVENT_NAVIGATION && (video_box->box_left != 0 || video_box->box_top != 0)) { structure = gst_event_get_structure (event); event_name = gst_structure_get_string (structure, "event"); if (event_name && (strcmp (event_name, "mouse-move") == 0 || strcmp (event_name, "mouse-button-press") == 0 || strcmp (event_name, "mouse-button-release") == 0)) { if (gst_structure_get_double (structure, "pointer_x", &pointer_x) && gst_structure_get_double (structure, "pointer_y", &pointer_y)) { gdouble new_pointer_x, new_pointer_y; GstEvent *new_event; new_pointer_x = pointer_x + video_box->box_left; new_pointer_y = pointer_y + video_box->box_top; new_structure = gst_structure_copy (structure); gst_structure_set (new_structure, "pointer_x", G_TYPE_DOUBLE, (gdouble) (new_pointer_x), "pointer_y", G_TYPE_DOUBLE, (gdouble) (new_pointer_y), NULL); new_event = gst_event_new_navigation (new_structure); gst_event_unref (event); event = new_event; } else { GST_WARNING_OBJECT (video_box, "Failed to read navigation event"); } } } GST_OBJECT_UNLOCK (video_box); return GST_BASE_TRANSFORM_CLASS (parent_class)->src_event (trans, event); } static void gst_video_box_process (GstVideoBox * video_box, const guint8 * src, guint8 * dest) { guint b_alpha = CLAMP (video_box->border_alpha * 256, 0, 256); guint i_alpha = CLAMP (video_box->alpha * 256, 0, 256); GstVideoBoxFill fill_type = video_box->fill_type; gint br, bl, bt, bb, crop_w, crop_h; crop_h = 0; crop_w = 0; br = video_box->box_right; bl = video_box->box_left; bt = video_box->box_top; bb = video_box->box_bottom; if (br >= 0 && bl >= 0) { crop_w = video_box->in_width - (br + bl); } else if (br >= 0 && bl < 0) { crop_w = video_box->in_width - (br); } else if (br < 0 && bl >= 0) { crop_w = video_box->in_width - (bl); } else if (br < 0 && bl < 0) { crop_w = video_box->in_width; } if (bb >= 0 && bt >= 0) { crop_h = video_box->in_height - (bb + bt); } else if (bb >= 0 && bt < 0) { crop_h = video_box->in_height - (bb); } else if (bb < 0 && bt >= 0) { crop_h = video_box->in_height - (bt); } else if (bb < 0 && bt < 0) { crop_h = video_box->in_height; } GST_DEBUG_OBJECT (video_box, "Borders are: L:%d, R:%d, T:%d, B:%d", bl, br, bt, bb); GST_DEBUG_OBJECT (video_box, "Alpha value is: %u (frame) %u (border)", i_alpha, b_alpha); if (crop_h < 0 || crop_w < 0) { video_box->fill (fill_type, b_alpha, video_box->out_format, dest, video_box->out_sdtv, video_box->out_width, video_box->out_height); } else if (bb == 0 && bt == 0 && br == 0 && bl == 0) { video_box->copy (i_alpha, video_box->out_format, dest, video_box->out_sdtv, video_box->out_width, video_box->out_height, 0, 0, video_box->in_format, src, video_box->in_sdtv, video_box->in_width, video_box->in_height, 0, 0, crop_w, crop_h); } else { gint src_x = 0, src_y = 0; gint dest_x = 0, dest_y = 0; /* Fill everything if a border should be added somewhere */ if (bt < 0 || bb < 0 || br < 0 || bl < 0) video_box->fill (fill_type, b_alpha, video_box->out_format, dest, video_box->out_sdtv, video_box->out_width, video_box->out_height); /* Top border */ if (bt < 0) { dest_y += -bt; } else { src_y += bt; } /* Left border */ if (bl < 0) { dest_x += -bl; } else { src_x += bl; } /* Frame */ video_box->copy (i_alpha, video_box->out_format, dest, video_box->out_sdtv, video_box->out_width, video_box->out_height, dest_x, dest_y, video_box->in_format, src, video_box->in_sdtv, video_box->in_width, video_box->in_height, src_x, src_y, crop_w, crop_h); } GST_LOG_OBJECT (video_box, "image created"); } static void gst_video_box_before_transform (GstBaseTransform * trans, GstBuffer * in) { GstVideoBox *video_box = GST_VIDEO_BOX (trans); GstClockTime timestamp, stream_time; timestamp = GST_BUFFER_TIMESTAMP (in); stream_time = gst_segment_to_stream_time (&trans->segment, GST_FORMAT_TIME, timestamp); GST_DEBUG_OBJECT (video_box, "sync to %" GST_TIME_FORMAT, GST_TIME_ARGS (timestamp)); if (GST_CLOCK_TIME_IS_VALID (stream_time)) gst_object_sync_values (G_OBJECT (video_box), stream_time); } static GstFlowReturn gst_video_box_transform (GstBaseTransform * trans, GstBuffer * in, GstBuffer * out) { GstVideoBox *video_box = GST_VIDEO_BOX (trans); const guint8 *indata; guint8 *outdata; indata = GST_BUFFER_DATA (in); outdata = GST_BUFFER_DATA (out); g_mutex_lock (video_box->mutex); gst_video_box_process (video_box, indata, outdata); g_mutex_unlock (video_box->mutex); return GST_FLOW_OK; } /* FIXME: 0.11 merge with videocrop plugin */ static gboolean plugin_init (GstPlugin * plugin) { gst_controller_init (NULL, NULL); GST_DEBUG_CATEGORY_INIT (videobox_debug, "videobox", 0, "Resizes a video by adding borders or cropping"); return gst_element_register (plugin, "videobox", GST_RANK_NONE, GST_TYPE_VIDEO_BOX); } GST_PLUGIN_DEFINE (GST_VERSION_MAJOR, GST_VERSION_MINOR, "videobox", "resizes a video by adding borders or cropping", plugin_init, VERSION, GST_LICENSE, GST_PACKAGE_NAME, GST_PACKAGE_ORIGIN)