videocrop: Add GstVideoCropMeta support

If downstream supports this meta, it will add or update it from
the GstBuffer in-place rather then copying.

https://bugzilla.gnome.org/show_bug.cgi?id=791453
This commit is contained in:
Nicolas Dufresne 2017-12-07 22:08:42 -05:00
parent de8f080a5c
commit 3a17f52ca3

View file

@ -114,6 +114,13 @@ static gboolean gst_video_crop_set_info (GstVideoFilter * vfilter, GstCaps * in,
static GstFlowReturn gst_video_crop_transform_frame (GstVideoFilter * vfilter, static GstFlowReturn gst_video_crop_transform_frame (GstVideoFilter * vfilter,
GstVideoFrame * in_frame, GstVideoFrame * out_frame); GstVideoFrame * in_frame, GstVideoFrame * out_frame);
static gboolean gst_video_crop_decide_allocation (GstBaseTransform * trans,
GstQuery * query);
static gboolean gst_video_crop_propose_allocation (GstBaseTransform * trans,
GstQuery * decide_query, GstQuery * query);
static GstFlowReturn gst_video_crop_transform_ip (GstBaseTransform * trans,
GstBuffer * buf);
static gboolean static gboolean
gst_video_crop_src_event (GstBaseTransform * trans, GstEvent * event) gst_video_crop_src_event (GstBaseTransform * trans, GstEvent * event)
{ {
@ -205,9 +212,16 @@ gst_video_crop_class_init (GstVideoCropClass * klass)
"Crops video into a user-defined region", "Crops video into a user-defined region",
"Tim-Philipp Müller <tim centricular net>"); "Tim-Philipp Müller <tim centricular net>");
basetransform_class->transform_ip_on_passthrough = FALSE;
basetransform_class->transform_caps = basetransform_class->transform_caps =
GST_DEBUG_FUNCPTR (gst_video_crop_transform_caps); GST_DEBUG_FUNCPTR (gst_video_crop_transform_caps);
basetransform_class->src_event = GST_DEBUG_FUNCPTR (gst_video_crop_src_event); basetransform_class->src_event = GST_DEBUG_FUNCPTR (gst_video_crop_src_event);
basetransform_class->decide_allocation =
GST_DEBUG_FUNCPTR (gst_video_crop_decide_allocation);
basetransform_class->propose_allocation =
GST_DEBUG_FUNCPTR (gst_video_crop_propose_allocation);
basetransform_class->transform_ip =
GST_DEBUG_FUNCPTR (gst_video_crop_transform_ip);
vfilter_class->set_info = GST_DEBUG_FUNCPTR (gst_video_crop_set_info); vfilter_class->set_info = GST_DEBUG_FUNCPTR (gst_video_crop_set_info);
vfilter_class->transform_frame = vfilter_class->transform_frame =
@ -227,7 +241,7 @@ gst_video_crop_init (GstVideoCrop * vcrop)
static void static void
gst_video_crop_transform_packed_complex (GstVideoCrop * vcrop, gst_video_crop_transform_packed_complex (GstVideoCrop * vcrop,
GstVideoFrame * in_frame, GstVideoFrame * out_frame) GstVideoFrame * in_frame, GstVideoFrame * out_frame, gint x, gint y)
{ {
guint8 *in_data, *out_data; guint8 *in_data, *out_data;
guint i, dx; guint i, dx;
@ -280,7 +294,7 @@ gst_video_crop_transform_packed_complex (GstVideoCrop * vcrop,
static void static void
gst_video_crop_transform_packed_simple (GstVideoCrop * vcrop, gst_video_crop_transform_packed_simple (GstVideoCrop * vcrop,
GstVideoFrame * in_frame, GstVideoFrame * out_frame) GstVideoFrame * in_frame, GstVideoFrame * out_frame, gint x, gint y)
{ {
guint8 *in_data, *out_data; guint8 *in_data, *out_data;
gint width, height; gint width, height;
@ -296,8 +310,9 @@ gst_video_crop_transform_packed_simple (GstVideoCrop * vcrop,
in_stride = GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 0); in_stride = GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 0);
out_stride = GST_VIDEO_FRAME_PLANE_STRIDE (out_frame, 0); out_stride = GST_VIDEO_FRAME_PLANE_STRIDE (out_frame, 0);
in_data += vcrop->crop_top * in_stride; in_data += (vcrop->crop_top + y) * in_stride;
in_data += vcrop->crop_left * GST_VIDEO_FRAME_COMP_PSTRIDE (in_frame, 0); in_data +=
(vcrop->crop_left + x) * GST_VIDEO_FRAME_COMP_PSTRIDE (in_frame, 0);
dx = width * GST_VIDEO_FRAME_COMP_PSTRIDE (out_frame, 0); dx = width * GST_VIDEO_FRAME_COMP_PSTRIDE (out_frame, 0);
@ -310,23 +325,24 @@ gst_video_crop_transform_packed_simple (GstVideoCrop * vcrop,
static void static void
gst_video_crop_transform_planar (GstVideoCrop * vcrop, gst_video_crop_transform_planar (GstVideoCrop * vcrop,
GstVideoFrame * in_frame, GstVideoFrame * out_frame) GstVideoFrame * in_frame, GstVideoFrame * out_frame, gint x, gint y)
{ {
gint width, height; gint width, height;
gint crop_top, crop_left;
guint8 *y_out, *u_out, *v_out; guint8 *y_out, *u_out, *v_out;
guint8 *y_in, *u_in, *v_in; guint8 *y_in, *u_in, *v_in;
guint i, dx; guint i, dx;
width = GST_VIDEO_FRAME_WIDTH (out_frame); width = GST_VIDEO_FRAME_WIDTH (out_frame);
height = GST_VIDEO_FRAME_HEIGHT (out_frame); height = GST_VIDEO_FRAME_HEIGHT (out_frame);
crop_left = vcrop->crop_left + x;
crop_top = vcrop->crop_top + y;
/* Y plane */ /* Y plane */
y_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 0); y_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 0);
y_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 0); y_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 0);
y_in += y_in += (crop_top * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 0)) + crop_left;
(vcrop->crop_top * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame,
0)) + vcrop->crop_left;
dx = width; dx = width;
for (i = 0; i < height; ++i) { for (i = 0; i < height; ++i) {
@ -339,14 +355,14 @@ gst_video_crop_transform_planar (GstVideoCrop * vcrop,
u_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 1); u_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 1);
u_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 1); u_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 1);
u_in += (vcrop->crop_top / 2) * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 1); u_in += (crop_top / 2) * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 1);
u_in += vcrop->crop_left / 2; u_in += crop_left / 2;
v_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 2); v_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 2);
v_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 2); v_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 2);
v_in += (vcrop->crop_top / 2) * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 2); v_in += (crop_top / 2) * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 2);
v_in += vcrop->crop_left / 2; v_in += crop_left / 2;
dx = GST_ROUND_UP_2 (width) / 2; dx = GST_ROUND_UP_2 (width) / 2;
@ -362,15 +378,18 @@ gst_video_crop_transform_planar (GstVideoCrop * vcrop,
static void static void
gst_video_crop_transform_semi_planar (GstVideoCrop * vcrop, gst_video_crop_transform_semi_planar (GstVideoCrop * vcrop,
GstVideoFrame * in_frame, GstVideoFrame * out_frame) GstVideoFrame * in_frame, GstVideoFrame * out_frame, gint x, gint y)
{ {
gint width, height; gint width, height;
gint crop_top, crop_left;
guint8 *y_out, *uv_out; guint8 *y_out, *uv_out;
guint8 *y_in, *uv_in; guint8 *y_in, *uv_in;
guint i, dx; guint i, dx;
width = GST_VIDEO_FRAME_WIDTH (out_frame); width = GST_VIDEO_FRAME_WIDTH (out_frame);
height = GST_VIDEO_FRAME_HEIGHT (out_frame); height = GST_VIDEO_FRAME_HEIGHT (out_frame);
crop_left = vcrop->crop_left + x;
crop_top = vcrop->crop_top + y;
/* Y plane */ /* Y plane */
y_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 0); y_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 0);
@ -380,8 +399,7 @@ gst_video_crop_transform_semi_planar (GstVideoCrop * vcrop,
uv_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 1); uv_in = GST_VIDEO_FRAME_PLANE_DATA (in_frame, 1);
uv_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 1); uv_out = GST_VIDEO_FRAME_PLANE_DATA (out_frame, 1);
y_in += vcrop->crop_top * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 0) + y_in += crop_top * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 0) + crop_left;
vcrop->crop_left;
dx = width; dx = width;
for (i = 0; i < height; ++i) { for (i = 0; i < height; ++i) {
@ -390,8 +408,8 @@ gst_video_crop_transform_semi_planar (GstVideoCrop * vcrop,
y_out += GST_VIDEO_FRAME_PLANE_STRIDE (out_frame, 0); y_out += GST_VIDEO_FRAME_PLANE_STRIDE (out_frame, 0);
} }
uv_in += (vcrop->crop_top / 2) * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 1); uv_in += (crop_top / 2) * GST_VIDEO_FRAME_PLANE_STRIDE (in_frame, 1);
uv_in += GST_ROUND_DOWN_2 (vcrop->crop_left); uv_in += GST_ROUND_DOWN_2 (crop_left);
dx = GST_ROUND_UP_2 (width); dx = GST_ROUND_UP_2 (width);
for (i = 0; i < GST_ROUND_UP_2 (height) / 2; i++) { for (i = 0; i < GST_ROUND_UP_2 (height) / 2; i++) {
@ -406,6 +424,8 @@ gst_video_crop_transform_frame (GstVideoFilter * vfilter,
GstVideoFrame * in_frame, GstVideoFrame * out_frame) GstVideoFrame * in_frame, GstVideoFrame * out_frame)
{ {
GstVideoCrop *vcrop = GST_VIDEO_CROP (vfilter); GstVideoCrop *vcrop = GST_VIDEO_CROP (vfilter);
GstVideoCropMeta *meta = gst_buffer_get_video_crop_meta (in_frame->buffer);
gint x = 0, y = 0;
if (G_UNLIKELY (vcrop->need_update)) { if (G_UNLIKELY (vcrop->need_update)) {
if (!gst_video_crop_set_info (vfilter, NULL, &vcrop->in_info, NULL, if (!gst_video_crop_set_info (vfilter, NULL, &vcrop->in_info, NULL,
@ -414,18 +434,24 @@ gst_video_crop_transform_frame (GstVideoFilter * vfilter,
} }
} }
if (meta) {
x = meta->x;
y = meta->y;
}
switch (vcrop->packing) { switch (vcrop->packing) {
case VIDEO_CROP_PIXEL_FORMAT_PACKED_SIMPLE: case VIDEO_CROP_PIXEL_FORMAT_PACKED_SIMPLE:
gst_video_crop_transform_packed_simple (vcrop, in_frame, out_frame); gst_video_crop_transform_packed_simple (vcrop, in_frame, out_frame, x, y);
break; break;
case VIDEO_CROP_PIXEL_FORMAT_PACKED_COMPLEX: case VIDEO_CROP_PIXEL_FORMAT_PACKED_COMPLEX:
gst_video_crop_transform_packed_complex (vcrop, in_frame, out_frame); gst_video_crop_transform_packed_complex (vcrop, in_frame, out_frame, x,
y);
break; break;
case VIDEO_CROP_PIXEL_FORMAT_PLANAR: case VIDEO_CROP_PIXEL_FORMAT_PLANAR:
gst_video_crop_transform_planar (vcrop, in_frame, out_frame); gst_video_crop_transform_planar (vcrop, in_frame, out_frame, x, y);
break; break;
case VIDEO_CROP_PIXEL_FORMAT_SEMI_PLANAR: case VIDEO_CROP_PIXEL_FORMAT_SEMI_PLANAR:
gst_video_crop_transform_semi_planar (vcrop, in_frame, out_frame); gst_video_crop_transform_semi_planar (vcrop, in_frame, out_frame, x, y);
break; break;
default: default:
g_assert_not_reached (); g_assert_not_reached ();
@ -434,6 +460,84 @@ gst_video_crop_transform_frame (GstVideoFilter * vfilter,
return GST_FLOW_OK; return GST_FLOW_OK;
} }
static gboolean
gst_video_crop_decide_allocation (GstBaseTransform * trans, GstQuery * query)
{
GstVideoCrop *crop = GST_VIDEO_CROP (trans);
gboolean use_crop_meta;
use_crop_meta = (gst_query_find_allocation_meta (query,
GST_VIDEO_CROP_META_API_TYPE, NULL) &&
gst_query_find_allocation_meta (query, GST_VIDEO_META_API_TYPE, NULL));
if ((crop->crop_left | crop->crop_right | crop->crop_top | crop->
crop_bottom) == 0) {
GST_INFO_OBJECT (crop, "we are using passthrough");
gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (crop), TRUE);
gst_base_transform_set_in_place (GST_BASE_TRANSFORM (crop), FALSE);
} else if (use_crop_meta) {
GST_INFO_OBJECT (crop, "we are doing in-place transform using crop meta");
gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (crop), FALSE);
gst_base_transform_set_in_place (GST_BASE_TRANSFORM (crop), TRUE);
} else {
GST_INFO_OBJECT (crop, "we are not using passthrough");
gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (crop), FALSE);
gst_base_transform_set_in_place (GST_BASE_TRANSFORM (crop), FALSE);
}
return GST_BASE_TRANSFORM_CLASS (parent_class)->decide_allocation (trans,
query);
}
static gboolean
gst_video_crop_propose_allocation (GstBaseTransform * trans,
GstQuery * decide_query, GstQuery * query)
{
/* if we are not passthrough, we can handle video meta and crop meta */
if (decide_query) {
GST_DEBUG_OBJECT (trans, "Advertising video meta and crop meta support");
gst_query_add_allocation_meta (query, GST_VIDEO_META_API_TYPE, NULL);
gst_query_add_allocation_meta (query, GST_VIDEO_CROP_META_API_TYPE, NULL);
}
return GST_BASE_TRANSFORM_CLASS (parent_class)->propose_allocation (trans,
decide_query, query);
}
static GstFlowReturn
gst_video_crop_transform_ip (GstBaseTransform * trans, GstBuffer * buf)
{
GstVideoCrop *vcrop = GST_VIDEO_CROP (trans);
GstVideoMeta *video_meta;
GstVideoCropMeta *crop_meta;
GST_LOG_OBJECT (trans, "Transforming in-place");
/* The video meta is required since we are going to make the caps
* width/height smaller, which would not result in a usable GstVideoInfo for
* mapping the buffer. */
video_meta = gst_buffer_get_video_meta (buf);
if (!video_meta) {
video_meta = gst_buffer_add_video_meta (buf, GST_VIDEO_FRAME_FLAG_NONE,
GST_VIDEO_INFO_FORMAT (&vcrop->in_info), vcrop->in_info.width,
vcrop->in_info.height);
}
crop_meta = gst_buffer_get_video_crop_meta (buf);
if (!crop_meta) {
crop_meta = gst_buffer_add_video_crop_meta (buf);
crop_meta->width = vcrop->in_info.width;
crop_meta->height = vcrop->in_info.height;
}
crop_meta->x += vcrop->crop_left;
crop_meta->y += vcrop->crop_top;
crop_meta->width = GST_VIDEO_INFO_WIDTH (&vcrop->out_info);
crop_meta->height = GST_VIDEO_INFO_HEIGHT (&vcrop->out_info);
return GST_FLOW_OK;
}
static gint static gint
gst_video_crop_transform_dimension (gint val, gint delta) gst_video_crop_transform_dimension (gint val, gint delta)
{ {
@ -661,15 +765,6 @@ gst_video_crop_set_info (GstVideoFilter * vfilter, GstCaps * in,
GST_LOG_OBJECT (crop, "incaps = %" GST_PTR_FORMAT ", outcaps = %" GST_LOG_OBJECT (crop, "incaps = %" GST_PTR_FORMAT ", outcaps = %"
GST_PTR_FORMAT, in, out); GST_PTR_FORMAT, in, out);
if ((crop->crop_left | crop->crop_right | crop->crop_top | crop->
crop_bottom) == 0) {
GST_LOG_OBJECT (crop, "we are using passthrough");
gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (crop), TRUE);
} else {
GST_LOG_OBJECT (crop, "we are not using passthrough");
gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (crop), FALSE);
}
if (GST_VIDEO_INFO_IS_RGB (in_info) if (GST_VIDEO_INFO_IS_RGB (in_info)
|| GST_VIDEO_INFO_IS_GRAY (in_info)) { || GST_VIDEO_INFO_IS_GRAY (in_info)) {
crop->packing = VIDEO_CROP_PIXEL_FORMAT_PACKED_SIMPLE; crop->packing = VIDEO_CROP_PIXEL_FORMAT_PACKED_SIMPLE;
@ -706,6 +801,10 @@ gst_video_crop_set_info (GstVideoFilter * vfilter, GstCaps * in,
crop->in_info = *in_info; crop->in_info = *in_info;
crop->out_info = *out_info; crop->out_info = *out_info;
/* Ensure our decide_allocation will be called again */
gst_base_transform_set_passthrough (GST_BASE_TRANSFORM (crop), FALSE);
gst_base_transform_set_in_place (GST_BASE_TRANSFORM (crop), FALSE);
return TRUE; return TRUE;
/* ERROR */ /* ERROR */