/* GStreamer * Copyright (C) <2017> Sean DuBois * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:element-av1dec * * AV1 Decoder. * * ## Example launch line * * |[ * gst-launch-1.0 -v filesrc location=videotestsrc.webm ! matroskademux ! av1dec ! videoconvert ! videoscale ! autovideosink * ]| */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include "gstav1dec.h" enum { PROP_0, }; static GstStaticPadTemplate gst_av1_dec_sink_pad_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS ("video/x-av1") ); static GstStaticPadTemplate gst_av1_dec_src_pad_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE ("{ I420, YV12, Y42B, Y444" #if G_BYTE_ORDER == G_LITTLE_ENDIAN ", I420_10LE, I420_12LE, I422_10LE, I422_12LE, Y444_10LE, Y444_12LE" #else ", I420_10BE, I420_12BE, I422_10BE, I422_12BE, Y444_10BE, Y444_12BE" #endif " }")) ); GST_DEBUG_CATEGORY_STATIC (av1_dec_debug); #define GST_CAT_DEFAULT av1_dec_debug #define GST_VIDEO_FORMAT_WITH_ENDIAN(fmt,endian) GST_VIDEO_FORMAT_##fmt##endian #if G_BYTE_ORDER == G_LITTLE_ENDIAN #define AOM_FMT_TO_GST(fmt) GST_VIDEO_FORMAT_WITH_ENDIAN(fmt,LE) #else #define AOM_FMT_TO_GST(fmt) GST_VIDEO_FORMAT_WITH_ENDIAN(fmt,BE) #endif static void gst_av1_dec_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec); static void gst_av1_dec_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec); static gboolean gst_av1_dec_start (GstVideoDecoder * dec); static gboolean gst_av1_dec_stop (GstVideoDecoder * dec); static gboolean gst_av1_dec_set_format (GstVideoDecoder * dec, GstVideoCodecState * state); static gboolean gst_av1_dec_flush (GstVideoDecoder * dec); static GstFlowReturn gst_av1_dec_handle_frame (GstVideoDecoder * decoder, GstVideoCodecFrame * frame); static void gst_av1_dec_image_to_buffer (GstAV1Dec * dec, const aom_image_t * img, GstBuffer * buffer); static GstFlowReturn gst_av1_dec_open_codec (GstAV1Dec * av1dec, GstVideoCodecFrame * frame); static gboolean gst_av1_dec_get_valid_format (GstAV1Dec * dec, const aom_image_t * img, GstVideoFormat * fmt); #define gst_av1_dec_parent_class parent_class G_DEFINE_TYPE (GstAV1Dec, gst_av1_dec, GST_TYPE_VIDEO_DECODER); GST_ELEMENT_REGISTER_DEFINE (av1dec, "av1dec", GST_RANK_SECONDARY, GST_TYPE_AV1_DEC); static void gst_av1_dec_class_init (GstAV1DecClass * klass) { GObjectClass *gobject_class; GstElementClass *element_class; GstVideoDecoderClass *vdec_class; gobject_class = (GObjectClass *) klass; element_class = (GstElementClass *) klass; vdec_class = (GstVideoDecoderClass *) klass; parent_class = g_type_class_peek_parent (klass); gobject_class->set_property = gst_av1_dec_set_property; gobject_class->get_property = gst_av1_dec_get_property; gst_element_class_add_static_pad_template (element_class, &gst_av1_dec_src_pad_template); gst_element_class_add_static_pad_template (element_class, &gst_av1_dec_sink_pad_template); gst_element_class_set_static_metadata (element_class, "AV1 Decoder", "Codec/Decoder/Video", "Decode AV1 video streams", "Sean DuBois "); vdec_class->start = GST_DEBUG_FUNCPTR (gst_av1_dec_start); vdec_class->stop = GST_DEBUG_FUNCPTR (gst_av1_dec_stop); vdec_class->flush = GST_DEBUG_FUNCPTR (gst_av1_dec_flush); vdec_class->set_format = GST_DEBUG_FUNCPTR (gst_av1_dec_set_format); vdec_class->handle_frame = GST_DEBUG_FUNCPTR (gst_av1_dec_handle_frame); klass->codec_algo = &aom_codec_av1_dx_algo; GST_DEBUG_CATEGORY_INIT (av1_dec_debug, "av1dec", 0, "AV1 decoding element"); } static void gst_av1_dec_init (GstAV1Dec * av1dec) { GstVideoDecoder *dec = (GstVideoDecoder *) av1dec; GST_DEBUG_OBJECT (dec, "gst_av1_dec_init"); gst_video_decoder_set_packetized (dec, TRUE); gst_video_decoder_set_needs_format (dec, TRUE); gst_video_decoder_set_use_default_pad_acceptcaps (dec, TRUE); GST_PAD_SET_ACCEPT_TEMPLATE (GST_VIDEO_DECODER_SINK_PAD (dec)); } static void gst_av1_dec_set_property (GObject * object, guint prop_id, const GValue * value, GParamSpec * pspec) { switch (prop_id) { default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static void gst_av1_dec_get_property (GObject * object, guint prop_id, GValue * value, GParamSpec * pspec) { switch (prop_id) { default: G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec); break; } } static gboolean gst_av1_dec_start (GstVideoDecoder * dec) { GstAV1Dec *av1dec = GST_AV1_DEC_CAST (dec); av1dec->decoder_inited = FALSE; av1dec->output_state = NULL; av1dec->input_state = NULL; return TRUE; } static gboolean gst_av1_dec_stop (GstVideoDecoder * dec) { GstAV1Dec *av1dec = GST_AV1_DEC_CAST (dec); if (av1dec->output_state) { gst_video_codec_state_unref (av1dec->output_state); av1dec->output_state = NULL; } if (av1dec->input_state) { gst_video_codec_state_unref (av1dec->input_state); av1dec->input_state = NULL; } if (av1dec->decoder_inited) { aom_codec_destroy (&av1dec->decoder); } av1dec->decoder_inited = FALSE; return TRUE; } static gboolean gst_av1_dec_set_format (GstVideoDecoder * dec, GstVideoCodecState * state) { GstAV1Dec *av1dec = GST_AV1_DEC_CAST (dec); if (av1dec->decoder_inited) { aom_codec_destroy (&av1dec->decoder); } av1dec->decoder_inited = FALSE; if (av1dec->output_state) { gst_video_codec_state_unref (av1dec->output_state); av1dec->output_state = NULL; } if (av1dec->input_state) { gst_video_codec_state_unref (av1dec->input_state); } av1dec->input_state = gst_video_codec_state_ref (state); return TRUE; } static gboolean gst_av1_dec_flush (GstVideoDecoder * dec) { GstAV1Dec *av1dec = GST_AV1_DEC_CAST (dec); if (av1dec->output_state) { gst_video_codec_state_unref (av1dec->output_state); av1dec->output_state = NULL; } if (av1dec->decoder_inited) { aom_codec_destroy (&av1dec->decoder); } av1dec->decoder_inited = FALSE; return TRUE; } static GstFlowReturn gst_av1_dec_open_codec (GstAV1Dec * av1dec, GstVideoCodecFrame * frame) { aom_codec_err_t status; GstAV1DecClass *av1class = GST_AV1_DEC_GET_CLASS (av1dec); status = aom_codec_dec_init (&av1dec->decoder, av1class->codec_algo, NULL, 0); if (status != AOM_CODEC_OK) { GST_ELEMENT_ERROR (av1dec, LIBRARY, INIT, ("Failed to initialize AOM decoder"), ("%s", "")); return GST_FLOW_ERROR; } av1dec->decoder_inited = TRUE; return GST_FLOW_OK; } static void gst_av1_dec_handle_resolution_change (GstAV1Dec * av1dec, aom_image_t * img, GstVideoFormat fmt) { if (!av1dec->output_state || av1dec->output_state->info.finfo->format != fmt || av1dec->output_state->info.width != img->d_w || av1dec->output_state->info.height != img->d_h) { if (av1dec->output_state) gst_video_codec_state_unref (av1dec->output_state); av1dec->output_state = gst_video_decoder_set_output_state (GST_VIDEO_DECODER (av1dec), fmt, img->d_w, img->d_h, av1dec->input_state); gst_video_decoder_negotiate (GST_VIDEO_DECODER (av1dec)); } } static void gst_av1_dec_image_to_buffer (GstAV1Dec * dec, const aom_image_t * img, GstBuffer * buffer) { int deststride, srcstride, height, width, line, comp, y; guint8 *dest, *src; GstVideoFrame frame; GstVideoInfo *info = &dec->output_state->info; if (!gst_video_frame_map (&frame, info, buffer, GST_MAP_WRITE)) { GST_ERROR_OBJECT (dec, "Could not map video buffer"); return; } for (comp = 0; comp < 3; comp++) { dest = GST_VIDEO_FRAME_COMP_DATA (&frame, comp); src = img->planes[comp]; width = GST_VIDEO_FRAME_COMP_WIDTH (&frame, comp) * GST_VIDEO_FRAME_COMP_PSTRIDE (&frame, comp); height = GST_VIDEO_FRAME_COMP_HEIGHT (&frame, comp); deststride = GST_VIDEO_FRAME_COMP_STRIDE (&frame, comp); srcstride = img->stride[comp]; if ((img->fmt & AOM_IMG_FMT_HIGHBITDEPTH) && img->bit_depth == 8) { GST_TRACE_OBJECT (dec, "HIGHBITDEPTH image with 8 bit_depth. Comp %d: %d != %d, copying " "line by line.", comp, srcstride, deststride); for (line = 0; line < height; line++) { for (y = 0; y < width; y++) { dest[y] = src[y * 2]; } dest += deststride; src += srcstride; } } else if (srcstride == deststride) { GST_TRACE_OBJECT (dec, "Stride matches. Comp %d: %d, copying full plane", comp, srcstride); memcpy (dest, src, srcstride * height); } else { GST_TRACE_OBJECT (dec, "Stride mismatch. Comp %d: %d != %d, copying " "line by line.", comp, srcstride, deststride); for (line = 0; line < height; line++) { memcpy (dest, src, width); dest += deststride; src += srcstride; } } } gst_video_frame_unmap (&frame); } gboolean gst_av1_dec_get_valid_format (GstAV1Dec * dec, const aom_image_t * img, GstVideoFormat * fmt) { switch (img->fmt) { case AOM_IMG_FMT_I420: case AOM_IMG_FMT_I42016: if (img->bit_depth == 8) { *fmt = img->monochrome ? GST_VIDEO_FORMAT_GRAY8 : GST_VIDEO_FORMAT_I420; return TRUE; } else if (img->bit_depth == 10) { *fmt = AOM_FMT_TO_GST (I420_10); return TRUE; } else if (img->bit_depth == 12) { *fmt = AOM_FMT_TO_GST (I420_12); return TRUE; } GST_FIXME_OBJECT (dec, "Please add a 4:2:0 planar %u bit depth frame format", img->bit_depth); GST_ELEMENT_WARNING (dec, STREAM, NOT_IMPLEMENTED, (NULL), ("Unsupported frame format - 4:2:0 planar %u bit depth", img->bit_depth)); return FALSE; case AOM_IMG_FMT_I422: case AOM_IMG_FMT_I42216: if (img->bit_depth == 8) { *fmt = GST_VIDEO_FORMAT_Y42B; return TRUE; } else if (img->bit_depth == 10) { *fmt = AOM_FMT_TO_GST (I422_10); return TRUE; } else if (img->bit_depth == 12) { *fmt = AOM_FMT_TO_GST (I422_12); return TRUE; } GST_FIXME_OBJECT (dec, "Please add a 4:2:2 planar %u bit depth frame format", img->bit_depth); GST_ELEMENT_WARNING (dec, STREAM, NOT_IMPLEMENTED, (NULL), ("Unsupported frame format - 4:2:2 planar %u bit depth", img->bit_depth)); return FALSE; case AOM_IMG_FMT_I444: case AOM_IMG_FMT_I44416: if (img->bit_depth == 8) { *fmt = GST_VIDEO_FORMAT_Y444; return TRUE; } else if (img->bit_depth == 10) { *fmt = AOM_FMT_TO_GST (Y444_10); return TRUE; } else if (img->bit_depth == 12) { *fmt = AOM_FMT_TO_GST (Y444_12); return TRUE; } GST_FIXME_OBJECT (dec, "Please add a 4:4:4 planar %u bit depth frame format", img->bit_depth); GST_ELEMENT_WARNING (dec, STREAM, NOT_IMPLEMENTED, (NULL), ("Unsupported frame format - 4:4:4 planar %u bit depth", img->bit_depth)); return FALSE; case AOM_IMG_FMT_YV12: *fmt = GST_VIDEO_FORMAT_YV12; return TRUE; default: return FALSE; } } static GstFlowReturn gst_av1_dec_handle_frame (GstVideoDecoder * dec, GstVideoCodecFrame * frame) { GstAV1Dec *av1dec = GST_AV1_DEC_CAST (dec); GstFlowReturn ret; GstMapInfo minfo; aom_codec_err_t status; aom_image_t *img; aom_codec_iter_t iter = NULL; GstVideoFormat fmt; if (!av1dec->decoder_inited) { ret = gst_av1_dec_open_codec (av1dec, frame); if (ret == GST_FLOW_CUSTOM_SUCCESS_1) { gst_video_decoder_drop_frame (dec, frame); return GST_FLOW_OK; } else if (ret != GST_FLOW_OK) { gst_video_codec_frame_unref (frame); return ret; } } if (!gst_buffer_map (frame->input_buffer, &minfo, GST_MAP_READ)) { GST_ERROR_OBJECT (dec, "Failed to map input buffer"); gst_video_codec_frame_unref (frame); return GST_FLOW_ERROR; } status = aom_codec_decode (&av1dec->decoder, minfo.data, minfo.size, NULL); gst_buffer_unmap (frame->input_buffer, &minfo); if (status) { GST_ELEMENT_ERROR (av1dec, LIBRARY, INIT, ("Failed to decode frame"), ("%s", "")); gst_video_codec_frame_unref (frame); return ret; } img = aom_codec_get_frame (&av1dec->decoder, &iter); if (img) { if (gst_av1_dec_get_valid_format (av1dec, img, &fmt) == FALSE) { aom_img_free (img); GST_ELEMENT_ERROR (dec, LIBRARY, ENCODE, ("Failed to decode frame"), ("Unsupported color format %d", img->fmt)); gst_video_codec_frame_unref (frame); return GST_FLOW_ERROR; } gst_av1_dec_handle_resolution_change (av1dec, img, fmt); ret = gst_video_decoder_allocate_output_frame (dec, frame); if (ret == GST_FLOW_OK) { gst_av1_dec_image_to_buffer (av1dec, img, frame->output_buffer); ret = gst_video_decoder_finish_frame (dec, frame); } else { gst_video_decoder_drop_frame (dec, frame); } aom_img_free (img); while ((img = aom_codec_get_frame (&av1dec->decoder, &iter))) { GST_WARNING_OBJECT (dec, "Multiple decoded frames... dropping"); aom_img_free (img); } } else { GST_VIDEO_CODEC_FRAME_SET_DECODE_ONLY (frame); gst_video_decoder_finish_frame (dec, frame); } return ret; }