gstreamer/subprojects/gst-plugins-bad/gst-libs/gst/codecs/gstmpeg2picture.h
Seungha Yang 2ede4011bf codecs: Keep track of non-decoding-essential input state change
In theory, input caps can be updated anytime at non-keyframe or
sequence boundary, such as HDR10 metadata, framerate, aspect-ratio
or so. Those information update might not trigger ::new_sequence()
or subclass may ignore the changes.

By this commit, input state change will be tracked by baseclass
and subclass will be able to know the non-decoding-essential
update by checking the codec specific picture struct
on ::output_picture()

Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/3328>
2022-11-16 13:12:38 +00:00

196 lines
4.9 KiB
C

/* GStreamer
* Copyright (C) 2020 Intel Corporation
* Author: He Junyan <junyan.he@intel.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Library General Public
* License as published by the Free Software Foundation; either
* version 2 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Library General Public License for more details.
*
* You should have received a copy of the GNU Library General Public
* License along with this library; if not, write to the
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
* Boston, MA 02110-1301, USA.
*/
#ifndef __GST_MPEG2_PICTURE_H__
#define __GST_MPEG2_PICTURE_H__
#include <gst/codecs/codecs-prelude.h>
#include <gst/codecparsers/gstmpegvideoparser.h>
#include <gst/video/video.h>
G_BEGIN_DECLS
/**
* GST_TYPE_MPEG2_PICTURE:
*
* Since: 1.20
*/
#define GST_TYPE_MPEG2_PICTURE (gst_mpeg2_picture_get_type())
/**
* GST_IS_MPEG2_PICTURE:
*
* Since: 1.20
*/
#define GST_IS_MPEG2_PICTURE(obj) (GST_IS_MINI_OBJECT_TYPE(obj, GST_TYPE_MPEG2_PICTURE))
/**
* GST_MPEG2_PICTURE:
*
* Since: 1.20
*/
#define GST_MPEG2_PICTURE(obj) ((GstMpeg2Picture *)obj)
typedef struct _GstMpeg2Slice GstMpeg2Slice;
typedef struct _GstMpeg2Picture GstMpeg2Picture;
/**
* GstMpeg2Slice:
*
* Since: 1.20
*/
struct _GstMpeg2Slice
{
/*< private >*/
GstMpegVideoQuantMatrixExt *quant_matrix; /* The parameter set */
GstMpegVideoPictureHdr *pic_hdr;
GstMpegVideoPictureExt *pic_ext;
GstMpegVideoSliceHdr header;
/* parsed video packet (doesn't take ownership of raw data) */
GstMpegVideoPacket packet;
/* offset of the start code for the slice */
guint sc_offset;
/* size, including the start code */
guint size;
};
/**
* GstMpeg2Picture:
*
* Since: 1.20
*/
struct _GstMpeg2Picture
{
/*< private >*/
GstMiniObject parent;
/* From GstVideoCodecFrame */
guint32 system_frame_number;
gboolean needed_for_output;
/* For interlaced streams */
GstMpeg2Picture *first_field;
GstVideoBufferFlags buffer_flags;
gint pic_order_cnt;
gint tsn;
GstMpegVideoPictureStructure structure;
GstMpegVideoPictureType type;
/* decoder input state if this picture is discont point */
GstVideoCodecState *discont_state;
gpointer user_data;
GDestroyNotify notify;
};
/**
* GST_MPEG2_PICTURE_IS_REF:
* @picture: a #GstMpeg2Picture
*
* Check whether @picture's type is I or P
*
* Since: 1.20
*/
#define GST_MPEG2_PICTURE_IS_REF(picture) \
(((GstMpeg2Picture *) picture)->type == GST_MPEG_VIDEO_PICTURE_TYPE_I || \
((GstMpeg2Picture *) picture)->type == GST_MPEG_VIDEO_PICTURE_TYPE_P)
GST_CODECS_API
GType gst_mpeg2_picture_get_type (void);
GST_CODECS_API
GstMpeg2Picture * gst_mpeg2_picture_new (void);
static inline GstMpeg2Picture *
gst_mpeg2_picture_ref (GstMpeg2Picture * picture)
{
return (GstMpeg2Picture *) gst_mini_object_ref (GST_MINI_OBJECT_CAST (picture));
}
static inline void
gst_mpeg2_picture_unref (GstMpeg2Picture * picture)
{
gst_mini_object_unref (GST_MINI_OBJECT_CAST (picture));
}
static inline gboolean
gst_mpeg2_picture_replace (GstMpeg2Picture ** old_picture,
GstMpeg2Picture * new_picture)
{
return gst_mini_object_replace ((GstMiniObject **) old_picture,
(GstMiniObject *) new_picture);
}
static inline void
gst_clear_mpeg2_picture (GstMpeg2Picture ** picture)
{
if (picture && *picture) {
gst_mpeg2_picture_unref (*picture);
*picture = NULL;
}
}
GST_CODECS_API
void gst_mpeg2_picture_set_user_data (GstMpeg2Picture * picture,
gpointer user_data,
GDestroyNotify notify);
GST_CODECS_API
gpointer gst_mpeg2_picture_get_user_data (GstMpeg2Picture * picture);
/**
* GstMpeg2Dpb:
*
* Since: 1.20
*/
typedef struct _GstMpeg2Dpb GstMpeg2Dpb;
GST_CODECS_API
GstMpeg2Dpb * gst_mpeg2_dpb_new (void);
GST_CODECS_API
void gst_mpeg2_dpb_free (GstMpeg2Dpb * dpb);
GST_CODECS_API
void gst_mpeg2_dpb_clear (GstMpeg2Dpb * dpb);
GST_CODECS_API
void gst_mpeg2_dpb_add (GstMpeg2Dpb * dpb,
GstMpeg2Picture * picture);
GST_CODECS_API
GstMpeg2Picture * gst_mpeg2_dpb_bump (GstMpeg2Dpb * dpb);
GST_CODECS_API
gboolean gst_mpeg2_dpb_need_bump (GstMpeg2Dpb * dpb);
GST_CODECS_API
void gst_mpeg2_dpb_get_neighbours (GstMpeg2Dpb * dpb,
GstMpeg2Picture * picture,
GstMpeg2Picture ** prev_picture_ptr,
GstMpeg2Picture ** next_picture_ptr);
G_DEFINE_AUTOPTR_CLEANUP_FUNC(GstMpeg2Picture, gst_mpeg2_picture_unref)
G_END_DECLS
#endif /* __GST_MPEG2_PICTURE_H__ */