mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-11-23 10:11:08 +00:00
814 lines
26 KiB
C
814 lines
26 KiB
C
/* GStreamer
|
|
* Copyright (C) <2005> Wim Taymans <wim.taymans@gmail.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
# include "config.h"
|
|
#endif
|
|
|
|
#include <string.h>
|
|
#include <stdlib.h>
|
|
#include <stdio.h>
|
|
|
|
#include <gst/rtp/gstrtpbuffer.h>
|
|
#include <gst/video/video.h>
|
|
|
|
#include "gstrtpelements.h"
|
|
#include "gstrtph263ppay.h"
|
|
#include "gstrtputils.h"
|
|
|
|
#define DEFAULT_FRAGMENTATION_MODE GST_FRAGMENTATION_MODE_NORMAL
|
|
|
|
enum
|
|
{
|
|
PROP_0,
|
|
PROP_FRAGMENTATION_MODE
|
|
};
|
|
|
|
#define GST_TYPE_FRAGMENTATION_MODE (gst_fragmentation_mode_get_type())
|
|
static GType
|
|
gst_fragmentation_mode_get_type (void)
|
|
{
|
|
static GType fragmentation_mode_type = 0;
|
|
static const GEnumValue fragmentation_mode[] = {
|
|
{GST_FRAGMENTATION_MODE_NORMAL, "Normal", "normal"},
|
|
{GST_FRAGMENTATION_MODE_SYNC, "Fragment at sync points", "sync"},
|
|
{0, NULL, NULL},
|
|
};
|
|
|
|
if (!fragmentation_mode_type) {
|
|
fragmentation_mode_type =
|
|
g_enum_register_static ("GstFragmentationMode", fragmentation_mode);
|
|
}
|
|
return fragmentation_mode_type;
|
|
}
|
|
|
|
|
|
GST_DEBUG_CATEGORY_STATIC (rtph263ppay_debug);
|
|
#define GST_CAT_DEFAULT rtph263ppay_debug
|
|
|
|
static GstStaticPadTemplate gst_rtp_h263p_pay_sink_template =
|
|
GST_STATIC_PAD_TEMPLATE ("sink",
|
|
GST_PAD_SINK,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("video/x-h263, variant = (string) itu")
|
|
);
|
|
|
|
/*
|
|
* We also return these in getcaps() as required by the SDP caps
|
|
*
|
|
* width = (int) [16, 4096]
|
|
* height = (int) [16, 4096]
|
|
* "annex-f = (boolean) {true, false},"
|
|
* "annex-i = (boolean) {true, false},"
|
|
* "annex-j = (boolean) {true, false},"
|
|
* "annex-l = (boolean) {true, false},"
|
|
* "annex-t = (boolean) {true, false},"
|
|
* "annex-v = (boolean) {true, false}")
|
|
*/
|
|
|
|
|
|
static GstStaticPadTemplate gst_rtp_h263p_pay_src_template =
|
|
GST_STATIC_PAD_TEMPLATE ("src",
|
|
GST_PAD_SRC,
|
|
GST_PAD_ALWAYS,
|
|
GST_STATIC_CAPS ("application/x-rtp, "
|
|
"media = (string) \"video\", "
|
|
"payload = (int) " GST_RTP_PAYLOAD_DYNAMIC_STRING ", "
|
|
"clock-rate = (int) 90000, " "encoding-name = (string) \"H263-1998\"; "
|
|
"application/x-rtp, "
|
|
"media = (string) \"video\", "
|
|
"payload = (int) " GST_RTP_PAYLOAD_DYNAMIC_STRING ", "
|
|
"clock-rate = (int) 90000, " "encoding-name = (string) \"H263-2000\"")
|
|
);
|
|
|
|
static void gst_rtp_h263p_pay_finalize (GObject * object);
|
|
|
|
static void gst_rtp_h263p_pay_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec);
|
|
static void gst_rtp_h263p_pay_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec);
|
|
|
|
static gboolean gst_rtp_h263p_pay_setcaps (GstRTPBasePayload * payload,
|
|
GstCaps * caps);
|
|
static GstCaps *gst_rtp_h263p_pay_sink_getcaps (GstRTPBasePayload * payload,
|
|
GstPad * pad, GstCaps * filter);
|
|
static GstFlowReturn gst_rtp_h263p_pay_handle_buffer (GstRTPBasePayload *
|
|
payload, GstBuffer * buffer);
|
|
|
|
#define gst_rtp_h263p_pay_parent_class parent_class
|
|
G_DEFINE_TYPE (GstRtpH263PPay, gst_rtp_h263p_pay, GST_TYPE_RTP_BASE_PAYLOAD);
|
|
GST_ELEMENT_REGISTER_DEFINE_WITH_CODE (rtph263ppay, "rtph263ppay",
|
|
GST_RANK_SECONDARY, GST_TYPE_RTP_H263P_PAY, rtp_element_init (plugin));
|
|
|
|
static void
|
|
gst_rtp_h263p_pay_class_init (GstRtpH263PPayClass * klass)
|
|
{
|
|
GObjectClass *gobject_class;
|
|
GstElementClass *gstelement_class;
|
|
GstRTPBasePayloadClass *gstrtpbasepayload_class;
|
|
|
|
gobject_class = (GObjectClass *) klass;
|
|
gstelement_class = (GstElementClass *) klass;
|
|
gstrtpbasepayload_class = (GstRTPBasePayloadClass *) klass;
|
|
|
|
gobject_class->finalize = gst_rtp_h263p_pay_finalize;
|
|
gobject_class->set_property = gst_rtp_h263p_pay_set_property;
|
|
gobject_class->get_property = gst_rtp_h263p_pay_get_property;
|
|
|
|
gstrtpbasepayload_class->set_caps = gst_rtp_h263p_pay_setcaps;
|
|
gstrtpbasepayload_class->get_caps = gst_rtp_h263p_pay_sink_getcaps;
|
|
gstrtpbasepayload_class->handle_buffer = gst_rtp_h263p_pay_handle_buffer;
|
|
|
|
g_object_class_install_property (G_OBJECT_CLASS (klass),
|
|
PROP_FRAGMENTATION_MODE, g_param_spec_enum ("fragmentation-mode",
|
|
"Fragmentation Mode",
|
|
"Packet Fragmentation Mode", GST_TYPE_FRAGMENTATION_MODE,
|
|
DEFAULT_FRAGMENTATION_MODE,
|
|
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS));
|
|
|
|
gst_element_class_add_static_pad_template (gstelement_class,
|
|
&gst_rtp_h263p_pay_src_template);
|
|
gst_element_class_add_static_pad_template (gstelement_class,
|
|
&gst_rtp_h263p_pay_sink_template);
|
|
|
|
gst_element_class_set_static_metadata (gstelement_class, "RTP H263 payloader",
|
|
"Codec/Payloader/Network/RTP",
|
|
"Payload-encodes H263/+/++ video in RTP packets (RFC 4629)",
|
|
"Wim Taymans <wim.taymans@gmail.com>");
|
|
|
|
GST_DEBUG_CATEGORY_INIT (rtph263ppay_debug, "rtph263ppay",
|
|
0, "rtph263ppay (RFC 4629)");
|
|
|
|
gst_type_mark_as_plugin_api (GST_TYPE_FRAGMENTATION_MODE, 0);
|
|
}
|
|
|
|
static void
|
|
gst_rtp_h263p_pay_init (GstRtpH263PPay * rtph263ppay)
|
|
{
|
|
rtph263ppay->adapter = gst_adapter_new ();
|
|
|
|
rtph263ppay->fragmentation_mode = DEFAULT_FRAGMENTATION_MODE;
|
|
}
|
|
|
|
static void
|
|
gst_rtp_h263p_pay_finalize (GObject * object)
|
|
{
|
|
GstRtpH263PPay *rtph263ppay;
|
|
|
|
rtph263ppay = GST_RTP_H263P_PAY (object);
|
|
|
|
g_object_unref (rtph263ppay->adapter);
|
|
rtph263ppay->adapter = NULL;
|
|
|
|
G_OBJECT_CLASS (parent_class)->finalize (object);
|
|
}
|
|
|
|
static gboolean
|
|
gst_rtp_h263p_pay_setcaps (GstRTPBasePayload * payload, GstCaps * caps)
|
|
{
|
|
gboolean res;
|
|
GstCaps *peercaps;
|
|
gchar *encoding_name = NULL;
|
|
|
|
g_return_val_if_fail (gst_caps_is_fixed (caps), FALSE);
|
|
|
|
peercaps =
|
|
gst_pad_peer_query_caps (GST_RTP_BASE_PAYLOAD_SRCPAD (payload), NULL);
|
|
if (peercaps) {
|
|
GstCaps *tcaps =
|
|
gst_pad_get_pad_template_caps (GST_RTP_BASE_PAYLOAD_SRCPAD (payload));
|
|
GstCaps *intersect = gst_caps_intersect (peercaps, tcaps);
|
|
gst_caps_unref (tcaps);
|
|
|
|
gst_caps_unref (peercaps);
|
|
if (!gst_caps_is_empty (intersect)) {
|
|
GstStructure *s = gst_caps_get_structure (intersect, 0);
|
|
encoding_name = g_strdup (gst_structure_get_string (s, "encoding-name"));
|
|
}
|
|
gst_caps_unref (intersect);
|
|
}
|
|
|
|
if (!encoding_name)
|
|
encoding_name = g_strdup ("H263-1998");
|
|
|
|
gst_rtp_base_payload_set_options (payload, "video", TRUE,
|
|
(gchar *) encoding_name, 90000);
|
|
res = gst_rtp_base_payload_set_outcaps (payload, NULL);
|
|
g_free (encoding_name);
|
|
|
|
return res;
|
|
}
|
|
|
|
static GstCaps *
|
|
caps_append (GstCaps * caps, GstStructure * in_s, guint x, guint y, guint mpi)
|
|
{
|
|
GstStructure *s;
|
|
|
|
if (!in_s)
|
|
return caps;
|
|
|
|
if (mpi < 1 || mpi > 32)
|
|
return caps;
|
|
|
|
s = gst_structure_copy (in_s);
|
|
|
|
gst_structure_set (s,
|
|
"width", GST_TYPE_INT_RANGE, 1, x,
|
|
"height", GST_TYPE_INT_RANGE, 1, y,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 30000, 1001 * mpi, NULL);
|
|
|
|
caps = gst_caps_merge_structure (caps, s);
|
|
|
|
return caps;
|
|
}
|
|
|
|
|
|
static GstCaps *
|
|
gst_rtp_h263p_pay_sink_getcaps (GstRTPBasePayload * payload, GstPad * pad,
|
|
GstCaps * filter)
|
|
{
|
|
GstRtpH263PPay *rtph263ppay;
|
|
GstCaps *caps = NULL, *templ;
|
|
GstCaps *peercaps = NULL;
|
|
GstCaps *intersect = NULL;
|
|
guint i;
|
|
|
|
rtph263ppay = GST_RTP_H263P_PAY (payload);
|
|
|
|
peercaps =
|
|
gst_pad_peer_query_caps (GST_RTP_BASE_PAYLOAD_SRCPAD (payload), NULL);
|
|
|
|
/* if we're just outputting to udpsink or fakesink or so, we should also
|
|
* accept any input compatible with our sink template caps */
|
|
if (!peercaps || gst_caps_is_any (peercaps)) {
|
|
if (peercaps)
|
|
gst_caps_unref (peercaps);
|
|
caps =
|
|
gst_pad_get_pad_template_caps (GST_RTP_BASE_PAYLOAD_SINKPAD (payload));
|
|
goto done;
|
|
}
|
|
|
|
/* We basically need to differentiate two use-cases here: One where there's
|
|
* a capsfilter after the payloader with caps created from an SDP; in this
|
|
* case the filter caps are fixed and we want to signal to an encoder what
|
|
* we want it to produce. The second case is simply payloader ! depayloader
|
|
* where we are dealing with the depayloader's template caps. In this case
|
|
* we should accept any input compatible with our sink template caps. */
|
|
if (!gst_caps_is_fixed (peercaps)) {
|
|
gst_caps_unref (peercaps);
|
|
caps =
|
|
gst_pad_get_pad_template_caps (GST_RTP_BASE_PAYLOAD_SINKPAD (payload));
|
|
goto done;
|
|
}
|
|
|
|
templ = gst_pad_get_pad_template_caps (GST_RTP_BASE_PAYLOAD_SRCPAD (payload));
|
|
intersect = gst_caps_intersect (peercaps, templ);
|
|
gst_caps_unref (peercaps);
|
|
gst_caps_unref (templ);
|
|
|
|
if (gst_caps_is_empty (intersect))
|
|
return intersect;
|
|
|
|
caps = gst_caps_new_empty ();
|
|
for (i = 0; i < gst_caps_get_size (intersect); i++) {
|
|
GstStructure *s = gst_caps_get_structure (intersect, i);
|
|
const gchar *encoding_name = gst_structure_get_string (s, "encoding-name");
|
|
|
|
if (!strcmp (encoding_name, "H263-2000")) {
|
|
const gchar *profile_str = gst_structure_get_string (s, "profile");
|
|
const gchar *level_str = gst_structure_get_string (s, "level");
|
|
int profile = 0;
|
|
int level = 0;
|
|
|
|
if (profile_str && level_str) {
|
|
gboolean i = FALSE, j = FALSE, l = FALSE, t = FALSE, f = FALSE,
|
|
v = FALSE;
|
|
GstStructure *new_s = gst_structure_new ("video/x-h263",
|
|
"variant", G_TYPE_STRING, "itu",
|
|
NULL);
|
|
|
|
profile = atoi (profile_str);
|
|
level = atoi (level_str);
|
|
|
|
/* These profiles are defined in the H.263 Annex X */
|
|
switch (profile) {
|
|
case 0:
|
|
/* The Baseline Profile (Profile 0) */
|
|
break;
|
|
case 1:
|
|
/* H.320 Coding Efficiency Version 2 Backward-Compatibility Profile
|
|
* (Profile 1)
|
|
* Baseline + Annexes I, J, L.4 and T
|
|
*/
|
|
i = j = l = t = TRUE;
|
|
break;
|
|
case 2:
|
|
/* Version 1 Backward-Compatibility Profile (Profile 2)
|
|
* Baseline + Annex F
|
|
*/
|
|
i = j = l = t = f = TRUE;
|
|
break;
|
|
case 3:
|
|
/* Version 2 Interactive and Streaming Wireless Profile
|
|
* Baseline + Annexes I, J, T
|
|
*/
|
|
i = j = t = TRUE;
|
|
break;
|
|
case 4:
|
|
/* Version 3 Interactive and Streaming Wireless Profile (Profile 4)
|
|
* Baseline + Annexes I, J, T, V, W.6.3.8,
|
|
*/
|
|
/* Missing W.6.3.8 */
|
|
i = j = t = v = TRUE;
|
|
break;
|
|
case 5:
|
|
/* Conversational High Compression Profile (Profile 5)
|
|
* Baseline + Annexes F, I, J, L.4, T, D, U
|
|
*/
|
|
/* Missing D, U */
|
|
f = i = j = l = t = TRUE;
|
|
break;
|
|
case 6:
|
|
/* Conversational Internet Profile (Profile 6)
|
|
* Baseline + Annexes F, I, J, L.4, T, D, U and
|
|
* K with arbitratry slice ordering
|
|
*/
|
|
/* Missing D, U, K with arbitratry slice ordering */
|
|
f = i = j = l = t = TRUE;
|
|
break;
|
|
case 7:
|
|
/* Conversational Interlace Profile (Profile 7)
|
|
* Baseline + Annexes F, I, J, L.4, T, D, U, W.6.3.11
|
|
*/
|
|
/* Missing D, U, W.6.3.11 */
|
|
f = i = j = l = t = TRUE;
|
|
break;
|
|
case 8:
|
|
/* High Latency Profile (Profile 8)
|
|
* Baseline + Annexes F, I, J, L.4, T, D, U, P.5, O.1.1 and
|
|
* K with arbitratry slice ordering
|
|
*/
|
|
/* Missing D, U, P.5, O.1.1 */
|
|
f = i = j = l = t = TRUE;
|
|
break;
|
|
}
|
|
|
|
|
|
if (f || i || j || t || l || v) {
|
|
GValue list = { 0 };
|
|
GValue vstr = { 0 };
|
|
|
|
g_value_init (&list, GST_TYPE_LIST);
|
|
g_value_init (&vstr, G_TYPE_STRING);
|
|
|
|
g_value_set_static_string (&vstr, "h263");
|
|
gst_value_list_append_value (&list, &vstr);
|
|
g_value_set_static_string (&vstr, "h263p");
|
|
gst_value_list_append_value (&list, &vstr);
|
|
|
|
if (l || v) {
|
|
g_value_set_static_string (&vstr, "h263pp");
|
|
gst_value_list_append_value (&list, &vstr);
|
|
}
|
|
g_value_unset (&vstr);
|
|
|
|
gst_structure_set_value (new_s, "h263version", &list);
|
|
g_value_unset (&list);
|
|
} else {
|
|
gst_structure_set (new_s, "h263version", G_TYPE_STRING, "h263", NULL);
|
|
}
|
|
|
|
|
|
if (!f)
|
|
gst_structure_set (new_s, "annex-f", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!i)
|
|
gst_structure_set (new_s, "annex-i", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!j)
|
|
gst_structure_set (new_s, "annex-j", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!t)
|
|
gst_structure_set (new_s, "annex-t", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!l)
|
|
gst_structure_set (new_s, "annex-l", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!v)
|
|
gst_structure_set (new_s, "annex-v", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
|
|
|
|
if (level <= 10 || level == 45) {
|
|
gst_structure_set (new_s,
|
|
"width", GST_TYPE_INT_RANGE, 1, 176,
|
|
"height", GST_TYPE_INT_RANGE, 1, 144,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 30000, 2002, NULL);
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
} else if (level <= 20) {
|
|
GstStructure *s_copy = gst_structure_copy (new_s);
|
|
|
|
gst_structure_set (new_s,
|
|
"width", GST_TYPE_INT_RANGE, 1, 352,
|
|
"height", GST_TYPE_INT_RANGE, 1, 288,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 30000, 2002, NULL);
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
|
|
gst_structure_set (s_copy,
|
|
"width", GST_TYPE_INT_RANGE, 1, 176,
|
|
"height", GST_TYPE_INT_RANGE, 1, 144,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 30000, 1001, NULL);
|
|
caps = gst_caps_merge_structure (caps, s_copy);
|
|
} else if (level <= 40) {
|
|
|
|
gst_structure_set (new_s,
|
|
"width", GST_TYPE_INT_RANGE, 1, 352,
|
|
"height", GST_TYPE_INT_RANGE, 1, 288,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 30000, 1001, NULL);
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
} else if (level <= 50) {
|
|
GstStructure *s_copy = gst_structure_copy (new_s);
|
|
|
|
gst_structure_set (new_s,
|
|
"width", GST_TYPE_INT_RANGE, 1, 352,
|
|
"height", GST_TYPE_INT_RANGE, 1, 288,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 50, 1, NULL);
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
|
|
gst_structure_set (s_copy,
|
|
"width", GST_TYPE_INT_RANGE, 1, 352,
|
|
"height", GST_TYPE_INT_RANGE, 1, 240,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 60000, 1001, NULL);
|
|
caps = gst_caps_merge_structure (caps, s_copy);
|
|
} else if (level <= 60) {
|
|
GstStructure *s_copy = gst_structure_copy (new_s);
|
|
|
|
gst_structure_set (new_s,
|
|
"width", GST_TYPE_INT_RANGE, 1, 720,
|
|
"height", GST_TYPE_INT_RANGE, 1, 288,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 50, 1, NULL);
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
|
|
gst_structure_set (s_copy,
|
|
"width", GST_TYPE_INT_RANGE, 1, 720,
|
|
"height", GST_TYPE_INT_RANGE, 1, 240,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 60000, 1001, NULL);
|
|
caps = gst_caps_merge_structure (caps, s_copy);
|
|
} else if (level <= 70) {
|
|
GstStructure *s_copy = gst_structure_copy (new_s);
|
|
|
|
gst_structure_set (new_s,
|
|
"width", GST_TYPE_INT_RANGE, 1, 720,
|
|
"height", GST_TYPE_INT_RANGE, 1, 576,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 50, 1, NULL);
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
|
|
gst_structure_set (s_copy,
|
|
"width", GST_TYPE_INT_RANGE, 1, 720,
|
|
"height", GST_TYPE_INT_RANGE, 1, 480,
|
|
"framerate", GST_TYPE_FRACTION_RANGE, 0, 1, 60000, 1001, NULL);
|
|
caps = gst_caps_merge_structure (caps, s_copy);
|
|
} else {
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
}
|
|
|
|
} else {
|
|
GstStructure *new_s = gst_structure_new ("video/x-h263",
|
|
"variant", G_TYPE_STRING, "itu",
|
|
"h263version", G_TYPE_STRING, "h263",
|
|
NULL);
|
|
|
|
GST_DEBUG_OBJECT (rtph263ppay, "No profile or level specified"
|
|
" for H263-2000, defaulting to baseline H263");
|
|
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
}
|
|
} else {
|
|
gboolean f = FALSE, i = FALSE, j = FALSE, t = FALSE;
|
|
/* FIXME: ffmpeg support the Appendix K too, how do we express it ?
|
|
* guint k;
|
|
*/
|
|
const gchar *str;
|
|
GstStructure *new_s = gst_structure_new ("video/x-h263",
|
|
"variant", G_TYPE_STRING, "itu",
|
|
NULL);
|
|
gboolean added = FALSE;
|
|
|
|
str = gst_structure_get_string (s, "f");
|
|
if (str && !strcmp (str, "1"))
|
|
f = TRUE;
|
|
|
|
str = gst_structure_get_string (s, "i");
|
|
if (str && !strcmp (str, "1"))
|
|
i = TRUE;
|
|
|
|
str = gst_structure_get_string (s, "j");
|
|
if (str && !strcmp (str, "1"))
|
|
j = TRUE;
|
|
|
|
str = gst_structure_get_string (s, "t");
|
|
if (str && !strcmp (str, "1"))
|
|
t = TRUE;
|
|
|
|
if (f || i || j || t) {
|
|
GValue list = { 0 };
|
|
GValue vstr = { 0 };
|
|
|
|
g_value_init (&list, GST_TYPE_LIST);
|
|
g_value_init (&vstr, G_TYPE_STRING);
|
|
|
|
g_value_set_static_string (&vstr, "h263");
|
|
gst_value_list_append_value (&list, &vstr);
|
|
g_value_set_static_string (&vstr, "h263p");
|
|
gst_value_list_append_value (&list, &vstr);
|
|
g_value_unset (&vstr);
|
|
|
|
gst_structure_set_value (new_s, "h263version", &list);
|
|
g_value_unset (&list);
|
|
} else {
|
|
gst_structure_set (new_s, "h263version", G_TYPE_STRING, "h263", NULL);
|
|
}
|
|
|
|
if (!f)
|
|
gst_structure_set (new_s, "annex-f", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!i)
|
|
gst_structure_set (new_s, "annex-i", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!j)
|
|
gst_structure_set (new_s, "annex-j", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
if (!t)
|
|
gst_structure_set (new_s, "annex-t", G_TYPE_BOOLEAN, FALSE, NULL);
|
|
|
|
|
|
str = gst_structure_get_string (s, "custom");
|
|
if (str) {
|
|
unsigned int xmax, ymax, mpi;
|
|
if (sscanf (str, "%u,%u,%u", &xmax, &ymax, &mpi) == 3) {
|
|
if (xmax % 4 && ymax % 4 && mpi >= 1 && mpi <= 32) {
|
|
caps = caps_append (caps, new_s, xmax, ymax, mpi);
|
|
added = TRUE;
|
|
} else {
|
|
GST_WARNING_OBJECT (rtph263ppay, "Invalid custom framesize/MPI"
|
|
" %u x %u at %u, ignoring", xmax, ymax, mpi);
|
|
}
|
|
} else {
|
|
GST_WARNING_OBJECT (rtph263ppay, "Invalid custom framesize/MPI: %s,"
|
|
" ignoring", str);
|
|
}
|
|
}
|
|
|
|
str = gst_structure_get_string (s, "16cif");
|
|
if (str) {
|
|
int mpi = atoi (str);
|
|
caps = caps_append (caps, new_s, 1408, 1152, mpi);
|
|
added = TRUE;
|
|
}
|
|
|
|
str = gst_structure_get_string (s, "4cif");
|
|
if (str) {
|
|
int mpi = atoi (str);
|
|
caps = caps_append (caps, new_s, 704, 576, mpi);
|
|
added = TRUE;
|
|
}
|
|
|
|
str = gst_structure_get_string (s, "cif");
|
|
if (str) {
|
|
int mpi = atoi (str);
|
|
caps = caps_append (caps, new_s, 352, 288, mpi);
|
|
added = TRUE;
|
|
}
|
|
|
|
str = gst_structure_get_string (s, "qcif");
|
|
if (str) {
|
|
int mpi = atoi (str);
|
|
caps = caps_append (caps, new_s, 176, 144, mpi);
|
|
added = TRUE;
|
|
}
|
|
|
|
str = gst_structure_get_string (s, "sqcif");
|
|
if (str) {
|
|
int mpi = atoi (str);
|
|
caps = caps_append (caps, new_s, 128, 96, mpi);
|
|
added = TRUE;
|
|
}
|
|
|
|
if (added)
|
|
gst_structure_free (new_s);
|
|
else
|
|
caps = gst_caps_merge_structure (caps, new_s);
|
|
}
|
|
}
|
|
|
|
gst_caps_unref (intersect);
|
|
|
|
done:
|
|
|
|
if (filter) {
|
|
GstCaps *tmp;
|
|
|
|
GST_DEBUG_OBJECT (payload, "Intersect %" GST_PTR_FORMAT " and filter %"
|
|
GST_PTR_FORMAT, caps, filter);
|
|
tmp = gst_caps_intersect_full (filter, caps, GST_CAPS_INTERSECT_FIRST);
|
|
gst_caps_unref (caps);
|
|
caps = tmp;
|
|
}
|
|
|
|
return caps;
|
|
}
|
|
|
|
|
|
static void
|
|
gst_rtp_h263p_pay_set_property (GObject * object, guint prop_id,
|
|
const GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstRtpH263PPay *rtph263ppay;
|
|
|
|
rtph263ppay = GST_RTP_H263P_PAY (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_FRAGMENTATION_MODE:
|
|
rtph263ppay->fragmentation_mode = g_value_get_enum (value);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void
|
|
gst_rtp_h263p_pay_get_property (GObject * object, guint prop_id,
|
|
GValue * value, GParamSpec * pspec)
|
|
{
|
|
GstRtpH263PPay *rtph263ppay;
|
|
|
|
rtph263ppay = GST_RTP_H263P_PAY (object);
|
|
|
|
switch (prop_id) {
|
|
case PROP_FRAGMENTATION_MODE:
|
|
g_value_set_enum (value, rtph263ppay->fragmentation_mode);
|
|
break;
|
|
default:
|
|
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_rtp_h263p_pay_flush (GstRtpH263PPay * rtph263ppay)
|
|
{
|
|
guint avail;
|
|
GstBufferList *list = NULL;
|
|
GstBuffer *outbuf = NULL;
|
|
GstFlowReturn ret;
|
|
gboolean fragmented = FALSE;
|
|
|
|
avail = gst_adapter_available (rtph263ppay->adapter);
|
|
if (avail == 0)
|
|
return GST_FLOW_OK;
|
|
|
|
fragmented = FALSE;
|
|
/* This algorithm assumes the H263/+/++ encoder sends complete frames in each
|
|
* buffer */
|
|
/* With Fragmentation Mode at GST_FRAGMENTATION_MODE_NORMAL:
|
|
* This algorithm implements the Follow-on packets method for packetization.
|
|
* This assumes low packet loss network.
|
|
* With Fragmentation Mode at GST_FRAGMENTATION_MODE_SYNC:
|
|
* This algorithm separates large frames at synchronisation points (Segments)
|
|
* (See RFC 4629 section 6). It would be interesting to have a property such as network
|
|
* quality to select between both packetization methods */
|
|
/* TODO Add VRC support (See RFC 4629 section 5.2) */
|
|
|
|
while (avail > 0) {
|
|
guint towrite;
|
|
guint8 *payload;
|
|
gint header_len;
|
|
guint next_gop = 0;
|
|
gboolean found_gob = FALSE;
|
|
GstRTPBuffer rtp = { NULL };
|
|
GstBuffer *payload_buf;
|
|
|
|
if (rtph263ppay->fragmentation_mode == GST_FRAGMENTATION_MODE_SYNC) {
|
|
/* start after 1st gop possible */
|
|
|
|
/* Check if we have a gob or eos , eossbs */
|
|
/* FIXME EOS and EOSSBS packets should never contain any gobs and vice-versa */
|
|
next_gop =
|
|
gst_adapter_masked_scan_uint32 (rtph263ppay->adapter, 0xffff8000,
|
|
0x00008000, 0, avail);
|
|
if (next_gop == 0) {
|
|
GST_DEBUG_OBJECT (rtph263ppay, " Found GOB header");
|
|
found_gob = TRUE;
|
|
}
|
|
|
|
/* Find next and cut the packet accordingly */
|
|
/* TODO we should get as many gobs as possible until MTU is reached, this
|
|
* code seems to just get one GOB per packet */
|
|
if (next_gop == 0 && avail > 3)
|
|
next_gop =
|
|
gst_adapter_masked_scan_uint32 (rtph263ppay->adapter, 0xffff8000,
|
|
0x00008000, 3, avail - 3);
|
|
GST_DEBUG_OBJECT (rtph263ppay, " Next GOB Detected at : %d", next_gop);
|
|
if (next_gop == -1)
|
|
next_gop = 0;
|
|
}
|
|
|
|
/* for picture start frames (non-fragmented), we need to remove the first
|
|
* two 0x00 bytes and set P=1 */
|
|
if (!fragmented || found_gob) {
|
|
gst_adapter_flush (rtph263ppay->adapter, 2);
|
|
avail -= 2;
|
|
}
|
|
header_len = 2;
|
|
|
|
towrite = MIN (avail, gst_rtp_buffer_calc_payload_len
|
|
(GST_RTP_BASE_PAYLOAD_MTU (rtph263ppay) - header_len, 0, 0));
|
|
|
|
if (next_gop > 0)
|
|
towrite = MIN (next_gop, towrite);
|
|
|
|
outbuf =
|
|
gst_rtp_base_payload_allocate_output_buffer (GST_RTP_BASE_PAYLOAD
|
|
(rtph263ppay), header_len, 0, 0);
|
|
|
|
gst_rtp_buffer_map (outbuf, GST_MAP_WRITE, &rtp);
|
|
/* last fragment gets the marker bit set */
|
|
gst_rtp_buffer_set_marker (&rtp, avail > towrite ? 0 : 1);
|
|
|
|
payload = gst_rtp_buffer_get_payload (&rtp);
|
|
|
|
/* 0 1
|
|
* 0 1 2 3 4 5 6 7 8 9 0 1 2 3 4 5
|
|
* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|
|
* | RR |P|V| PLEN |PEBIT|
|
|
* +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
|
|
*/
|
|
/* if fragmented or gop header , write p bit =1 */
|
|
payload[0] = (fragmented && !found_gob) ? 0x00 : 0x04;
|
|
payload[1] = 0;
|
|
|
|
GST_BUFFER_PTS (outbuf) = rtph263ppay->first_timestamp;
|
|
GST_BUFFER_DURATION (outbuf) = rtph263ppay->first_duration;
|
|
gst_rtp_buffer_unmap (&rtp);
|
|
|
|
payload_buf = gst_adapter_take_buffer_fast (rtph263ppay->adapter, towrite);
|
|
gst_rtp_copy_video_meta (rtph263ppay, outbuf, payload_buf);
|
|
outbuf = gst_buffer_append (outbuf, payload_buf);
|
|
avail -= towrite;
|
|
|
|
/* If more data is available and this is our first iteration,
|
|
* we create a buffer list and remember that we're fragmented.
|
|
*
|
|
* If we're fragmented already, add buffers to the previously
|
|
* created buffer list.
|
|
*
|
|
* Otherwise fragmented will be FALSE and we just push the single output
|
|
* buffer, and no list is allocated.
|
|
*/
|
|
if (avail && !fragmented) {
|
|
fragmented = TRUE;
|
|
list = gst_buffer_list_new ();
|
|
gst_buffer_list_add (list, outbuf);
|
|
} else if (fragmented) {
|
|
gst_buffer_list_add (list, outbuf);
|
|
}
|
|
}
|
|
|
|
if (fragmented) {
|
|
ret =
|
|
gst_rtp_base_payload_push_list (GST_RTP_BASE_PAYLOAD (rtph263ppay),
|
|
list);
|
|
} else {
|
|
ret =
|
|
gst_rtp_base_payload_push (GST_RTP_BASE_PAYLOAD (rtph263ppay), outbuf);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstFlowReturn
|
|
gst_rtp_h263p_pay_handle_buffer (GstRTPBasePayload * payload,
|
|
GstBuffer * buffer)
|
|
{
|
|
GstRtpH263PPay *rtph263ppay;
|
|
GstFlowReturn ret;
|
|
|
|
rtph263ppay = GST_RTP_H263P_PAY (payload);
|
|
|
|
rtph263ppay->first_timestamp = GST_BUFFER_PTS (buffer);
|
|
rtph263ppay->first_duration = GST_BUFFER_DURATION (buffer);
|
|
|
|
/* we always encode and flush a full picture */
|
|
gst_adapter_push (rtph263ppay->adapter, buffer);
|
|
ret = gst_rtp_h263p_pay_flush (rtph263ppay);
|
|
|
|
return ret;
|
|
}
|