/* GStreamer * Copyright (C) <1999> Erik Walthinsen * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 02111-1307, USA. */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include #include "gstjpegdec.h" #include #include "gst/gst-i18n-plugin.h" #include GstElementDetails gst_jpeg_dec_details = { "JPEG image decoder", "Codec/Decoder/Image", "Decode images from JPEG format", "Wim Taymans ", }; #define MIN_WIDTH 16 #define MAX_WIDTH 4096 #define MIN_HEIGHT 16 #define MAX_HEIGHT 4096 static GstStaticPadTemplate gst_jpeg_dec_src_pad_template = GST_STATIC_PAD_TEMPLATE ("src", GST_PAD_SRC, GST_PAD_ALWAYS, GST_STATIC_CAPS (GST_VIDEO_CAPS_YUV ("I420")) ); static GstStaticPadTemplate gst_jpeg_dec_sink_pad_template = GST_STATIC_PAD_TEMPLATE ("sink", GST_PAD_SINK, GST_PAD_ALWAYS, GST_STATIC_CAPS ("image/jpeg, " "width = (int) [ " G_STRINGIFY (MIN_WIDTH) ", " G_STRINGIFY (MAX_WIDTH) " ], " "height = (int) [ " G_STRINGIFY (MIN_HEIGHT) ", " G_STRINGIFY (MAX_HEIGHT) " ], " "framerate = (fraction) [ 0/1, MAX ]") ); GST_DEBUG_CATEGORY_STATIC (jpeg_dec_debug); #define GST_CAT_DEFAULT jpeg_dec_debug /* These macros are adapted from videotestsrc.c * and/or gst-plugins/gst/games/gstvideoimage.c */ #define I420_Y_ROWSTRIDE(width) (GST_ROUND_UP_4(width)) #define I420_U_ROWSTRIDE(width) (GST_ROUND_UP_8(width)/2) #define I420_V_ROWSTRIDE(width) ((GST_ROUND_UP_8(I420_Y_ROWSTRIDE(width)))/2) #define I420_Y_OFFSET(w,h) (0) #define I420_U_OFFSET(w,h) (I420_Y_OFFSET(w,h)+(I420_Y_ROWSTRIDE(w)*GST_ROUND_UP_2(h))) #define I420_V_OFFSET(w,h) (I420_U_OFFSET(w,h)+(I420_U_ROWSTRIDE(w)*GST_ROUND_UP_2(h)/2)) #define I420_SIZE(w,h) (I420_V_OFFSET(w,h)+(I420_V_ROWSTRIDE(w)*GST_ROUND_UP_2(h)/2)) static GstElementClass *parent_class; /* NULL */ static void gst_jpeg_dec_base_init (gpointer g_class); static void gst_jpeg_dec_class_init (GstJpegDecClass * klass); static void gst_jpeg_dec_init (GstJpegDec * jpegdec); static GstFlowReturn gst_jpeg_dec_chain (GstPad * pad, GstBuffer * buffer); static gboolean gst_jpeg_dec_setcaps (GstPad * pad, GstCaps * caps); static GstStateChangeReturn gst_jpeg_dec_change_state (GstElement * element, GstStateChange transition); GType gst_jpeg_dec_get_type (void) { static GType type = 0; if (!type) { static const GTypeInfo jpeg_dec_info = { sizeof (GstJpegDecClass), (GBaseInitFunc) gst_jpeg_dec_base_init, NULL, (GClassInitFunc) gst_jpeg_dec_class_init, NULL, NULL, sizeof (GstJpegDec), 0, (GInstanceInitFunc) gst_jpeg_dec_init, }; type = g_type_register_static (GST_TYPE_ELEMENT, "GstJpegDec", &jpeg_dec_info, 0); } return type; } static void gst_jpeg_dec_finalize (GObject * object) { GstJpegDec *dec = GST_JPEG_DEC (object); jpeg_destroy_decompress (&dec->cinfo); if (dec->tempbuf) gst_buffer_unref (dec->tempbuf); G_OBJECT_CLASS (parent_class)->finalize (object); } static void gst_jpeg_dec_base_init (gpointer g_class) { GstElementClass *element_class = GST_ELEMENT_CLASS (g_class); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_jpeg_dec_src_pad_template)); gst_element_class_add_pad_template (element_class, gst_static_pad_template_get (&gst_jpeg_dec_sink_pad_template)); gst_element_class_set_details (element_class, &gst_jpeg_dec_details); } static void gst_jpeg_dec_class_init (GstJpegDecClass * klass) { GstElementClass *gstelement_class; GObjectClass *gobject_class; gstelement_class = (GstElementClass *) klass; gobject_class = (GObjectClass *) klass; parent_class = g_type_class_peek_parent (klass); gobject_class->finalize = gst_jpeg_dec_finalize; gstelement_class->change_state = GST_DEBUG_FUNCPTR (gst_jpeg_dec_change_state); GST_DEBUG_CATEGORY_INIT (jpeg_dec_debug, "jpegdec", 0, "JPEG decoder"); } static gboolean gst_jpeg_dec_fill_input_buffer (j_decompress_ptr cinfo) { /* struct GstJpegDecSourceMgr *src_mgr; GstJpegDec *dec; src_mgr = (struct GstJpegDecSourceMgr*) &cinfo->src; dec = GST_JPEG_DEC (src_mgr->dec); */ GST_DEBUG ("fill_input_buffer"); /* g_return_val_if_fail (dec != NULL, TRUE); src_mgr->pub.next_input_byte = GST_BUFFER_DATA (dec->tempbuf); src_mgr->pub.bytes_in_buffer = GST_BUFFER_SIZE (dec->tempbuf); */ return TRUE; } static void gst_jpeg_dec_init_source (j_decompress_ptr cinfo) { GST_DEBUG ("init_source"); } static void gst_jpeg_dec_skip_input_data (j_decompress_ptr cinfo, glong num_bytes) { GST_DEBUG ("skip_input_data: %ld bytes", num_bytes); if (num_bytes > 0 && cinfo->src->bytes_in_buffer >= num_bytes) { cinfo->src->next_input_byte += (size_t) num_bytes; cinfo->src->bytes_in_buffer -= (size_t) num_bytes; } } static gboolean gst_jpeg_dec_resync_to_restart (j_decompress_ptr cinfo, gint desired) { GST_DEBUG ("resync_to_start"); return TRUE; } static void gst_jpeg_dec_term_source (j_decompress_ptr cinfo) { GST_DEBUG ("term_source"); return; } METHODDEF (void) gst_jpeg_dec_my_output_message (j_common_ptr cinfo) { return; /* do nothing */ } METHODDEF (void) gst_jpeg_dec_my_emit_message (j_common_ptr cinfo, int msg_level) { /* GST_DEBUG ("emit_message: msg_level = %d", msg_level); */ return; } METHODDEF (void) gst_jpeg_dec_my_error_exit (j_common_ptr cinfo) { struct GstJpegDecErrorMgr *err_mgr = (struct GstJpegDecErrorMgr *) cinfo->err; (*cinfo->err->output_message) (cinfo); longjmp (err_mgr->setjmp_buffer, 1); } static void gst_jpeg_dec_init (GstJpegDec * dec) { GST_DEBUG ("initializing"); /* create the sink and src pads */ dec->sinkpad = gst_pad_new_from_static_template (&gst_jpeg_dec_sink_pad_template, "sink"); gst_element_add_pad (GST_ELEMENT (dec), dec->sinkpad); gst_pad_set_setcaps_function (dec->sinkpad, GST_DEBUG_FUNCPTR (gst_jpeg_dec_setcaps)); gst_pad_set_chain_function (dec->sinkpad, GST_DEBUG_FUNCPTR (gst_jpeg_dec_chain)); dec->srcpad = gst_pad_new_from_static_template (&gst_jpeg_dec_src_pad_template, "src"); gst_element_add_pad (GST_ELEMENT (dec), dec->srcpad); /* setup jpeglib */ memset (&dec->cinfo, 0, sizeof (dec->cinfo)); memset (&dec->jerr, 0, sizeof (dec->jerr)); dec->cinfo.err = jpeg_std_error (&dec->jerr.pub); dec->jerr.pub.output_message = gst_jpeg_dec_my_output_message; dec->jerr.pub.emit_message = gst_jpeg_dec_my_emit_message; dec->jerr.pub.error_exit = gst_jpeg_dec_my_error_exit; jpeg_create_decompress (&dec->cinfo); dec->cinfo.src = (struct jpeg_source_mgr *) &dec->jsrc; dec->cinfo.src->init_source = gst_jpeg_dec_init_source; dec->cinfo.src->fill_input_buffer = gst_jpeg_dec_fill_input_buffer; dec->cinfo.src->skip_input_data = gst_jpeg_dec_skip_input_data; dec->cinfo.src->resync_to_restart = gst_jpeg_dec_resync_to_restart; dec->cinfo.src->term_source = gst_jpeg_dec_term_source; dec->jsrc.dec = dec; } static inline gboolean is_jpeg_start_marker (const guint8 * data) { return (data[0] == 0xff && data[1] == 0xd8); } static inline gboolean is_jpeg_end_marker (const guint8 * data) { return (data[0] == 0xff && data[1] == 0xd9); } static gboolean gst_jpeg_dec_find_jpeg_header (GstJpegDec * dec) { const guint8 *data; guint size; data = GST_BUFFER_DATA (dec->tempbuf); size = GST_BUFFER_SIZE (dec->tempbuf); g_return_val_if_fail (size >= 2, FALSE); while (!is_jpeg_start_marker (data) || data[2] != 0xff) { const guint8 *marker; GstBuffer *tmp; guint off; marker = memchr (data + 1, 0xff, size - 1 - 2); if (marker == NULL) { off = size - 1; /* keep last byte */ } else { off = marker - data; } tmp = gst_buffer_create_sub (dec->tempbuf, off, size - off); gst_buffer_unref (dec->tempbuf); dec->tempbuf = tmp; data = GST_BUFFER_DATA (dec->tempbuf); size = GST_BUFFER_SIZE (dec->tempbuf); if (size < 2) return FALSE; /* wait for more data */ } return TRUE; /* got header */ } static gboolean gst_jpeg_dec_ensure_header (GstJpegDec * dec) { g_return_val_if_fail (dec->tempbuf != NULL, FALSE); check_header: /* we need at least a start marker (0xff 0xd8) * and an end marker (0xff 0xd9) */ if (GST_BUFFER_SIZE (dec->tempbuf) <= 4) { GST_DEBUG ("Not enough data"); return FALSE; /* we need more data */ } if (!is_jpeg_start_marker (GST_BUFFER_DATA (dec->tempbuf))) { GST_DEBUG ("Not a JPEG header, resyncing to header..."); if (!gst_jpeg_dec_find_jpeg_header (dec)) { GST_DEBUG ("No JPEG header in current buffer"); return FALSE; /* we need more data */ } GST_DEBUG ("Found JPEG header"); goto check_header; /* buffer might have changed */ } return TRUE; } #if 0 static gboolean gst_jpeg_dec_have_end_marker (GstJpegDec * dec) { guint8 *data = GST_BUFFER_DATA (dec->tempbuf); guint size = GST_BUFFER_SIZE (dec->tempbuf); return (size > 2 && data && is_jpeg_end_marker (data + size - 2)); } #endif static inline gboolean gst_jpeg_dec_parse_tag_has_entropy_segment (guint8 tag) { if (tag == 0xda || (tag >= 0xd0 && tag <= 0xd7)) return TRUE; return FALSE; } /* returns image length in bytes if parsed * successfully, otherwise 0 */ static guint gst_jpeg_dec_parse_image_data (GstJpegDec * dec) { guint8 *start, *data, *end; guint size; size = GST_BUFFER_SIZE (dec->tempbuf); start = GST_BUFFER_DATA (dec->tempbuf); end = start + size; data = start; g_return_val_if_fail (is_jpeg_start_marker (data), 0); GST_DEBUG ("Parsing jpeg image data (%u bytes)", size); /* skip start marker */ data += 2; while (1) { guint frame_len; /* enough bytes left for EOI marker? (we need 0xff 0xNN, thus end-1) */ if (data >= end - 1) { GST_DEBUG ("at end of input and no EOI marker found, need more data"); return 0; } if (is_jpeg_end_marker (data)) { GST_DEBUG ("0x%08x: end marker", data - start); goto found_eoi; } /* do we need to resync? */ if (*data != 0xff) { GST_DEBUG ("Lost sync at 0x%08x, resyncing", data - start); /* at the very least we expect 0xff 0xNN, thus end-1 */ while (*data != 0xff && data < end - 1) ++data; if (is_jpeg_end_marker (data)) { GST_DEBUG ("resynced to end marker"); goto found_eoi; } /* we need 0xFF 0xNN 0xLL 0xLL */ if (data >= end - 1 - 2) { GST_DEBUG ("at end of input, without new sync, need more data"); return 0; } /* check if we will still be in sync if we interpret * this as a sync point and skip this frame */ frame_len = GST_READ_UINT16_BE (data + 2); GST_DEBUG ("possible sync at 0x%08x, frame_len=%u", data - start, frame_len); if (data + 2 + frame_len >= end - 1 || data[2 + frame_len] != 0xff) { /* ignore and continue resyncing until we hit the end * of our data or find a sync point that looks okay */ ++data; continue; } GST_DEBUG ("found sync at 0x%08x", data - size); } while (*data == 0xff) ++data; if (data + 2 >= end) return 0; if (*data >= 0xd0 && *data <= 0xd7) frame_len = 0; else frame_len = GST_READ_UINT16_BE (data + 1); GST_DEBUG ("0x%08x: tag %02x, frame_len=%u", data - start - 1, *data, frame_len); /* the frame length includes the 2 bytes for the length; here we want at * least 2 more bytes at the end for an end marker, thus end-2 */ if (data + 1 + frame_len >= end - 2) { /* theoretically we could have lost sync and not really need more * data, but that's just tough luck and a broken image then */ GST_DEBUG ("at end of input and no EOI marker found, need more data"); return 0; } if (gst_jpeg_dec_parse_tag_has_entropy_segment (*data)) { guint8 *d2 = data + 1 + frame_len; guint eseglen = 0; GST_DEBUG ("0x%08x: finding entropy segment length", data - start - 1); while (1) { if (d2[eseglen] == 0xff && d2[eseglen + 1] != 0x00) break; if (d2 + eseglen >= end - 1) return 0; /* need more data */ ++eseglen; } frame_len += eseglen; GST_DEBUG ("entropy segment length=%u => frame_len=%u", eseglen, frame_len); } data += 1 + frame_len; } found_eoi: /* data is assumed to point to the 0xff sync point of the * EOI marker (so there is one more byte after that) */ g_assert (is_jpeg_end_marker (data)); return ((data + 1) - start + 1); } /* shamelessly ripped from jpegutils.c in mjpegtools */ static void add_huff_table (j_decompress_ptr dinfo, JHUFF_TBL ** htblptr, const UINT8 * bits, const UINT8 * val) /* Define a Huffman table */ { int nsymbols, len; if (*htblptr == NULL) *htblptr = jpeg_alloc_huff_table ((j_common_ptr) dinfo); /* Copy the number-of-symbols-of-each-code-length counts */ memcpy ((*htblptr)->bits, bits, sizeof ((*htblptr)->bits)); /* Validate the counts. We do this here mainly so we can copy the right * number of symbols from the val[] array, without risking marching off * the end of memory. jchuff.c will do a more thorough test later. */ nsymbols = 0; for (len = 1; len <= 16; len++) nsymbols += bits[len]; if (nsymbols < 1 || nsymbols > 256) g_error ("jpegutils.c: add_huff_table failed badly. "); memcpy ((*htblptr)->huffval, val, nsymbols * sizeof (UINT8)); } static void std_huff_tables (j_decompress_ptr dinfo) /* Set up the standard Huffman tables (cf. JPEG standard section K.3) */ /* IMPORTANT: these are only valid for 8-bit data precision! */ { static const UINT8 bits_dc_luminance[17] = { /* 0-base */ 0, 0, 1, 5, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0 }; static const UINT8 val_dc_luminance[] = { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 }; static const UINT8 bits_dc_chrominance[17] = { /* 0-base */ 0, 0, 3, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0 }; static const UINT8 val_dc_chrominance[] = { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 }; static const UINT8 bits_ac_luminance[17] = { /* 0-base */ 0, 0, 2, 1, 3, 3, 2, 4, 3, 5, 5, 4, 4, 0, 0, 1, 0x7d }; static const UINT8 val_ac_luminance[] = { 0x01, 0x02, 0x03, 0x00, 0x04, 0x11, 0x05, 0x12, 0x21, 0x31, 0x41, 0x06, 0x13, 0x51, 0x61, 0x07, 0x22, 0x71, 0x14, 0x32, 0x81, 0x91, 0xa1, 0x08, 0x23, 0x42, 0xb1, 0xc1, 0x15, 0x52, 0xd1, 0xf0, 0x24, 0x33, 0x62, 0x72, 0x82, 0x09, 0x0a, 0x16, 0x17, 0x18, 0x19, 0x1a, 0x25, 0x26, 0x27, 0x28, 0x29, 0x2a, 0x34, 0x35, 0x36, 0x37, 0x38, 0x39, 0x3a, 0x43, 0x44, 0x45, 0x46, 0x47, 0x48, 0x49, 0x4a, 0x53, 0x54, 0x55, 0x56, 0x57, 0x58, 0x59, 0x5a, 0x63, 0x64, 0x65, 0x66, 0x67, 0x68, 0x69, 0x6a, 0x73, 0x74, 0x75, 0x76, 0x77, 0x78, 0x79, 0x7a, 0x83, 0x84, 0x85, 0x86, 0x87, 0x88, 0x89, 0x8a, 0x92, 0x93, 0x94, 0x95, 0x96, 0x97, 0x98, 0x99, 0x9a, 0xa2, 0xa3, 0xa4, 0xa5, 0xa6, 0xa7, 0xa8, 0xa9, 0xaa, 0xb2, 0xb3, 0xb4, 0xb5, 0xb6, 0xb7, 0xb8, 0xb9, 0xba, 0xc2, 0xc3, 0xc4, 0xc5, 0xc6, 0xc7, 0xc8, 0xc9, 0xca, 0xd2, 0xd3, 0xd4, 0xd5, 0xd6, 0xd7, 0xd8, 0xd9, 0xda, 0xe1, 0xe2, 0xe3, 0xe4, 0xe5, 0xe6, 0xe7, 0xe8, 0xe9, 0xea, 0xf1, 0xf2, 0xf3, 0xf4, 0xf5, 0xf6, 0xf7, 0xf8, 0xf9, 0xfa }; static const UINT8 bits_ac_chrominance[17] = { /* 0-base */ 0, 0, 2, 1, 2, 4, 4, 3, 4, 7, 5, 4, 4, 0, 1, 2, 0x77 }; static const UINT8 val_ac_chrominance[] = { 0x00, 0x01, 0x02, 0x03, 0x11, 0x04, 0x05, 0x21, 0x31, 0x06, 0x12, 0x41, 0x51, 0x07, 0x61, 0x71, 0x13, 0x22, 0x32, 0x81, 0x08, 0x14, 0x42, 0x91, 0xa1, 0xb1, 0xc1, 0x09, 0x23, 0x33, 0x52, 0xf0, 0x15, 0x62, 0x72, 0xd1, 0x0a, 0x16, 0x24, 0x34, 0xe1, 0x25, 0xf1, 0x17, 0x18, 0x19, 0x1a, 0x26, 0x27, 0x28, 0x29, 0x2a, 0x35, 0x36, 0x37, 0x38, 0x39, 0x3a, 0x43, 0x44, 0x45, 0x46, 0x47, 0x48, 0x49, 0x4a, 0x53, 0x54, 0x55, 0x56, 0x57, 0x58, 0x59, 0x5a, 0x63, 0x64, 0x65, 0x66, 0x67, 0x68, 0x69, 0x6a, 0x73, 0x74, 0x75, 0x76, 0x77, 0x78, 0x79, 0x7a, 0x82, 0x83, 0x84, 0x85, 0x86, 0x87, 0x88, 0x89, 0x8a, 0x92, 0x93, 0x94, 0x95, 0x96, 0x97, 0x98, 0x99, 0x9a, 0xa2, 0xa3, 0xa4, 0xa5, 0xa6, 0xa7, 0xa8, 0xa9, 0xaa, 0xb2, 0xb3, 0xb4, 0xb5, 0xb6, 0xb7, 0xb8, 0xb9, 0xba, 0xc2, 0xc3, 0xc4, 0xc5, 0xc6, 0xc7, 0xc8, 0xc9, 0xca, 0xd2, 0xd3, 0xd4, 0xd5, 0xd6, 0xd7, 0xd8, 0xd9, 0xda, 0xe2, 0xe3, 0xe4, 0xe5, 0xe6, 0xe7, 0xe8, 0xe9, 0xea, 0xf2, 0xf3, 0xf4, 0xf5, 0xf6, 0xf7, 0xf8, 0xf9, 0xfa }; add_huff_table (dinfo, &dinfo->dc_huff_tbl_ptrs[0], bits_dc_luminance, val_dc_luminance); add_huff_table (dinfo, &dinfo->ac_huff_tbl_ptrs[0], bits_ac_luminance, val_ac_luminance); add_huff_table (dinfo, &dinfo->dc_huff_tbl_ptrs[1], bits_dc_chrominance, val_dc_chrominance); add_huff_table (dinfo, &dinfo->ac_huff_tbl_ptrs[1], bits_ac_chrominance, val_ac_chrominance); } static void guarantee_huff_tables (j_decompress_ptr dinfo) { if ((dinfo->dc_huff_tbl_ptrs[0] == NULL) && (dinfo->dc_huff_tbl_ptrs[1] == NULL) && (dinfo->ac_huff_tbl_ptrs[0] == NULL) && (dinfo->ac_huff_tbl_ptrs[1] == NULL)) { GST_DEBUG ("Generating standard Huffman tables for this frame."); std_huff_tables (dinfo); } } static gboolean gst_jpeg_dec_setcaps (GstPad * pad, GstCaps * caps) { GstStructure *s; GstJpegDec *dec; const GValue *framerate; dec = GST_JPEG_DEC (GST_OBJECT_PARENT (pad)); s = gst_caps_get_structure (caps, 0); if ((framerate = gst_structure_get_value (s, "framerate")) != NULL) { dec->framerate_numerator = gst_value_get_fraction_numerator (framerate); dec->framerate_denominator = gst_value_get_fraction_denominator (framerate); dec->packetized = TRUE; GST_DEBUG ("got framerate of %d/%d fps => packetized mode", dec->framerate_numerator, dec->framerate_denominator); } /* do not extract width/height here. we do that in the chain * function on a per-frame basis (including the line[] array * setup) */ /* But we can take the framerate values and set them on the src pad */ return TRUE; } static void gst_jpeg_dec_decode_indirect (GstJpegDec * dec, guchar * base[3], guchar * last[3], guint width, guint height, gint r_v) { guchar y[16][MAX_WIDTH]; guchar u[8][MAX_WIDTH / 2]; guchar v[8][MAX_WIDTH / 2]; guchar *y_rows[16] = { y[0], y[1], y[2], y[3], y[4], y[5], y[6], y[7], y[8], y[9], y[10], y[11], y[12], y[13], y[14], y[15] }; guchar *u_rows[8] = { u[0], u[1], u[2], u[3], u[4], u[5], u[6], u[7] }; guchar *v_rows[8] = { v[0], v[1], v[2], v[3], v[4], v[5], v[6], v[7] }; guchar **scanarray[3] = { y_rows, u_rows, v_rows }; gint i, j, k; GST_DEBUG_OBJECT (dec, "unadvantageous width, taking slow route involving memcpy"); for (i = 0; i < height; i += r_v * DCTSIZE) { jpeg_read_raw_data (&dec->cinfo, scanarray, r_v * DCTSIZE); for (j = 0, k = 0; j < (r_v * DCTSIZE); j += r_v, k++) { memcpy (base[0], y_rows[j], I420_Y_ROWSTRIDE (width)); if (base[0] < last[0]) base[0] += I420_Y_ROWSTRIDE (width); if (r_v == 2) { memcpy (base[0], y_rows[j + 1], I420_Y_ROWSTRIDE (width)); if (base[0] < last[0]) base[0] += I420_Y_ROWSTRIDE (width); } memcpy (base[1], u_rows[k], I420_U_ROWSTRIDE (width)); memcpy (base[2], v_rows[k], I420_V_ROWSTRIDE (width)); if (r_v == 2 || (k & 1) != 0) { if (base[1] < last[1] && base[2] < last[2]) { base[1] += I420_U_ROWSTRIDE (width); base[2] += I420_V_ROWSTRIDE (width); } } } } } static void gst_jpeg_dec_decode_direct (GstJpegDec * dec, guchar * base[3], guchar * last[3], guint width, guint height, gint r_v) { guchar **line[3]; /* the jpeg line buffer */ gint i, j, k; line[0] = g_new0 (guchar *, (r_v * DCTSIZE)); line[1] = g_new0 (guchar *, (r_v * DCTSIZE)); line[2] = g_new0 (guchar *, (r_v * DCTSIZE)); /* let jpeglib decode directly into our final buffer */ GST_DEBUG_OBJECT (dec, "decoding directly into output buffer"); for (i = 0; i < height; i += r_v * DCTSIZE) { for (j = 0, k = 0; j < (r_v * DCTSIZE); j += r_v, k++) { line[0][j] = base[0]; if (base[0] < last[0]) base[0] += I420_Y_ROWSTRIDE (width); if (r_v == 2) { line[0][j + 1] = base[0]; if (base[0] < last[0]) base[0] += I420_Y_ROWSTRIDE (width); } line[1][k] = base[1]; line[2][k] = base[2]; if (r_v == 2 || (k & 1) != 0) { if (base[1] < last[1] && base[2] < last[2]) { base[1] += I420_U_ROWSTRIDE (width); base[2] += I420_V_ROWSTRIDE (width); } } } jpeg_read_raw_data (&dec->cinfo, line, r_v * DCTSIZE); } g_free (line[0]); g_free (line[1]); g_free (line[2]); } static GstFlowReturn gst_jpeg_dec_chain (GstPad * pad, GstBuffer * buf) { GstFlowReturn ret; GstJpegDec *dec; GstBuffer *outbuf; gulong size; guchar *data, *outdata; guchar *base[3], *last[3]; guint img_len; gint width, height; gint r_h, r_v; gint i; guint code; GstClockTime timestamp, duration; dec = GST_JPEG_DEC (GST_OBJECT_PARENT (pad)); timestamp = GST_BUFFER_TIMESTAMP (buf); duration = GST_BUFFER_DURATION (buf); /* GST_LOG_OBJECT (dec, "Received buffer: %d bytes, ts=%" GST_TIME_FORMAT ", dur=%" GST_TIME_FORMAT, GST_BUFFER_SIZE (buf), GST_TIME_ARGS (timestamp), GST_TIME_ARGS (duration)); */ if (GST_CLOCK_TIME_IS_VALID (timestamp)) { dec->next_ts = timestamp; } if (dec->tempbuf) { dec->tempbuf = gst_buffer_join (dec->tempbuf, buf); } else { dec->tempbuf = buf; } buf = NULL; if (!gst_jpeg_dec_ensure_header (dec)) goto need_more_data; /* If we know that each input buffer contains data * for a whole jpeg image (e.g. MJPEG streams), just * do some sanity checking instead of parsing all of * the jpeg data */ if (dec->packetized) { img_len = GST_BUFFER_SIZE (dec->tempbuf); } else { /* Parse jpeg image to handle jpeg input that * is not aligned to buffer boundaries */ img_len = gst_jpeg_dec_parse_image_data (dec); if (img_len == 0) goto need_more_data; } data = (guchar *) GST_BUFFER_DATA (dec->tempbuf); size = img_len; GST_LOG_OBJECT (dec, "image size = %u", img_len); dec->jsrc.pub.next_input_byte = data; dec->jsrc.pub.bytes_in_buffer = size; if (setjmp (dec->jerr.setjmp_buffer)) { code = dec->jerr.pub.msg_code; if (code == JERR_INPUT_EOF) { GST_DEBUG ("jpeg input EOF error, we probably need more data"); goto need_more_data; } goto decode_error; } GST_LOG_OBJECT (dec, "reading header %02x %02x %02x %02x", data[0], data[1], data[2], data[3]); jpeg_read_header (&dec->cinfo, TRUE); r_h = dec->cinfo.cur_comp_info[0]->h_samp_factor; r_v = dec->cinfo.cur_comp_info[0]->v_samp_factor; GST_DEBUG ("num_components=%d, comps_in_scan=%d\n", dec->cinfo.num_components, dec->cinfo.comps_in_scan); for (i = 0; i < dec->cinfo.num_components; ++i) { GST_DEBUG ("[%d] h_samp_factor=%d, v_samp_factor=%d\n", i, dec->cinfo.cur_comp_info[i]->h_samp_factor, dec->cinfo.cur_comp_info[i]->v_samp_factor); } dec->cinfo.do_fancy_upsampling = FALSE; dec->cinfo.do_block_smoothing = FALSE; dec->cinfo.out_color_space = JCS_YCbCr; dec->cinfo.dct_method = JDCT_IFAST; dec->cinfo.raw_data_out = TRUE; GST_LOG_OBJECT (dec, "starting decompress"); guarantee_huff_tables (&dec->cinfo); jpeg_start_decompress (&dec->cinfo); width = dec->cinfo.output_width; height = dec->cinfo.output_height; if (width < MIN_WIDTH || width > MAX_WIDTH || height < MIN_HEIGHT || height > MAX_HEIGHT) goto wrong_size; if (width != dec->caps_width || height != dec->caps_height || dec->framerate_numerator != dec->caps_framerate_numerator || dec->framerate_denominator != dec->caps_framerate_denominator) { GstCaps *caps; /* framerate == 0/1 is a still frame */ if (dec->framerate_denominator == 0) { dec->framerate_numerator = 0; dec->framerate_denominator = 1; } caps = gst_caps_new_simple ("video/x-raw-yuv", "format", GST_TYPE_FOURCC, GST_MAKE_FOURCC ('I', '4', '2', '0'), "width", G_TYPE_INT, width, "height", G_TYPE_INT, height, "framerate", GST_TYPE_FRACTION, dec->framerate_numerator, dec->framerate_denominator, NULL); GST_DEBUG_OBJECT (dec, "setting caps %" GST_PTR_FORMAT, caps); GST_DEBUG_OBJECT (dec, "max_v_samp_factor=%d", dec->cinfo.max_v_samp_factor); gst_pad_set_caps (dec->srcpad, caps); gst_caps_unref (caps); dec->caps_width = width; dec->caps_height = height; dec->caps_framerate_numerator = dec->framerate_numerator; dec->caps_framerate_denominator = dec->framerate_denominator; } ret = gst_pad_alloc_buffer_and_set_caps (dec->srcpad, GST_BUFFER_OFFSET_NONE, I420_SIZE (width, height), GST_PAD_CAPS (dec->srcpad), &outbuf); if (ret != GST_FLOW_OK) goto alloc_failed; outdata = GST_BUFFER_DATA (outbuf); GST_BUFFER_TIMESTAMP (outbuf) = dec->next_ts; if (dec->packetized && GST_CLOCK_TIME_IS_VALID (dec->next_ts)) { if (GST_CLOCK_TIME_IS_VALID (duration)) { /* use duration from incoming buffer for outgoing buffer */ dec->next_ts += duration; } else if (dec->framerate_numerator != 0) { duration = gst_util_uint64_scale (GST_SECOND, dec->framerate_denominator, dec->framerate_numerator); dec->next_ts += duration; } else { duration = GST_CLOCK_TIME_NONE; dec->next_ts = GST_CLOCK_TIME_NONE; } } else { duration = GST_CLOCK_TIME_NONE; dec->next_ts = GST_CLOCK_TIME_NONE; } GST_BUFFER_DURATION (outbuf) = duration; GST_LOG_OBJECT (dec, "width %d, height %d, buffer size %d, required size %d", width, height, GST_BUFFER_SIZE (outbuf), I420_SIZE (width, height)); /* mind the swap, jpeglib outputs blue chroma first */ base[0] = outdata + I420_Y_OFFSET (width, height); base[1] = outdata + I420_U_OFFSET (width, height); base[2] = outdata + I420_V_OFFSET (width, height); /* make sure we don't make jpeglib write beyond our buffer, * which might happen if (height % (r_v*DCTSIZE)) != 0 */ last[0] = base[0] + (I420_Y_ROWSTRIDE (width) * (height - 1)); last[1] = base[1] + (I420_U_ROWSTRIDE (width) * ((GST_ROUND_UP_2 (height) / 2) - 1)); last[2] = base[2] + (I420_V_ROWSTRIDE (width) * ((GST_ROUND_UP_2 (height) / 2) - 1)); GST_LOG_OBJECT (dec, "decompressing %u", dec->cinfo.rec_outbuf_height); GST_LOG_OBJECT (dec, "max_h_samp_factor=%u", dec->cinfo.max_h_samp_factor); /* For some widths jpeglib requires more horizontal padding than I420 * provides. In those cases we need to decode into separate buffers and then * copy over the data into our final picture buffer, otherwise jpeglib might * write over the end of a line into the beginning of the next line, * resulting in blocky artifacts on the left side of the picture. */ if (width % (dec->cinfo.max_h_samp_factor * DCTSIZE) != 0) { gst_jpeg_dec_decode_indirect (dec, base, last, width, height, r_v); } else { gst_jpeg_dec_decode_direct (dec, base, last, width, height, r_v); } GST_LOG_OBJECT (dec, "decompressing finished"); jpeg_finish_decompress (&dec->cinfo); GST_LOG_OBJECT (dec, "pushing buffer (ts=%" GST_TIME_FORMAT ", dur=%" GST_TIME_FORMAT, GST_TIME_ARGS (GST_BUFFER_TIMESTAMP (outbuf)), GST_TIME_ARGS (GST_BUFFER_DURATION (outbuf))); ret = gst_pad_push (dec->srcpad, outbuf); done: if (GST_BUFFER_SIZE (dec->tempbuf) == img_len) { gst_buffer_unref (dec->tempbuf); dec->tempbuf = NULL; } else { GstBuffer *buf = gst_buffer_create_sub (dec->tempbuf, img_len, GST_BUFFER_SIZE (dec->tempbuf) - img_len); gst_buffer_unref (dec->tempbuf); dec->tempbuf = buf; } return ret; /* special cases */ need_more_data: { GST_LOG_OBJECT (dec, "we need more data"); return GST_FLOW_OK; } /* ERRORS */ wrong_size: { GST_ELEMENT_ERROR (dec, STREAM, DECODE, ("Picture is too small or too big (%ux%u)", width, height), ("Picture is too small or too big (%ux%u)", width, height)); ret = GST_FLOW_ERROR; goto done; } decode_error: { GST_ELEMENT_ERROR (dec, STREAM, DECODE, (_("Failed to decode JPEG image")), ("Error #%u: %s", code, dec->jerr.pub.jpeg_message_table[code])); ret = GST_FLOW_ERROR; goto done; } alloc_failed: { const gchar *reason; reason = gst_flow_get_name (ret); GST_DEBUG_OBJECT (dec, "failed to alloc buffer, reason %s", reason); if (GST_FLOW_IS_FATAL (ret)) { GST_ELEMENT_ERROR (dec, STREAM, DECODE, ("Buffer allocation failed, reason: %s", reason), ("Buffer allocation failed, reason: %s", reason)); } return ret; } } static GstStateChangeReturn gst_jpeg_dec_change_state (GstElement * element, GstStateChange transition) { GstStateChangeReturn ret; GstJpegDec *dec; dec = GST_JPEG_DEC (element); switch (transition) { case GST_STATE_CHANGE_READY_TO_PAUSED: dec->framerate_numerator = 0; dec->framerate_denominator = 1; dec->caps_framerate_numerator = dec->caps_framerate_denominator = 0; dec->caps_width = -1; dec->caps_height = -1; dec->packetized = FALSE; dec->next_ts = 0; default: break; } ret = GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); if (ret != GST_STATE_CHANGE_SUCCESS) return ret; switch (transition) { case GST_STATE_CHANGE_PAUSED_TO_READY: if (dec->tempbuf) { gst_buffer_unref (dec->tempbuf); dec->tempbuf = NULL; } break; default: break; } return ret; }