gstreamer/gst-libs/gst/vaapi/gstvaapiencoder_jpeg.c
He Junyan b50f06309a libs: encoder: set context info profile by encoder
Instead of init_context_info() setting the passed profile, it is
assumed that it has to be set by each encoder.

Signed-off-by: Víctor Manuel Jáquez Leal <vjaquez@igalia.com>
2019-12-29 17:58:36 +01:00

893 lines
30 KiB
C

/*
* gstvaapiencoder_jpeg.c - JPEG encoder
*
* Copyright (C) 2015 Intel Corporation
* Author: Sreerenj Balachandran <sreerenj.balachandran@intel.com>
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public License
* as published by the Free Software Foundation; either version 2.1
* of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, write to the Free
* Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
* Boston, MA 02110-1301 USA
*/
#include "sysdeps.h"
#include <gst/base/gstbitwriter.h>
#include <gst/codecparsers/gstjpegparser.h>
#include "gstvaapicompat.h"
#include "gstvaapiencoder_priv.h"
#include "gstvaapiencoder_jpeg.h"
#include "gstvaapicodedbufferproxy_priv.h"
#include "gstvaapisurface.h"
#define DEBUG 1
#include "gstvaapidebug.h"
/* Define default rate control mode ("constant-qp") */
#define DEFAULT_RATECONTROL GST_VAAPI_RATECONTROL_NONE
/* Supported set of VA rate controls, within this implementation */
#define SUPPORTED_RATECONTROLS \
(GST_VAAPI_RATECONTROL_MASK (NONE))
/* Supported set of tuning options, within this implementation */
#define SUPPORTED_TUNE_OPTIONS \
(GST_VAAPI_ENCODER_TUNE_MASK (NONE))
/* Supported set of VA packed headers, within this implementation */
#define SUPPORTED_PACKED_HEADERS \
(VA_ENC_PACKED_HEADER_RAW_DATA)
#define NUM_DC_RUN_SIZE_BITS 16
#define NUM_AC_RUN_SIZE_BITS 16
#define NUM_AC_CODE_WORDS_HUFFVAL 162
#define NUM_DC_CODE_WORDS_HUFFVAL 12
/* ------------------------------------------------------------------------- */
/* --- JPEG Encoder --- */
/* ------------------------------------------------------------------------- */
struct _GstVaapiEncoderJpeg
{
GstVaapiEncoder parent_instance;
GstVaapiProfile profile;
guint quality;
GstJpegQuantTables quant_tables;
GstJpegQuantTables scaled_quant_tables;
gboolean has_quant_tables;
GstJpegHuffmanTables huff_tables;
gboolean has_huff_tables;
gint cwidth[GST_VIDEO_MAX_COMPONENTS];
gint cheight[GST_VIDEO_MAX_COMPONENTS];
gint h_samp[GST_VIDEO_MAX_COMPONENTS];
gint v_samp[GST_VIDEO_MAX_COMPONENTS];
gint h_max_samp;
gint v_max_samp;
guint n_components;
};
/* based on upstream gst-plugins-good jpegencoder */
static void
generate_sampling_factors (GstVaapiEncoderJpeg * encoder)
{
GstVideoInfo *vinfo;
gint i;
vinfo = GST_VAAPI_ENCODER_VIDEO_INFO (encoder);
if (GST_VIDEO_INFO_FORMAT (vinfo) == GST_VIDEO_FORMAT_ENCODED) {
/* Use native I420 format */
encoder->n_components = 3;
for (i = 0; i < encoder->n_components; ++i) {
if (i == 0)
encoder->h_samp[i] = encoder->v_samp[i] = 2;
else
encoder->h_samp[i] = encoder->v_samp[i] = 1;
GST_DEBUG ("sampling factors: %d %d", encoder->h_samp[i],
encoder->v_samp[i]);
}
return;
}
encoder->n_components = GST_VIDEO_INFO_N_COMPONENTS (vinfo);
encoder->h_max_samp = 0;
encoder->v_max_samp = 0;
for (i = 0; i < encoder->n_components; ++i) {
encoder->cwidth[i] = GST_VIDEO_INFO_COMP_WIDTH (vinfo, i);
encoder->cheight[i] = GST_VIDEO_INFO_COMP_HEIGHT (vinfo, i);
encoder->h_samp[i] =
GST_ROUND_UP_4 (GST_VIDEO_INFO_WIDTH (vinfo)) / encoder->cwidth[i];
encoder->h_max_samp = MAX (encoder->h_max_samp, encoder->h_samp[i]);
encoder->v_samp[i] =
GST_ROUND_UP_4 (GST_VIDEO_INFO_HEIGHT (vinfo)) / encoder->cheight[i];
encoder->v_max_samp = MAX (encoder->v_max_samp, encoder->v_samp[i]);
}
/* samp should only be 1, 2 or 4 */
g_assert (encoder->h_max_samp <= 4);
g_assert (encoder->v_max_samp <= 4);
/* now invert */
/* maximum is invariant, as one of the components should have samp 1 */
for (i = 0; i < encoder->n_components; ++i) {
encoder->h_samp[i] = encoder->h_max_samp / encoder->h_samp[i];
encoder->v_samp[i] = encoder->v_max_samp / encoder->v_samp[i];
GST_DEBUG ("sampling factors: %d %d", encoder->h_samp[i],
encoder->v_samp[i]);
}
}
/* Derives the profile that suits best to the configuration */
static GstVaapiEncoderStatus
ensure_profile (GstVaapiEncoderJpeg * encoder)
{
/* Always start from "simple" profile for maximum compatibility */
encoder->profile = GST_VAAPI_PROFILE_JPEG_BASELINE;
return GST_VAAPI_ENCODER_STATUS_SUCCESS;
}
/* Derives the profile supported by the underlying hardware */
static gboolean
ensure_hw_profile (GstVaapiEncoderJpeg * encoder)
{
GstVaapiDisplay *const display = GST_VAAPI_ENCODER_DISPLAY (encoder);
GstVaapiEntrypoint entrypoint = GST_VAAPI_ENTRYPOINT_PICTURE_ENCODE;
GstVaapiProfile profile, profiles[2];
guint i, num_profiles = 0;
profiles[num_profiles++] = encoder->profile;
profile = GST_VAAPI_PROFILE_UNKNOWN;
for (i = 0; i < num_profiles; i++) {
if (gst_vaapi_display_has_encoder (display, profiles[i], entrypoint)) {
profile = profiles[i];
break;
}
}
if (profile == GST_VAAPI_PROFILE_UNKNOWN)
goto error_unsupported_profile;
GST_VAAPI_ENCODER_CAST (encoder)->profile = profile;
return TRUE;
/* ERRORS */
error_unsupported_profile:
{
GST_ERROR ("unsupported HW profile %s",
gst_vaapi_profile_get_va_name (encoder->profile));
return FALSE;
}
}
static GstVaapiEncoderStatus
set_context_info (GstVaapiEncoder * base_encoder)
{
GstVaapiEncoderJpeg *encoder = GST_VAAPI_ENCODER_JPEG (base_encoder);
GstVideoInfo *const vip = GST_VAAPI_ENCODER_VIDEO_INFO (encoder);
/* Maximum sizes for common headers (in bytes) */
enum
{
MAX_APP_HDR_SIZE = 20,
MAX_FRAME_HDR_SIZE = 19,
MAX_QUANT_TABLE_SIZE = 138,
MAX_HUFFMAN_TABLE_SIZE = 432,
MAX_SCAN_HDR_SIZE = 14
};
if (!ensure_hw_profile (encoder))
return GST_VAAPI_ENCODER_STATUS_ERROR_UNSUPPORTED_PROFILE;
base_encoder->num_ref_frames = 0;
/* Only YUV 4:2:0 formats are supported for now. */
base_encoder->codedbuf_size = GST_ROUND_UP_16 (vip->width) *
GST_ROUND_UP_16 (vip->height) * 3 / 2;
base_encoder->codedbuf_size += MAX_APP_HDR_SIZE + MAX_FRAME_HDR_SIZE +
MAX_QUANT_TABLE_SIZE + MAX_HUFFMAN_TABLE_SIZE + MAX_SCAN_HDR_SIZE;
base_encoder->context_info.profile = base_encoder->profile;
base_encoder->context_info.entrypoint = GST_VAAPI_ENTRYPOINT_PICTURE_ENCODE;
return GST_VAAPI_ENCODER_STATUS_SUCCESS;
}
static gboolean
fill_picture (GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture,
GstVaapiCodedBuffer * codedbuf, GstVaapiSurfaceProxy * surface)
{
guint i;
VAEncPictureParameterBufferJPEG *const pic_param = picture->param;
memset (pic_param, 0, sizeof (VAEncPictureParameterBufferJPEG));
pic_param->reconstructed_picture =
GST_VAAPI_SURFACE_PROXY_SURFACE_ID (surface);
pic_param->picture_width = GST_VAAPI_ENCODER_WIDTH (encoder);
pic_param->picture_height = GST_VAAPI_ENCODER_HEIGHT (encoder);
pic_param->coded_buf = GST_VAAPI_CODED_BUFFER_ID (codedbuf);
pic_param->pic_flags.bits.profile = 0; /* Profile = Baseline */
pic_param->pic_flags.bits.progressive = 0; /* Sequential encoding */
pic_param->pic_flags.bits.huffman = 1; /* Uses Huffman coding */
pic_param->pic_flags.bits.interleaved = 0; /* Input format is non interleaved (YUV) */
pic_param->pic_flags.bits.differential = 0; /* non-Differential Encoding */
pic_param->sample_bit_depth = 8;
pic_param->num_scan = 1;
pic_param->num_components = encoder->n_components;
pic_param->quality = encoder->quality;
for (i = 0; i < pic_param->num_components; i++) {
pic_param->component_id[i] = i + 1;
if (i != 0)
pic_param->quantiser_table_selector[i] = 1;
}
return TRUE;
}
static gboolean
ensure_picture (GstVaapiEncoderJpeg * encoder, GstVaapiEncPicture * picture,
GstVaapiCodedBufferProxy * codedbuf_proxy, GstVaapiSurfaceProxy * surface)
{
GstVaapiCodedBuffer *const codedbuf =
GST_VAAPI_CODED_BUFFER_PROXY_BUFFER (codedbuf_proxy);
if (!fill_picture (encoder, picture, codedbuf, surface))
return FALSE;
return TRUE;
}
/* This is a work-around: Normalize the quality factor and scale QM
* values similar to what VA-Intel driver is doing. Otherwise the
* generated packed headers will be wrong, since the driver itself
* is scaling the QM values using the normalized quality factor */
static void
generate_scaled_qm (GstJpegQuantTables * quant_tables,
GstJpegQuantTables * scaled_quant_tables, guint quality)
{
guint qt_val, nm_quality, i;
nm_quality = quality == 0 ? 1 : quality;
nm_quality =
(nm_quality < 50) ? (5000 / nm_quality) : (200 - (nm_quality * 2));
g_assert (quant_tables != NULL);
g_assert (scaled_quant_tables != NULL);
for (i = 0; i < GST_JPEG_MAX_QUANT_ELEMENTS; i++) {
/* Luma QM */
qt_val = (quant_tables->quant_tables[0].quant_table[i] * nm_quality) / 100;
scaled_quant_tables->quant_tables[0].quant_table[i] =
CLAMP (qt_val, 1, 255);
/* Chroma QM */
qt_val = (quant_tables->quant_tables[1].quant_table[i] * nm_quality) / 100;
scaled_quant_tables->quant_tables[1].quant_table[i] =
CLAMP (qt_val, 1, 255);
}
}
static gboolean
fill_quantization_table (GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture)
{
VAQMatrixBufferJPEG *q_matrix;
int i;
g_assert (picture);
picture->q_matrix = GST_VAAPI_ENC_Q_MATRIX_NEW (JPEG, encoder);
if (!picture->q_matrix) {
GST_ERROR ("failed to allocate quantiser table");
return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED;
}
q_matrix = picture->q_matrix->param;
if (!encoder->has_quant_tables) {
gst_jpeg_get_default_quantization_tables (&encoder->quant_tables);
encoder->has_quant_tables = TRUE;
generate_scaled_qm (&encoder->quant_tables, &encoder->scaled_quant_tables,
encoder->quality);
}
q_matrix->load_lum_quantiser_matrix = 1;
for (i = 0; i < GST_JPEG_MAX_QUANT_ELEMENTS; i++) {
q_matrix->lum_quantiser_matrix[i] =
encoder->quant_tables.quant_tables[0].quant_table[i];
}
q_matrix->load_chroma_quantiser_matrix = 1;
for (i = 0; i < GST_JPEG_MAX_QUANT_ELEMENTS; i++) {
q_matrix->chroma_quantiser_matrix[i] =
encoder->quant_tables.quant_tables[1].quant_table[i];
}
return TRUE;
}
static gboolean
ensure_quantization_table (GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture)
{
g_assert (picture);
if (!fill_quantization_table (encoder, picture))
return FALSE;
return TRUE;
}
static gboolean
fill_huffman_table (GstVaapiEncoderJpeg * encoder, GstVaapiEncPicture * picture)
{
VAHuffmanTableBufferJPEGBaseline *huffman_table;
guint i, num_tables;
g_assert (picture);
picture->huf_table = GST_VAAPI_ENC_HUFFMAN_TABLE_NEW (JPEGBaseline, encoder);
if (!picture->huf_table) {
GST_ERROR ("failed to allocate Huffman tables");
return GST_VAAPI_DECODER_STATUS_ERROR_ALLOCATION_FAILED;
}
huffman_table = picture->huf_table->param;
num_tables = MIN (G_N_ELEMENTS (huffman_table->huffman_table),
GST_JPEG_MAX_SCAN_COMPONENTS);
if (!encoder->has_huff_tables) {
gst_jpeg_get_default_huffman_tables (&encoder->huff_tables);
encoder->has_huff_tables = TRUE;
}
for (i = 0; i < num_tables; i++) {
huffman_table->load_huffman_table[i] =
encoder->huff_tables.dc_tables[i].valid
&& encoder->huff_tables.ac_tables[i].valid;
if (!huffman_table->load_huffman_table[i])
continue;
memcpy (huffman_table->huffman_table[i].num_dc_codes,
encoder->huff_tables.dc_tables[i].huf_bits,
sizeof (huffman_table->huffman_table[i].num_dc_codes));
memcpy (huffman_table->huffman_table[i].dc_values,
encoder->huff_tables.dc_tables[i].huf_values,
sizeof (huffman_table->huffman_table[i].dc_values));
memcpy (huffman_table->huffman_table[i].num_ac_codes,
encoder->huff_tables.ac_tables[i].huf_bits,
sizeof (huffman_table->huffman_table[i].num_ac_codes));
memcpy (huffman_table->huffman_table[i].ac_values,
encoder->huff_tables.ac_tables[i].huf_values,
sizeof (huffman_table->huffman_table[i].ac_values));
memset (huffman_table->huffman_table[i].pad,
0, sizeof (huffman_table->huffman_table[i].pad));
}
return TRUE;
}
static gboolean
ensure_huffman_table (GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture)
{
g_assert (picture);
if (!fill_huffman_table (encoder, picture))
return FALSE;
return TRUE;
}
static gboolean
fill_slices (GstVaapiEncoderJpeg * encoder, GstVaapiEncPicture * picture)
{
VAEncSliceParameterBufferJPEG *slice_param;
GstVaapiEncSlice *slice;
VAEncPictureParameterBufferJPEG *const pic_param = picture->param;
slice = GST_VAAPI_ENC_SLICE_NEW (JPEG, encoder);
g_assert (slice && slice->param_id != VA_INVALID_ID);
slice_param = slice->param;
memset (slice_param, 0, sizeof (VAEncSliceParameterBufferJPEG));
slice_param->restart_interval = 0;
slice_param->num_components = pic_param->num_components;
slice_param->components[0].component_selector = 1;
slice_param->components[0].dc_table_selector = 0;
slice_param->components[0].ac_table_selector = 0;
slice_param->components[1].component_selector = 2;
slice_param->components[1].dc_table_selector = 1;
slice_param->components[1].ac_table_selector = 1;
slice_param->components[2].component_selector = 3;
slice_param->components[2].dc_table_selector = 1;
slice_param->components[2].ac_table_selector = 1;
gst_vaapi_enc_picture_add_slice (picture, slice);
gst_vaapi_codec_object_replace (&slice, NULL);
return TRUE;
}
static gboolean
ensure_slices (GstVaapiEncoderJpeg * encoder, GstVaapiEncPicture * picture)
{
g_assert (picture);
if (!fill_slices (encoder, picture))
return FALSE;
return TRUE;
}
static void
generate_frame_hdr (GstJpegFrameHdr * frame_hdr, GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture)
{
VAEncPictureParameterBufferJPEG *const pic_param = picture->param;
guint i;
memset (frame_hdr, 0, sizeof (GstJpegFrameHdr));
frame_hdr->sample_precision = 8;
frame_hdr->width = pic_param->picture_width;
frame_hdr->height = pic_param->picture_height;
frame_hdr->num_components = pic_param->num_components;
for (i = 0; i < frame_hdr->num_components; i++) {
frame_hdr->components[i].identifier = pic_param->component_id[i];
frame_hdr->components[i].horizontal_factor = encoder->h_samp[i];
frame_hdr->components[i].vertical_factor = encoder->v_samp[i];
frame_hdr->components[i].quant_table_selector =
pic_param->quantiser_table_selector[i];
}
}
static void
generate_scan_hdr (GstJpegScanHdr * scan_hdr, GstVaapiEncPicture * picture)
{
VAEncPictureParameterBufferJPEG *const pic_param = picture->param;
memset (scan_hdr, 0, sizeof (GstJpegScanHdr));
scan_hdr->num_components = pic_param->num_components;
//Y Component
scan_hdr->components[0].component_selector = 1;
scan_hdr->components[0].dc_selector = 0;
scan_hdr->components[0].ac_selector = 0;
//U Component
scan_hdr->components[1].component_selector = 2;
scan_hdr->components[1].dc_selector = 1;
scan_hdr->components[1].ac_selector = 1;
//V Component
scan_hdr->components[2].component_selector = 3;
scan_hdr->components[2].dc_selector = 1;
scan_hdr->components[2].ac_selector = 1;
}
static gboolean
bs_write_jpeg_header (GstBitWriter * bs, GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture)
{
GstJpegFrameHdr frame_hdr;
GstJpegScanHdr scan_hdr;
guint i, j;
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_SOI, 8);
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_APP_MIN, 8);
gst_bit_writer_put_bits_uint16 (bs, 16, 16);
gst_bit_writer_put_bits_uint8 (bs, 0x4A, 8); //J
gst_bit_writer_put_bits_uint8 (bs, 0x46, 8); //F
gst_bit_writer_put_bits_uint8 (bs, 0x49, 8); //I
gst_bit_writer_put_bits_uint8 (bs, 0x46, 8); //F
gst_bit_writer_put_bits_uint8 (bs, 0x00, 8); //0
gst_bit_writer_put_bits_uint8 (bs, 1, 8); //Major Version
gst_bit_writer_put_bits_uint8 (bs, 1, 8); //Minor Version
gst_bit_writer_put_bits_uint8 (bs, 0, 8); //Density units 0:no units, 1:pixels per inch, 2: pixels per cm
gst_bit_writer_put_bits_uint16 (bs, 1, 16); //X density (pixel-aspect-ratio)
gst_bit_writer_put_bits_uint16 (bs, 1, 16); //Y density (pixel-aspect-ratio)
gst_bit_writer_put_bits_uint8 (bs, 0, 8); //Thumbnail width
gst_bit_writer_put_bits_uint8 (bs, 0, 8); //Thumbnail height
/* Add quantization table */
if (!encoder->has_quant_tables) {
gst_jpeg_get_default_quantization_tables (&encoder->quant_tables);
generate_scaled_qm (&encoder->quant_tables, &encoder->scaled_quant_tables,
encoder->quality);
encoder->has_quant_tables = TRUE;
}
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_DQT, 8);
gst_bit_writer_put_bits_uint16 (bs, 3 + GST_JPEG_MAX_QUANT_ELEMENTS, 16); //Lq
gst_bit_writer_put_bits_uint8 (bs, encoder->quant_tables.quant_tables[0].quant_precision, 4); //Pq
gst_bit_writer_put_bits_uint8 (bs, 0, 4); //Tq
for (i = 0; i < GST_JPEG_MAX_QUANT_ELEMENTS; i++) {
gst_bit_writer_put_bits_uint16 (bs,
encoder->scaled_quant_tables.quant_tables[0].quant_table[i], 8);
}
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_DQT, 8);
gst_bit_writer_put_bits_uint16 (bs, 3 + GST_JPEG_MAX_QUANT_ELEMENTS, 16); //Lq
gst_bit_writer_put_bits_uint8 (bs, encoder->quant_tables.quant_tables[1].quant_precision, 4); //Pq
gst_bit_writer_put_bits_uint8 (bs, 1, 4); //Tq
for (i = 0; i < GST_JPEG_MAX_QUANT_ELEMENTS; i++) {
gst_bit_writer_put_bits_uint16 (bs,
encoder->scaled_quant_tables.quant_tables[1].quant_table[i], 8);
}
/*Add frame header */
generate_frame_hdr (&frame_hdr, encoder, picture);
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_SOF_MIN, 8);
gst_bit_writer_put_bits_uint16 (bs, 8 + (3 * 3), 16); //lf, Size of FrameHeader in bytes without the Marker SOF
gst_bit_writer_put_bits_uint8 (bs, frame_hdr.sample_precision, 8);
gst_bit_writer_put_bits_uint16 (bs, frame_hdr.height, 16);
gst_bit_writer_put_bits_uint16 (bs, frame_hdr.width, 16);
gst_bit_writer_put_bits_uint8 (bs, frame_hdr.num_components, 8);
for (i = 0; i < frame_hdr.num_components; i++) {
gst_bit_writer_put_bits_uint8 (bs, frame_hdr.components[i].identifier, 8);
gst_bit_writer_put_bits_uint8 (bs,
frame_hdr.components[i].horizontal_factor, 4);
gst_bit_writer_put_bits_uint8 (bs, frame_hdr.components[i].vertical_factor,
4);
gst_bit_writer_put_bits_uint8 (bs,
frame_hdr.components[i].quant_table_selector, 8);
}
/* Add Huffman table */
if (!encoder->has_huff_tables) {
gst_jpeg_get_default_huffman_tables (&encoder->huff_tables);
encoder->has_huff_tables = TRUE;
}
for (i = 0; i < 2; i++) {
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_DHT, 8);
gst_bit_writer_put_bits_uint16 (bs, 0x1F, 16); //length of table
gst_bit_writer_put_bits_uint8 (bs, 0, 4);
gst_bit_writer_put_bits_uint8 (bs, i, 4);
for (j = 0; j < NUM_DC_RUN_SIZE_BITS; j++) {
gst_bit_writer_put_bits_uint8 (bs,
encoder->huff_tables.dc_tables[i].huf_bits[j], 8);
}
for (j = 0; j < NUM_DC_CODE_WORDS_HUFFVAL; j++) {
gst_bit_writer_put_bits_uint8 (bs,
encoder->huff_tables.dc_tables[i].huf_values[j], 8);
}
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_DHT, 8);
gst_bit_writer_put_bits_uint16 (bs, 0xB5, 16); //length of table
gst_bit_writer_put_bits_uint8 (bs, 1, 4);
gst_bit_writer_put_bits_uint8 (bs, i, 4);
for (j = 0; j < NUM_AC_RUN_SIZE_BITS; j++) {
gst_bit_writer_put_bits_uint8 (bs,
encoder->huff_tables.ac_tables[i].huf_bits[j], 8);
}
for (j = 0; j < NUM_AC_CODE_WORDS_HUFFVAL; j++) {
gst_bit_writer_put_bits_uint8 (bs,
encoder->huff_tables.ac_tables[i].huf_values[j], 8);
}
}
/* Add ScanHeader */
generate_scan_hdr (&scan_hdr, picture);
gst_bit_writer_put_bits_uint8 (bs, 0xFF, 8);
gst_bit_writer_put_bits_uint8 (bs, GST_JPEG_MARKER_SOS, 8);
gst_bit_writer_put_bits_uint16 (bs, 12, 16); //Length of Scan
gst_bit_writer_put_bits_uint8 (bs, scan_hdr.num_components, 8);
for (i = 0; i < scan_hdr.num_components; i++) {
gst_bit_writer_put_bits_uint8 (bs,
scan_hdr.components[i].component_selector, 8);
gst_bit_writer_put_bits_uint8 (bs, scan_hdr.components[i].dc_selector, 4);
gst_bit_writer_put_bits_uint8 (bs, scan_hdr.components[i].ac_selector, 4);
}
gst_bit_writer_put_bits_uint8 (bs, 0, 8); //0 for Baseline
gst_bit_writer_put_bits_uint8 (bs, 63, 8); //63 for Baseline
gst_bit_writer_put_bits_uint8 (bs, 0, 4); //0 for Baseline
gst_bit_writer_put_bits_uint8 (bs, 0, 4); //0 for Baseline
return TRUE;
}
static gboolean
add_packed_header (GstVaapiEncoderJpeg * encoder, GstVaapiEncPicture * picture)
{
GstVaapiEncPackedHeader *packed_raw_data_hdr;
GstBitWriter bs;
VAEncPackedHeaderParameterBuffer packed_raw_data_hdr_param = { 0 };
guint32 data_bit_size;
guint8 *data;
gst_bit_writer_init_with_size (&bs, 128, FALSE);
bs_write_jpeg_header (&bs, encoder, picture);
data_bit_size = GST_BIT_WRITER_BIT_SIZE (&bs);
data = GST_BIT_WRITER_DATA (&bs);
packed_raw_data_hdr_param.type = VAEncPackedHeaderRawData;
packed_raw_data_hdr_param.bit_length = data_bit_size;
packed_raw_data_hdr_param.has_emulation_bytes = 0;
packed_raw_data_hdr =
gst_vaapi_enc_packed_header_new (GST_VAAPI_ENCODER (encoder),
&packed_raw_data_hdr_param, sizeof (packed_raw_data_hdr_param), data,
(data_bit_size + 7) / 8);
g_assert (packed_raw_data_hdr);
gst_vaapi_enc_picture_add_packed_header (picture, packed_raw_data_hdr);
gst_vaapi_codec_object_replace (&packed_raw_data_hdr, NULL);
gst_bit_writer_reset (&bs);
return TRUE;
}
static gboolean
ensure_packed_headers (GstVaapiEncoderJpeg * encoder,
GstVaapiEncPicture * picture)
{
g_assert (picture);
if ((GST_VAAPI_ENCODER_PACKED_HEADERS (encoder) &
VA_ENC_PACKED_HEADER_RAW_DATA)
&& !add_packed_header (encoder, picture))
goto error_create_packed_hdr;
return TRUE;
/* ERRORS */
error_create_packed_hdr:
{
GST_ERROR ("failed to create packed raw data header buffer");
return FALSE;
}
}
static GstVaapiEncoderStatus
gst_vaapi_encoder_jpeg_encode (GstVaapiEncoder * base_encoder,
GstVaapiEncPicture * picture, GstVaapiCodedBufferProxy * codedbuf)
{
GstVaapiEncoderJpeg *const encoder = GST_VAAPI_ENCODER_JPEG (base_encoder);
GstVaapiEncoderStatus ret = GST_VAAPI_ENCODER_STATUS_ERROR_UNKNOWN;
GstVaapiSurfaceProxy *reconstruct = NULL;
reconstruct = gst_vaapi_encoder_create_surface (base_encoder);
g_assert (GST_VAAPI_SURFACE_PROXY_SURFACE (reconstruct));
if (!ensure_picture (encoder, picture, codedbuf, reconstruct))
goto error;
if (!ensure_quantization_table (encoder, picture))
goto error;
if (!ensure_huffman_table (encoder, picture))
goto error;
if (!ensure_slices (encoder, picture))
goto error;
if (!ensure_packed_headers (encoder, picture))
goto error;
if (!gst_vaapi_enc_picture_encode (picture))
goto error;
if (reconstruct)
gst_vaapi_encoder_release_surface (GST_VAAPI_ENCODER (encoder),
reconstruct);
return GST_VAAPI_ENCODER_STATUS_SUCCESS;
/* ERRORS */
error:
{
if (reconstruct)
gst_vaapi_encoder_release_surface (GST_VAAPI_ENCODER (encoder),
reconstruct);
return ret;
}
}
static GstVaapiEncoderStatus
gst_vaapi_encoder_jpeg_flush (GstVaapiEncoder * base_encoder)
{
return GST_VAAPI_ENCODER_STATUS_SUCCESS;
}
static GstVaapiEncoderStatus
gst_vaapi_encoder_jpeg_reordering (GstVaapiEncoder * base_encoder,
GstVideoCodecFrame * frame, GstVaapiEncPicture ** output)
{
GstVaapiEncoderJpeg *const encoder = GST_VAAPI_ENCODER_JPEG (base_encoder);
GstVaapiEncPicture *picture = NULL;
GstVaapiEncoderStatus status = GST_VAAPI_ENCODER_STATUS_SUCCESS;
if (!frame)
return GST_VAAPI_ENCODER_STATUS_NO_SURFACE;
picture = GST_VAAPI_ENC_PICTURE_NEW (JPEG, encoder, frame);
if (!picture) {
GST_WARNING ("create JPEG picture failed, frame timestamp:%"
GST_TIME_FORMAT, GST_TIME_ARGS (frame->pts));
return GST_VAAPI_ENCODER_STATUS_ERROR_ALLOCATION_FAILED;
}
*output = picture;
return status;
}
static GstVaapiEncoderStatus
gst_vaapi_encoder_jpeg_reconfigure (GstVaapiEncoder * base_encoder)
{
GstVaapiEncoderJpeg *const encoder = GST_VAAPI_ENCODER_JPEG (base_encoder);
GstVaapiEncoderStatus status;
status = ensure_profile (encoder);
if (status != GST_VAAPI_ENCODER_STATUS_SUCCESS)
return status;
/* generate sampling factors (A.1.1) */
generate_sampling_factors (encoder);
return set_context_info (base_encoder);
}
struct _GstVaapiEncoderJpegClass
{
GstVaapiEncoderClass parent_class;
};
G_DEFINE_TYPE (GstVaapiEncoderJpeg, gst_vaapi_encoder_jpeg,
GST_TYPE_VAAPI_ENCODER);
static void
gst_vaapi_encoder_jpeg_init (GstVaapiEncoderJpeg * encoder)
{
encoder->has_quant_tables = FALSE;
memset (&encoder->quant_tables, 0, sizeof (encoder->quant_tables));
memset (&encoder->scaled_quant_tables, 0,
sizeof (encoder->scaled_quant_tables));
encoder->has_huff_tables = FALSE;
memset (&encoder->huff_tables, 0, sizeof (encoder->huff_tables));
}
/**
* @ENCODER_JPEG_PROP_RATECONTROL: Rate control (#GstVaapiRateControl).
* @ENCODER_JPEG_PROP_TUNE: The tuning options (#GstVaapiEncoderTune).
* @ENCODER_JPEG_PROP_QUALITY: Quality Factor value (uint).
*
* The set of JPEG encoder specific configurable properties.
*/
enum
{
ENCODER_JPEG_PROP_RATECONTROL = 1,
ENCODER_JPEG_PROP_TUNE,
ENCODER_JPEG_PROP_QUALITY,
ENCODER_JPEG_N_PROPERTIES
};
static GParamSpec *properties[ENCODER_JPEG_N_PROPERTIES];
static void
gst_vaapi_encoder_jpeg_set_property (GObject * object, guint prop_id,
const GValue * value, GParamSpec * pspec)
{
GstVaapiEncoder *const base_encoder = GST_VAAPI_ENCODER (object);
GstVaapiEncoderJpeg *const encoder = GST_VAAPI_ENCODER_JPEG (object);
if (base_encoder->num_codedbuf_queued > 0) {
GST_ERROR_OBJECT (object,
"failed to set any property after encoding started");
return;
}
switch (prop_id) {
case ENCODER_JPEG_PROP_RATECONTROL:
gst_vaapi_encoder_set_rate_control (base_encoder,
g_value_get_enum (value));
break;
case ENCODER_JPEG_PROP_TUNE:
gst_vaapi_encoder_set_tuning (base_encoder, g_value_get_enum (value));
break;
case ENCODER_JPEG_PROP_QUALITY:
encoder->quality = g_value_get_uint (value);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
}
}
static void
gst_vaapi_encoder_jpeg_get_property (GObject * object, guint prop_id,
GValue * value, GParamSpec * pspec)
{
GstVaapiEncoderJpeg *const encoder = GST_VAAPI_ENCODER_JPEG (object);
GstVaapiEncoder *const base_encoder = GST_VAAPI_ENCODER (object);
switch (prop_id) {
case ENCODER_JPEG_PROP_RATECONTROL:
g_value_set_enum (value, base_encoder->rate_control);
break;
case ENCODER_JPEG_PROP_TUNE:
g_value_set_enum (value, base_encoder->tune);
break;
case ENCODER_JPEG_PROP_QUALITY:
g_value_set_uint (value, encoder->quality);
break;
default:
G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
}
}
GST_VAAPI_ENCODER_DEFINE_CLASS_DATA (JPEG);
static void
gst_vaapi_encoder_jpeg_class_init (GstVaapiEncoderJpegClass * klass)
{
GObjectClass *const object_class = G_OBJECT_CLASS (klass);
GstVaapiEncoderClass *const encoder_class = GST_VAAPI_ENCODER_CLASS (klass);
encoder_class->class_data = &g_class_data;
encoder_class->reconfigure = gst_vaapi_encoder_jpeg_reconfigure;
encoder_class->reordering = gst_vaapi_encoder_jpeg_reordering;
encoder_class->encode = gst_vaapi_encoder_jpeg_encode;
encoder_class->flush = gst_vaapi_encoder_jpeg_flush;
object_class->set_property = gst_vaapi_encoder_jpeg_set_property;
object_class->get_property = gst_vaapi_encoder_jpeg_get_property;
properties[ENCODER_JPEG_PROP_RATECONTROL] =
g_param_spec_enum ("rate-control",
"Rate Control", "Rate control mode",
g_class_data.rate_control_get_type (),
g_class_data.default_rate_control,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT |
GST_VAAPI_PARAM_ENCODER_EXPOSURE);
properties[ENCODER_JPEG_PROP_TUNE] =
g_param_spec_enum ("tune",
"Encoder Tuning",
"Encoder tuning option",
g_class_data.encoder_tune_get_type (),
g_class_data.default_encoder_tune,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT |
GST_VAAPI_PARAM_ENCODER_EXPOSURE);
properties[ENCODER_JPEG_PROP_QUALITY] =
g_param_spec_uint ("quality",
"Quality factor",
"Quality factor", 0, 100, 50,
G_PARAM_READWRITE | G_PARAM_STATIC_STRINGS | G_PARAM_CONSTRUCT |
GST_VAAPI_PARAM_ENCODER_EXPOSURE);
g_object_class_install_properties (object_class, ENCODER_JPEG_N_PROPERTIES,
properties);
}
/**
* gst_vaapi_encoder_jpeg_new:
* @display: a #GstVaapiDisplay
*
* Creates a new #GstVaapiEncoder for JPEG encoding.
*
* Return value: the newly allocated #GstVaapiEncoder object
*/
GstVaapiEncoder *
gst_vaapi_encoder_jpeg_new (GstVaapiDisplay * display)
{
return g_object_new (GST_TYPE_VAAPI_ENCODER_JPEG, "display", display, NULL);
}