mirror of
https://gitlab.freedesktop.org/gstreamer/gstreamer.git
synced 2024-10-22 10:23:47 +00:00
72ab56c376
An application, using for example appsink, can hold buffers from any va allocator after setting the pipeline to NULL. We need to destroy the allocator when that memory is unrefed. This patch juggles a bit with the allocator reference count in memories in order to achieve this: 1. When memory is created no alloc ref is modified 2. When memory is released, alloc ref is decreased 3. When memory is reassiged to a buffer, alloc ref is increased 4. When memory is flushed, alloc ref is increased becase it is going to be decreased in gst_memory_unref() Also this patch moves the deallocation of member variables to finalize() rather than dispose() Part-of: <https://gitlab.freedesktop.org/gstreamer/gst-plugins-bad/-/merge_requests/1815>
1453 lines
38 KiB
C
1453 lines
38 KiB
C
/* GStreamer
|
|
* Copyright (C) 2020 Igalia, S.L.
|
|
* Author: Víctor Jáquez <vjaquez@igalia.com>
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Library General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Library General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Library General Public
|
|
* License along with this library; if not, write to the
|
|
* Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
|
|
* Boston, MA 02110-1301, USA.
|
|
*/
|
|
|
|
#ifdef HAVE_CONFIG_H
|
|
#include "config.h"
|
|
#endif
|
|
|
|
#include "gstvaallocator.h"
|
|
|
|
#include <sys/types.h>
|
|
#include <unistd.h>
|
|
#include <va/va_drmcommon.h>
|
|
|
|
#include "gstvacaps.h"
|
|
#include "gstvavideoformat.h"
|
|
|
|
#define GST_CAT_DEFAULT gst_va_memory_debug
|
|
GST_DEBUG_CATEGORY_STATIC (gst_va_memory_debug);
|
|
|
|
static void
|
|
_init_debug_category (void)
|
|
{
|
|
#ifndef GST_DISABLE_GST_DEBUG
|
|
static volatile gsize _init = 0;
|
|
|
|
if (g_once_init_enter (&_init)) {
|
|
GST_DEBUG_CATEGORY_INIT (gst_va_memory_debug, "vamemory", 0, "VA memory");
|
|
g_once_init_leave (&_init, 1);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
static gboolean
|
|
_destroy_surfaces (GstVaDisplay * display, VASurfaceID * surfaces,
|
|
gint num_surfaces)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
g_return_val_if_fail (num_surfaces > 0, FALSE);
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaDestroySurfaces (dpy, surfaces, num_surfaces);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaDestroySurfaces: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
|
|
return TRUE;
|
|
|
|
}
|
|
|
|
static gboolean
|
|
_create_surfaces (GstVaDisplay * display, guint rt_format, guint fourcc,
|
|
guint width, guint height, gint usage_hint,
|
|
VASurfaceAttribExternalBuffers * ext_buf, VASurfaceID * surfaces,
|
|
guint num_surfaces)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
/* *INDENT-OFF* */
|
|
VASurfaceAttrib attrs[5] = {
|
|
{
|
|
.type = VASurfaceAttribUsageHint,
|
|
.flags = VA_SURFACE_ATTRIB_SETTABLE,
|
|
.value.type = VAGenericValueTypeInteger,
|
|
.value.value.i = usage_hint,
|
|
},
|
|
{
|
|
.type = VASurfaceAttribMemoryType,
|
|
.flags = VA_SURFACE_ATTRIB_SETTABLE,
|
|
.value.type = VAGenericValueTypeInteger,
|
|
.value.value.i = ext_buf ? VA_SURFACE_ATTRIB_MEM_TYPE_DRM_PRIME
|
|
: VA_SURFACE_ATTRIB_MEM_TYPE_VA,
|
|
},
|
|
};
|
|
/* *INDENT-ON* */
|
|
VAStatus status;
|
|
guint num_attrs = 2;
|
|
|
|
g_return_val_if_fail (num_surfaces > 0, FALSE);
|
|
|
|
if (fourcc > 0) {
|
|
/* *INDENT-OFF* */
|
|
attrs[num_attrs++] = (VASurfaceAttrib) {
|
|
.type = VASurfaceAttribPixelFormat,
|
|
.flags = VA_SURFACE_ATTRIB_SETTABLE,
|
|
.value.type = VAGenericValueTypeInteger,
|
|
.value.value.i = fourcc,
|
|
};
|
|
/* *INDENT-ON* */
|
|
}
|
|
|
|
if (ext_buf) {
|
|
/* *INDENT-OFF* */
|
|
attrs[num_attrs++] = (VASurfaceAttrib) {
|
|
.type = VASurfaceAttribExternalBufferDescriptor,
|
|
.flags = VA_SURFACE_ATTRIB_SETTABLE,
|
|
.value.type = VAGenericValueTypePointer,
|
|
.value.value.p = ext_buf,
|
|
};
|
|
/* *INDENT-ON* */
|
|
}
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaCreateSurfaces (dpy, rt_format, width, height, surfaces,
|
|
num_surfaces, attrs, num_attrs);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaCreateSurfaces: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_export_surface_to_dmabuf (GstVaDisplay * display, VASurfaceID surface,
|
|
guint32 flags, VADRMPRIMESurfaceDescriptor * desc)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaExportSurfaceHandle (dpy, surface,
|
|
VA_SURFACE_ATTRIB_MEM_TYPE_DRM_PRIME_2, flags, desc);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaExportSurfaceHandle: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_destroy_image (GstVaDisplay * display, VAImageID image_id)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaDestroyImage (dpy, image_id);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaDestroyImage: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_get_derive_image (GstVaDisplay * display, VASurfaceID surface, VAImage * image)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaDeriveImage (dpy, surface, image);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_WARNING ("vaDeriveImage: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_create_image (GstVaDisplay * display, GstVideoFormat format, gint width,
|
|
gint height, VAImage * image)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
const VAImageFormat *va_format;
|
|
VAStatus status;
|
|
|
|
va_format = gst_va_image_format_from_video_format (format);
|
|
if (!va_format)
|
|
return FALSE;
|
|
|
|
gst_va_display_lock (display);
|
|
status =
|
|
vaCreateImage (dpy, (VAImageFormat *) va_format, width, height, image);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaCreateImage: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_get_image (GstVaDisplay * display, VASurfaceID surface, VAImage * image)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaGetImage (dpy, surface, 0, 0, image->width, image->height,
|
|
image->image_id);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaGetImage: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_sync_surface (GstVaDisplay * display, VASurfaceID surface)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaSyncSurface (dpy, surface);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_WARNING ("vaSyncSurface: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_map_buffer (GstVaDisplay * display, VABufferID buffer, gpointer * data)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaMapBuffer (dpy, buffer, data);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_WARNING ("vaMapBuffer: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_unmap_buffer (GstVaDisplay * display, VABufferID buffer)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaUnmapBuffer (dpy, buffer);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_WARNING ("vaUnmapBuffer: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
static gboolean
|
|
_put_image (GstVaDisplay * display, VASurfaceID surface, VAImage * image)
|
|
{
|
|
VADisplay dpy = gst_va_display_get_va_dpy (display);
|
|
VAStatus status;
|
|
|
|
if (!_sync_surface (display, surface))
|
|
return FALSE;
|
|
|
|
gst_va_display_lock (display);
|
|
status = vaPutImage (dpy, surface, image->image_id, 0, 0, image->width,
|
|
image->height, 0, 0, image->width, image->height);
|
|
gst_va_display_unlock (display);
|
|
if (status != VA_STATUS_SUCCESS) {
|
|
GST_ERROR ("vaPutImage: %s", vaErrorStr (status));
|
|
return FALSE;
|
|
}
|
|
return TRUE;
|
|
}
|
|
|
|
/*=========================== Quarks for GstMemory ===========================*/
|
|
|
|
static GQuark
|
|
gst_va_buffer_surface_quark (void)
|
|
{
|
|
static gsize surface_quark = 0;
|
|
|
|
if (g_once_init_enter (&surface_quark)) {
|
|
GQuark quark = g_quark_from_string ("GstVaBufferSurface");
|
|
g_once_init_leave (&surface_quark, quark);
|
|
}
|
|
|
|
return surface_quark;
|
|
}
|
|
|
|
static GQuark
|
|
gst_va_drm_mod_quark (void)
|
|
{
|
|
static gsize drm_mod_quark = 0;
|
|
|
|
if (g_once_init_enter (&drm_mod_quark)) {
|
|
GQuark quark = g_quark_from_string ("DRMModifier");
|
|
g_once_init_leave (&drm_mod_quark, quark);
|
|
}
|
|
|
|
return drm_mod_quark;
|
|
}
|
|
|
|
/*========================= GstVaBufferSurface ===============================*/
|
|
|
|
typedef struct _GstVaBufferSurface GstVaBufferSurface;
|
|
struct _GstVaBufferSurface
|
|
{
|
|
GstVaDisplay *display;
|
|
VASurfaceID surface;
|
|
volatile gint ref_count;
|
|
};
|
|
|
|
static void
|
|
gst_va_buffer_surface_unref (gpointer data)
|
|
{
|
|
GstVaBufferSurface *buf = data;
|
|
|
|
g_return_if_fail (buf && GST_IS_VA_DISPLAY (buf->display));
|
|
|
|
if (g_atomic_int_dec_and_test (&buf->ref_count)) {
|
|
GST_LOG_OBJECT (buf->display, "Destroying surface %#x", buf->surface);
|
|
_destroy_surfaces (buf->display, &buf->surface, 1);
|
|
gst_clear_object (&buf->display);
|
|
g_slice_free (GstVaBufferSurface, buf);
|
|
}
|
|
}
|
|
|
|
static GstVaBufferSurface *
|
|
gst_va_buffer_surface_new (VASurfaceID surface, GstVideoFormat format,
|
|
gint width, gint height)
|
|
{
|
|
GstVaBufferSurface *buf = g_slice_new (GstVaBufferSurface);
|
|
|
|
g_atomic_int_set (&buf->ref_count, 0);
|
|
buf->surface = surface;
|
|
buf->display = NULL;
|
|
|
|
return buf;
|
|
}
|
|
|
|
static void
|
|
_available_mems_flush (GstVaDisplay * display, GstAtomicQueue * available_mems,
|
|
gint * surface_count)
|
|
{
|
|
GstMemory *mem;
|
|
GstVaBufferSurface *buf;
|
|
|
|
while ((mem = gst_atomic_queue_pop (available_mems))) {
|
|
/* destroy the surface */
|
|
buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem),
|
|
gst_va_buffer_surface_quark ());
|
|
if (buf) {
|
|
if (g_atomic_int_dec_and_test (&buf->ref_count)) {
|
|
GST_LOG ("Destroying surface %#x", buf->surface);
|
|
_destroy_surfaces (display, &buf->surface, 1);
|
|
*surface_count -= 1; /* GstVaDmabufAllocator */
|
|
g_slice_free (GstVaBufferSurface, buf);
|
|
}
|
|
} else {
|
|
*surface_count -= 1; /* GstVaAllocator */
|
|
}
|
|
|
|
GST_MINI_OBJECT_CAST (mem)->dispose = NULL;
|
|
/* when mem are pushed available queue its allocator is unref,
|
|
* then now it is required to ref the allocator here because
|
|
* memory's finalize will unref it again */
|
|
gst_object_ref (mem->allocator);
|
|
gst_memory_unref (mem);
|
|
}
|
|
}
|
|
|
|
/*=========================== GstVaDmabufAllocator ===========================*/
|
|
|
|
struct _GstVaDmabufAllocator
|
|
{
|
|
GstDmaBufAllocator parent;
|
|
|
|
/* queue for disposed surfaces */
|
|
GstAtomicQueue *available_mems;
|
|
gint surface_count;
|
|
|
|
GstVaDisplay *display;
|
|
|
|
GstMemoryMapFunction parent_map;
|
|
|
|
GCond buffer_cond;
|
|
|
|
GstVideoInfo info;
|
|
guint usage_hint;
|
|
};
|
|
|
|
#define gst_va_dmabuf_allocator_parent_class dmabuf_parent_class
|
|
G_DEFINE_TYPE_WITH_CODE (GstVaDmabufAllocator, gst_va_dmabuf_allocator,
|
|
GST_TYPE_DMABUF_ALLOCATOR, _init_debug_category ());
|
|
|
|
static gpointer
|
|
gst_va_dmabuf_mem_map (GstMemory * gmem, gsize maxsize, GstMapFlags flags)
|
|
{
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (gmem->allocator);
|
|
VASurfaceID surface = gst_va_memory_get_surface (gmem);
|
|
|
|
_sync_surface (self->display, surface);
|
|
|
|
/* @TODO: if mapping with flag GST_MAP_VASURFACE return the
|
|
* VA_SURFACE_ID.
|
|
* if mapping and drm_modifers are not lineal, use vaDeriveImage */
|
|
#ifndef GST_DISABLE_GST_DEBUG
|
|
{
|
|
guint64 *drm_mod;
|
|
|
|
drm_mod = gst_mini_object_get_qdata (GST_MINI_OBJECT (gmem),
|
|
gst_va_drm_mod_quark ());
|
|
GST_TRACE_OBJECT (self, "DRM modifiers: %#lx", *drm_mod);
|
|
}
|
|
#endif
|
|
|
|
return self->parent_map (gmem, maxsize, flags);
|
|
}
|
|
|
|
static void
|
|
gst_va_dmabuf_allocator_finalize (GObject * object)
|
|
{
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (object);
|
|
|
|
gst_atomic_queue_unref (self->available_mems);
|
|
gst_clear_object (&self->display);
|
|
g_cond_clear (&self->buffer_cond);
|
|
|
|
G_OBJECT_CLASS (dmabuf_parent_class)->finalize (object);
|
|
}
|
|
|
|
static void
|
|
gst_va_dmabuf_allocator_dispose (GObject * object)
|
|
{
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (object);
|
|
|
|
_available_mems_flush (self->display, self->available_mems,
|
|
&self->surface_count);
|
|
if (self->surface_count != 0)
|
|
GST_WARNING_OBJECT (self, "Surfaces leaked: %d", self->surface_count);
|
|
|
|
G_OBJECT_CLASS (dmabuf_parent_class)->dispose (object);
|
|
}
|
|
|
|
static void
|
|
gst_va_dmabuf_allocator_class_init (GstVaDmabufAllocatorClass * klass)
|
|
{
|
|
GObjectClass *object_class = G_OBJECT_CLASS (klass);
|
|
|
|
object_class->dispose = gst_va_dmabuf_allocator_dispose;
|
|
object_class->finalize = gst_va_dmabuf_allocator_finalize;
|
|
}
|
|
|
|
static void
|
|
gst_va_dmabuf_allocator_init (GstVaDmabufAllocator * self)
|
|
{
|
|
self->available_mems = gst_atomic_queue_new (2);
|
|
g_cond_init (&self->buffer_cond);
|
|
|
|
self->parent_map = GST_ALLOCATOR (self)->mem_map;
|
|
GST_ALLOCATOR (self)->mem_map = gst_va_dmabuf_mem_map;
|
|
}
|
|
|
|
GstAllocator *
|
|
gst_va_dmabuf_allocator_new (GstVaDisplay * display)
|
|
{
|
|
GstVaDmabufAllocator *self;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_DISPLAY (display), NULL);
|
|
|
|
self = g_object_new (GST_TYPE_VA_DMABUF_ALLOCATOR, NULL);
|
|
self->display = gst_object_ref (display);
|
|
gst_object_ref_sink (self);
|
|
|
|
return GST_ALLOCATOR (self);
|
|
}
|
|
|
|
static inline goffset
|
|
_get_fd_size (gint fd)
|
|
{
|
|
return lseek (fd, 0, SEEK_END);
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_dmabuf_memory_release (GstMiniObject * mini_object)
|
|
{
|
|
GstMemory *mem = GST_MEMORY_CAST (mini_object);
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (mem->allocator);
|
|
|
|
GST_OBJECT_LOCK (self);
|
|
|
|
GST_LOG ("releasing %p", mem);
|
|
gst_atomic_queue_push (self->available_mems, gst_memory_ref (mem));
|
|
g_cond_signal (&self->buffer_cond);
|
|
|
|
GST_OBJECT_UNLOCK (self);
|
|
|
|
/* Keep last in case we are holding on the last allocator ref */
|
|
gst_object_unref (mem->allocator);
|
|
|
|
/* don't call mini_object's free */
|
|
return FALSE;
|
|
}
|
|
|
|
/* Creates an exported VASurface and adds it as @buffer's memories
|
|
* qdata
|
|
*
|
|
* If @info is not NULL, a dummy (non-pooled) buffer is created to
|
|
* update offsets and strides, and it has to be unrefed immediately.
|
|
*/
|
|
static gboolean
|
|
gst_va_dmabuf_allocator_setup_buffer_full (GstAllocator * allocator,
|
|
GstBuffer * buffer, GstVideoInfo * info)
|
|
{
|
|
GstVaBufferSurface *buf;
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator);
|
|
GstVideoFormat format;
|
|
VADRMPRIMESurfaceDescriptor desc = { 0, };
|
|
VASurfaceID surface;
|
|
guint32 i, fourcc, rt_format, export_flags;
|
|
GDestroyNotify buffer_destroy = NULL;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator), FALSE);
|
|
|
|
format = GST_VIDEO_INFO_FORMAT (&self->info);
|
|
fourcc = gst_va_fourcc_from_video_format (format);
|
|
rt_format = gst_va_chroma_from_video_format (format);
|
|
if (fourcc == 0 || rt_format == 0) {
|
|
GST_ERROR_OBJECT (allocator, "Unsupported format: %s",
|
|
gst_video_format_to_string (GST_VIDEO_INFO_FORMAT (&self->info)));
|
|
return FALSE;
|
|
}
|
|
|
|
if (!_create_surfaces (self->display, rt_format, fourcc,
|
|
GST_VIDEO_INFO_WIDTH (&self->info),
|
|
GST_VIDEO_INFO_HEIGHT (&self->info), self->usage_hint, NULL,
|
|
&surface, 1))
|
|
return FALSE;
|
|
|
|
/* FIXME(victor): find a better way since this is just a hack for
|
|
* i965 driver */
|
|
if (fourcc == VA_FOURCC_YUY2 || fourcc == VA_FOURCC_UYVY) {
|
|
/* These are not representable as separate planes */
|
|
export_flags = VA_EXPORT_SURFACE_COMPOSED_LAYERS;
|
|
} else {
|
|
/* Each layer will contain exactly one plane. For example, an NV12
|
|
* surface will be exported as two layers */
|
|
export_flags = VA_EXPORT_SURFACE_SEPARATE_LAYERS;
|
|
}
|
|
|
|
export_flags |= VA_EXPORT_SURFACE_READ_WRITE;
|
|
|
|
if (!_export_surface_to_dmabuf (self->display, surface, export_flags, &desc))
|
|
goto failed;
|
|
|
|
g_assert (GST_VIDEO_INFO_N_PLANES (&self->info) == desc.num_layers);
|
|
|
|
if (fourcc != desc.fourcc) {
|
|
GST_ERROR ("Unsupported fourcc: %" GST_FOURCC_FORMAT,
|
|
GST_FOURCC_ARGS (desc.fourcc));
|
|
goto failed;
|
|
}
|
|
|
|
buf = gst_va_buffer_surface_new (surface, format, desc.width, desc.height);
|
|
if (G_UNLIKELY (info)) {
|
|
*info = self->info;
|
|
GST_VIDEO_INFO_SIZE (info) = 0;
|
|
}
|
|
|
|
for (i = 0; i < desc.num_objects; i++) {
|
|
gint fd = desc.objects[i].fd;
|
|
gsize size = desc.objects[i].size > 0 ?
|
|
desc.objects[i].size : _get_fd_size (fd);
|
|
GstMemory *mem = gst_dmabuf_allocator_alloc (allocator, fd, size);
|
|
guint64 *drm_mod = g_new (guint64, 1);
|
|
|
|
gst_buffer_append_memory (buffer, mem);
|
|
|
|
if (G_LIKELY (!info)) {
|
|
GST_MINI_OBJECT (mem)->dispose = gst_va_dmabuf_memory_release;
|
|
} else {
|
|
/* if no @info, surface will be destroyed as soon as buffer is
|
|
* destroyed (e.g. gst_va_dmabuf_allocator_try()) */
|
|
buf->display = gst_object_ref (self->display);
|
|
buffer_destroy = gst_va_buffer_surface_unref;
|
|
}
|
|
|
|
g_atomic_int_add (&buf->ref_count, 1);
|
|
gst_mini_object_set_qdata (GST_MINI_OBJECT (mem),
|
|
gst_va_buffer_surface_quark (), buf, buffer_destroy);
|
|
|
|
*drm_mod = desc.objects[i].drm_format_modifier;
|
|
gst_mini_object_set_qdata (GST_MINI_OBJECT (mem), gst_va_drm_mod_quark (),
|
|
drm_mod, g_free);
|
|
|
|
if (G_UNLIKELY (info))
|
|
GST_VIDEO_INFO_SIZE (info) += size;
|
|
}
|
|
|
|
if (G_UNLIKELY (info)) {
|
|
for (i = 0; i < desc.num_layers; i++) {
|
|
g_assert (desc.layers[i].num_planes == 1);
|
|
GST_VIDEO_INFO_PLANE_OFFSET (info, i) = desc.layers[i].offset[0];
|
|
GST_VIDEO_INFO_PLANE_STRIDE (info, i) = desc.layers[i].pitch[0];
|
|
}
|
|
} else {
|
|
g_atomic_int_inc (&self->surface_count);
|
|
}
|
|
|
|
GST_LOG_OBJECT (self, "Created surface %#x [%dx%d] size %" G_GSIZE_FORMAT,
|
|
buf->surface, GST_VIDEO_INFO_WIDTH (&self->info),
|
|
GST_VIDEO_INFO_HEIGHT (&self->info), GST_VIDEO_INFO_SIZE (&self->info));
|
|
|
|
return TRUE;
|
|
|
|
failed:
|
|
{
|
|
_destroy_surfaces (self->display, &surface, 1);
|
|
return FALSE;
|
|
}
|
|
}
|
|
|
|
gboolean
|
|
gst_va_dmabuf_allocator_setup_buffer (GstAllocator * allocator,
|
|
GstBuffer * buffer)
|
|
{
|
|
return gst_va_dmabuf_allocator_setup_buffer_full (allocator, buffer, NULL);
|
|
}
|
|
|
|
gboolean
|
|
gst_va_dmabuf_allocator_prepare_buffer (GstAllocator * allocator,
|
|
GstBuffer * buffer)
|
|
{
|
|
GstMemory *pmem, *mem[GST_VIDEO_MAX_PLANES] = { 0, };
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator);
|
|
VASurfaceID surface, psurface;
|
|
gint j, idx = 1;
|
|
|
|
GST_OBJECT_LOCK (self);
|
|
|
|
/* if available mems, use them */
|
|
if (gst_atomic_queue_length (self->available_mems) == 0)
|
|
g_cond_wait (&self->buffer_cond, GST_OBJECT_GET_LOCK (self));
|
|
|
|
mem[0] = gst_atomic_queue_pop (self->available_mems);
|
|
surface = gst_va_memory_get_surface (mem[0]);
|
|
|
|
do {
|
|
pmem = gst_atomic_queue_peek (self->available_mems);
|
|
if (!pmem)
|
|
break;
|
|
|
|
psurface = gst_va_memory_get_surface (pmem);
|
|
if (psurface != surface)
|
|
break;
|
|
|
|
mem[idx++] = gst_atomic_queue_pop (self->available_mems);
|
|
} while (TRUE);
|
|
|
|
GST_OBJECT_UNLOCK (self);
|
|
|
|
/* append them in reverse order */
|
|
for (j = idx - 1; j >= 0; j--) {
|
|
gst_object_ref (mem[j]->allocator);
|
|
gst_buffer_append_memory (buffer, mem[j]);
|
|
}
|
|
|
|
GST_TRACE_OBJECT (self, "Prepared surface %#x in buffer %p", surface, buffer);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
void
|
|
gst_va_dmabuf_allocator_flush (GstAllocator * allocator)
|
|
{
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator);
|
|
|
|
GST_OBJECT_LOCK (self);
|
|
_available_mems_flush (self->display, self->available_mems,
|
|
&self->surface_count);
|
|
g_cond_signal (&self->buffer_cond);
|
|
GST_OBJECT_UNLOCK (self);
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_dmabuf_allocator_try (GstAllocator * allocator)
|
|
{
|
|
GstBuffer *buffer;
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator);
|
|
GstVideoInfo info = self->info;
|
|
gboolean ret;
|
|
|
|
buffer = gst_buffer_new ();
|
|
ret = gst_va_dmabuf_allocator_setup_buffer_full (allocator, buffer, &info);
|
|
gst_buffer_unref (buffer);
|
|
|
|
if (ret)
|
|
self->info = info;
|
|
|
|
return ret;
|
|
}
|
|
|
|
gboolean
|
|
gst_va_dmabuf_allocator_set_format (GstAllocator * allocator,
|
|
GstVideoInfo * info, guint usage_hint)
|
|
{
|
|
GstVaDmabufAllocator *self;
|
|
gboolean ret;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator), FALSE);
|
|
g_return_val_if_fail (info, FALSE);
|
|
|
|
self = GST_VA_DMABUF_ALLOCATOR (allocator);
|
|
|
|
if (self->surface_count != 0) {
|
|
if (GST_VIDEO_INFO_FORMAT (info) == GST_VIDEO_INFO_FORMAT (&self->info)
|
|
&& GST_VIDEO_INFO_WIDTH (info) == GST_VIDEO_INFO_WIDTH (&self->info)
|
|
&& GST_VIDEO_INFO_HEIGHT (info) == GST_VIDEO_INFO_HEIGHT (&self->info)
|
|
&& usage_hint == self->usage_hint) {
|
|
*info = self->info; /* update callee info (offset & stride) */
|
|
return TRUE;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
self->usage_hint = usage_hint;
|
|
self->info = *info;
|
|
|
|
ret = gst_va_dmabuf_allocator_try (allocator);
|
|
|
|
if (ret)
|
|
*info = self->info;
|
|
|
|
return ret;
|
|
}
|
|
|
|
gboolean
|
|
gst_va_dmabuf_allocator_get_format (GstAllocator * allocator,
|
|
GstVideoInfo * info, guint * usage_hint)
|
|
{
|
|
GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator);
|
|
|
|
if (GST_VIDEO_INFO_FORMAT (&self->info) == GST_VIDEO_FORMAT_UNKNOWN)
|
|
return FALSE;
|
|
|
|
if (info)
|
|
*info = self->info;
|
|
if (usage_hint)
|
|
*usage_hint = self->usage_hint;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/* XXX: use a surface pool to control the created surfaces */
|
|
gboolean
|
|
gst_va_dmabuf_memories_setup (GstVaDisplay * display, GstVideoInfo * info,
|
|
guint n_planes, GstMemory * mem[GST_VIDEO_MAX_PLANES],
|
|
uintptr_t * fds, gsize offset[GST_VIDEO_MAX_PLANES], guint usage_hint)
|
|
{
|
|
GstVideoFormat format;
|
|
GstVaBufferSurface *buf;
|
|
/* *INDENT-OFF* */
|
|
VASurfaceAttribExternalBuffers ext_buf = {
|
|
.width = GST_VIDEO_INFO_WIDTH (info),
|
|
.height = GST_VIDEO_INFO_HEIGHT (info),
|
|
.data_size = GST_VIDEO_INFO_SIZE (info),
|
|
.num_planes = GST_VIDEO_INFO_N_PLANES (info),
|
|
.buffers = fds,
|
|
.num_buffers = GST_VIDEO_INFO_N_PLANES (info),
|
|
};
|
|
/* *INDENT-ON* */
|
|
VASurfaceID surface;
|
|
guint32 fourcc, rt_format;
|
|
guint i;
|
|
gboolean ret;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_DISPLAY (display), FALSE);
|
|
|
|
format = GST_VIDEO_INFO_FORMAT (info);
|
|
if (format == GST_VIDEO_FORMAT_UNKNOWN)
|
|
return FALSE;
|
|
|
|
rt_format = gst_va_chroma_from_video_format (format);
|
|
if (rt_format == 0)
|
|
return FALSE;
|
|
|
|
fourcc = gst_va_fourcc_from_video_format (format);
|
|
if (fourcc == 0)
|
|
return FALSE;
|
|
|
|
ext_buf.pixel_format = fourcc;
|
|
|
|
for (i = 0; i < MIN (n_planes, 4); i++) {
|
|
ext_buf.pitches[i] = GST_VIDEO_INFO_PLANE_STRIDE (info, i);
|
|
ext_buf.offsets[i] = offset[i];
|
|
}
|
|
|
|
ret = _create_surfaces (display, rt_format, ext_buf.pixel_format,
|
|
ext_buf.width, ext_buf.height, usage_hint, &ext_buf, &surface, 1);
|
|
if (!ret)
|
|
return FALSE;
|
|
|
|
GST_LOG_OBJECT (display, "Created surface %#x [%dx%d]", surface,
|
|
ext_buf.width, ext_buf.height);
|
|
|
|
buf = gst_va_buffer_surface_new (surface, rt_format, ext_buf.width,
|
|
ext_buf.height);
|
|
buf->display = gst_object_ref (display);
|
|
|
|
for (i = 0; i < n_planes; i++) {
|
|
g_atomic_int_add (&buf->ref_count, 1);
|
|
gst_mini_object_set_qdata (GST_MINI_OBJECT (mem[i]),
|
|
gst_va_buffer_surface_quark (), buf, gst_va_buffer_surface_unref);
|
|
GST_INFO_OBJECT (display, "setting surface %#x to dmabuf fd %d",
|
|
buf->surface, gst_dmabuf_memory_get_fd (mem[i]));
|
|
}
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*===================== GstVaAllocator / GstVaMemory =========================*/
|
|
|
|
struct _GstVaAllocator
|
|
{
|
|
GstAllocator parent;
|
|
|
|
/* queue for disposed surfaces */
|
|
GstAtomicQueue *available_mems;
|
|
gint surface_count;
|
|
|
|
GstVaDisplay *display;
|
|
|
|
gboolean use_derived;
|
|
GArray *surface_formats;
|
|
|
|
GstVideoFormat surface_format;
|
|
GstVideoFormat img_format;
|
|
guint32 fourcc;
|
|
guint32 rt_format;
|
|
|
|
GCond buffer_cond;
|
|
|
|
GstVideoInfo info;
|
|
guint usage_hint;
|
|
};
|
|
|
|
typedef struct _GstVaMemory GstVaMemory;
|
|
struct _GstVaMemory
|
|
{
|
|
GstMemory mem;
|
|
|
|
VASurfaceID surface;
|
|
GstVideoFormat surface_format;
|
|
VAImage image;
|
|
gpointer mapped_data;
|
|
|
|
GstMapFlags prev_mapflags;
|
|
volatile gint map_count;
|
|
|
|
gboolean is_derived;
|
|
gboolean is_dirty;
|
|
GMutex lock;
|
|
};
|
|
|
|
G_DEFINE_TYPE_WITH_CODE (GstVaAllocator, gst_va_allocator, GST_TYPE_ALLOCATOR,
|
|
_init_debug_category ());
|
|
|
|
static gboolean _va_unmap (GstVaMemory * mem);
|
|
|
|
static void
|
|
gst_va_allocator_finalize (GObject * object)
|
|
{
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (object);
|
|
|
|
gst_atomic_queue_unref (self->available_mems);
|
|
gst_clear_object (&self->display);
|
|
g_clear_pointer (&self->surface_formats, g_array_unref);
|
|
g_cond_clear (&self->buffer_cond);
|
|
|
|
G_OBJECT_CLASS (gst_va_allocator_parent_class)->finalize (object);
|
|
}
|
|
|
|
static void
|
|
gst_va_allocator_dispose (GObject * object)
|
|
{
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (object);
|
|
|
|
_available_mems_flush (self->display, self->available_mems,
|
|
&self->surface_count);
|
|
if (self->surface_count != 0)
|
|
GST_WARNING_OBJECT (self, "Surfaces leaked: %d", self->surface_count);
|
|
|
|
G_OBJECT_CLASS (gst_va_allocator_parent_class)->dispose (object);
|
|
}
|
|
|
|
static void
|
|
_va_free (GstAllocator * allocator, GstMemory * mem)
|
|
{
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (allocator);
|
|
GstVaMemory *va_mem = (GstVaMemory *) mem;
|
|
|
|
if (va_mem->mapped_data) {
|
|
g_warning (G_STRLOC ":%s: Freeing memory %p still mapped", G_STRFUNC,
|
|
va_mem);
|
|
_va_unmap (va_mem);
|
|
}
|
|
|
|
if (va_mem->surface != VA_INVALID_ID && mem->parent == NULL) {
|
|
GST_LOG_OBJECT (self, "Destroying surface %#x", va_mem->surface);
|
|
_destroy_surfaces (self->display, &va_mem->surface, 1);
|
|
}
|
|
|
|
g_mutex_clear (&va_mem->lock);
|
|
|
|
g_slice_free (GstVaMemory, va_mem);
|
|
}
|
|
|
|
static void
|
|
gst_va_allocator_class_init (GstVaAllocatorClass * klass)
|
|
{
|
|
GstAllocatorClass *allocator_class = GST_ALLOCATOR_CLASS (klass);
|
|
GObjectClass *object_class = G_OBJECT_CLASS (klass);
|
|
|
|
object_class->dispose = gst_va_allocator_dispose;
|
|
object_class->finalize = gst_va_allocator_finalize;
|
|
allocator_class->free = _va_free;
|
|
}
|
|
|
|
static inline void
|
|
_clean_mem (GstVaMemory * mem)
|
|
{
|
|
memset (&mem->image, 0, sizeof (mem->image));
|
|
mem->image.image_id = VA_INVALID_ID;
|
|
mem->image.buf = VA_INVALID_ID;
|
|
|
|
mem->is_derived = TRUE;
|
|
mem->is_dirty = FALSE;
|
|
mem->prev_mapflags = 0;
|
|
mem->mapped_data = NULL;
|
|
}
|
|
|
|
static void
|
|
_reset_mem (GstVaMemory * mem, GstAllocator * allocator, gsize size)
|
|
{
|
|
_clean_mem (mem);
|
|
g_atomic_int_set (&mem->map_count, 0);
|
|
g_mutex_init (&mem->lock);
|
|
|
|
gst_memory_init (GST_MEMORY_CAST (mem), 0, allocator, NULL, size,
|
|
0 /* align */ , 0 /* offset */ , size);
|
|
}
|
|
|
|
static inline gboolean
|
|
_ensure_image (GstVaDisplay * display, VASurfaceID surface,
|
|
GstVideoInfo * info, VAImage * image, gboolean derived)
|
|
{
|
|
gboolean ret = TRUE;
|
|
|
|
if (image->image_id != VA_INVALID_ID)
|
|
return TRUE;
|
|
|
|
if (!_sync_surface (display, surface))
|
|
return FALSE;
|
|
|
|
if (derived) {
|
|
ret = _get_derive_image (display, surface, image);
|
|
} else {
|
|
ret = _create_image (display, GST_VIDEO_INFO_FORMAT (info),
|
|
GST_VIDEO_INFO_WIDTH (info), GST_VIDEO_INFO_HEIGHT (info), image);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline gboolean
|
|
_update_image_info (GstVaAllocator * va_allocator)
|
|
{
|
|
VASurfaceID surface;
|
|
VAImage image = {.image_id = VA_INVALID_ID, };
|
|
gboolean derived;
|
|
guint i;
|
|
|
|
/* Create a test surface first */
|
|
if (!_create_surfaces (va_allocator->display, va_allocator->rt_format,
|
|
va_allocator->fourcc, GST_VIDEO_INFO_WIDTH (&va_allocator->info),
|
|
GST_VIDEO_INFO_HEIGHT (&va_allocator->info), va_allocator->usage_hint,
|
|
NULL, &surface, 1)) {
|
|
GST_ERROR_OBJECT (va_allocator, "Failed to create a test surface");
|
|
return FALSE;
|
|
}
|
|
|
|
GST_DEBUG_OBJECT (va_allocator, "Created surface %#x [%dx%d]", surface,
|
|
GST_VIDEO_INFO_WIDTH (&va_allocator->info),
|
|
GST_VIDEO_INFO_HEIGHT (&va_allocator->info));
|
|
|
|
/* Try derived first, but different formats can never derive */
|
|
if (va_allocator->surface_format == va_allocator->img_format) {
|
|
derived = TRUE;
|
|
if (_get_derive_image (va_allocator->display, surface, &image))
|
|
goto update;
|
|
}
|
|
|
|
/* Then we try to create a image. */
|
|
derived = FALSE;
|
|
if (!_create_image (va_allocator->display, va_allocator->img_format,
|
|
GST_VIDEO_INFO_WIDTH (&va_allocator->info),
|
|
GST_VIDEO_INFO_HEIGHT (&va_allocator->info), &image)) {
|
|
_destroy_surfaces (va_allocator->display, &surface, 1);
|
|
return FALSE;
|
|
}
|
|
|
|
update:
|
|
va_allocator->use_derived = derived;
|
|
|
|
for (i = 0; i < image.num_planes; i++) {
|
|
GST_VIDEO_INFO_PLANE_OFFSET (&va_allocator->info, i) = image.offsets[i];
|
|
GST_VIDEO_INFO_PLANE_STRIDE (&va_allocator->info, i) = image.pitches[i];
|
|
}
|
|
|
|
GST_VIDEO_INFO_SIZE (&va_allocator->info) = image.data_size;
|
|
|
|
_destroy_image (va_allocator->display, image.image_id);
|
|
_destroy_surfaces (va_allocator->display, &surface, 1);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
static gpointer
|
|
_va_map_unlocked (GstVaMemory * mem, GstMapFlags flags)
|
|
{
|
|
GstAllocator *allocator = GST_MEMORY_CAST (mem)->allocator;
|
|
GstVaAllocator *va_allocator;
|
|
GstVaDisplay *display;
|
|
|
|
g_return_val_if_fail (mem->surface != VA_INVALID_ID, NULL);
|
|
g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), NULL);
|
|
|
|
if (g_atomic_int_get (&mem->map_count) > 0) {
|
|
if (mem->prev_mapflags != flags || !mem->mapped_data)
|
|
return NULL;
|
|
else
|
|
goto success;
|
|
}
|
|
|
|
va_allocator = GST_VA_ALLOCATOR (allocator);
|
|
display = va_allocator->display;
|
|
|
|
if (flags & GST_MAP_WRITE) {
|
|
mem->is_dirty = TRUE;
|
|
} else { /* GST_MAP_READ only */
|
|
mem->is_dirty = FALSE;
|
|
}
|
|
|
|
if (flags & GST_MAP_VA) {
|
|
mem->mapped_data = &mem->surface;
|
|
goto success;
|
|
}
|
|
|
|
if (!_ensure_image (display, mem->surface, &va_allocator->info, &mem->image,
|
|
va_allocator->use_derived))
|
|
return NULL;
|
|
|
|
mem->is_derived = va_allocator->use_derived;
|
|
|
|
if (!mem->is_derived) {
|
|
if (!_get_image (display, mem->surface, &mem->image))
|
|
goto fail;
|
|
}
|
|
|
|
if (!_map_buffer (display, mem->image.buf, &mem->mapped_data))
|
|
goto fail;
|
|
|
|
success:
|
|
{
|
|
mem->prev_mapflags = flags;
|
|
g_atomic_int_add (&mem->map_count, 1);
|
|
return mem->mapped_data;
|
|
}
|
|
|
|
fail:
|
|
{
|
|
_destroy_image (display, mem->image.image_id);
|
|
_clean_mem (mem);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
static gpointer
|
|
_va_map (GstVaMemory * mem, gsize maxsize, GstMapFlags flags)
|
|
{
|
|
gpointer data;
|
|
|
|
g_mutex_lock (&mem->lock);
|
|
data = _va_map_unlocked (mem, flags);
|
|
g_mutex_unlock (&mem->lock);
|
|
|
|
return data;
|
|
}
|
|
|
|
static gboolean
|
|
_va_unmap_unlocked (GstVaMemory * mem)
|
|
{
|
|
GstAllocator *allocator = GST_MEMORY_CAST (mem)->allocator;
|
|
GstVaDisplay *display;
|
|
gboolean ret = TRUE;
|
|
|
|
if (!g_atomic_int_dec_and_test (&mem->map_count))
|
|
return TRUE;
|
|
|
|
if (mem->prev_mapflags & GST_MAP_VA)
|
|
goto bail;
|
|
|
|
display = GST_VA_ALLOCATOR (allocator)->display;
|
|
|
|
if (mem->image.image_id != VA_INVALID_ID) {
|
|
if (mem->is_dirty && !mem->is_derived) {
|
|
ret = _put_image (display, mem->surface, &mem->image);
|
|
mem->is_dirty = FALSE;
|
|
}
|
|
/* XXX(victor): if is derived and is dirty, create another surface
|
|
* an replace it in mem */
|
|
}
|
|
|
|
ret &= _unmap_buffer (display, mem->image.buf);
|
|
ret &= _destroy_image (display, mem->image.image_id);
|
|
|
|
bail:
|
|
_clean_mem (mem);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static gboolean
|
|
_va_unmap (GstVaMemory * mem)
|
|
{
|
|
gboolean ret;
|
|
|
|
g_mutex_lock (&mem->lock);
|
|
ret = _va_unmap_unlocked (mem);
|
|
g_mutex_unlock (&mem->lock);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static GstMemory *
|
|
_va_share (GstMemory * mem, gssize offset, gssize size)
|
|
{
|
|
GstVaMemory *vamem = (GstVaMemory *) mem;
|
|
GstVaMemory *sub;
|
|
GstMemory *parent;
|
|
GST_DEBUG ("%p: share %" G_GSSIZE_FORMAT ", %" G_GSIZE_FORMAT, mem, offset,
|
|
size);
|
|
|
|
/* find real parent */
|
|
if ((parent = vamem->mem.parent) == NULL)
|
|
parent = (GstMemory *) vamem;
|
|
|
|
if (size == -1)
|
|
size = mem->maxsize - offset;
|
|
|
|
sub = g_slice_new (GstVaMemory);
|
|
/* the shared memory is alwyas readonly */
|
|
gst_memory_init (GST_MEMORY_CAST (sub), GST_MINI_OBJECT_FLAGS (parent) |
|
|
GST_MINI_OBJECT_FLAG_LOCK_READONLY, vamem->mem.allocator, parent,
|
|
vamem->mem.maxsize, vamem->mem.align, vamem->mem.offset + offset, size);
|
|
|
|
sub->surface = vamem->surface;
|
|
sub->surface_format = vamem->surface_format;
|
|
|
|
_clean_mem (sub);
|
|
|
|
g_atomic_int_set (&sub->map_count, 0);
|
|
g_mutex_init (&sub->lock);
|
|
|
|
return GST_MEMORY_CAST (sub);
|
|
}
|
|
|
|
static void
|
|
gst_va_allocator_init (GstVaAllocator * self)
|
|
{
|
|
GstAllocator *allocator = GST_ALLOCATOR (self);
|
|
|
|
self->available_mems = gst_atomic_queue_new (2);
|
|
|
|
allocator->mem_type = GST_ALLOCATOR_VASURFACE;
|
|
allocator->mem_map = (GstMemoryMapFunction) _va_map;
|
|
allocator->mem_unmap = (GstMemoryUnmapFunction) _va_unmap;
|
|
allocator->mem_share = _va_share;
|
|
|
|
self->use_derived = TRUE;
|
|
|
|
g_cond_init (&self->buffer_cond);
|
|
|
|
GST_OBJECT_FLAG_SET (self, GST_ALLOCATOR_FLAG_CUSTOM_ALLOC);
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_memory_release (GstMiniObject * mini_object)
|
|
{
|
|
GstMemory *mem = GST_MEMORY_CAST (mini_object);
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (mem->allocator);
|
|
|
|
GST_OBJECT_LOCK (self);
|
|
|
|
GST_LOG ("releasing %p", mem);
|
|
gst_atomic_queue_push (self->available_mems, gst_memory_ref (mem));
|
|
g_cond_signal (&self->buffer_cond);
|
|
|
|
GST_OBJECT_UNLOCK (self);
|
|
|
|
/* Keep last in case we are holding on the last allocator ref */
|
|
gst_object_unref (mem->allocator);
|
|
|
|
/* don't call mini_object's free */
|
|
return FALSE;
|
|
}
|
|
|
|
GstMemory *
|
|
gst_va_allocator_alloc (GstAllocator * allocator)
|
|
{
|
|
GstVaAllocator *self;
|
|
GstVaMemory *mem;
|
|
VASurfaceID surface;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), NULL);
|
|
|
|
self = GST_VA_ALLOCATOR (allocator);
|
|
|
|
g_return_val_if_fail (self->rt_format != 0, NULL);
|
|
|
|
if (!_create_surfaces (self->display, self->rt_format, self->fourcc,
|
|
GST_VIDEO_INFO_WIDTH (&self->info),
|
|
GST_VIDEO_INFO_HEIGHT (&self->info), self->usage_hint, NULL,
|
|
&surface, 1))
|
|
return NULL;
|
|
|
|
mem = g_slice_new (GstVaMemory);
|
|
|
|
mem->surface = surface;
|
|
mem->surface_format = self->surface_format;
|
|
|
|
_reset_mem (mem, allocator, GST_VIDEO_INFO_SIZE (&self->info));
|
|
|
|
GST_MINI_OBJECT (mem)->dispose = gst_va_memory_release;
|
|
g_atomic_int_inc (&self->surface_count);
|
|
|
|
GST_LOG_OBJECT (self, "Created surface %#x [%dx%d]", mem->surface,
|
|
GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info));
|
|
|
|
return GST_MEMORY_CAST (mem);
|
|
}
|
|
|
|
GstAllocator *
|
|
gst_va_allocator_new (GstVaDisplay * display, GArray * surface_formats)
|
|
{
|
|
GstVaAllocator *self;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_DISPLAY (display), NULL);
|
|
|
|
self = g_object_new (GST_TYPE_VA_ALLOCATOR, NULL);
|
|
self->display = gst_object_ref (display);
|
|
self->surface_formats = surface_formats;
|
|
gst_object_ref_sink (self);
|
|
|
|
return GST_ALLOCATOR (self);
|
|
}
|
|
|
|
gboolean
|
|
gst_va_allocator_prepare_buffer (GstAllocator * allocator, GstBuffer * buffer)
|
|
{
|
|
GstMemory *mem;
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (allocator);
|
|
VASurfaceID surface;
|
|
|
|
GST_OBJECT_LOCK (self);
|
|
/* if available mems, use them */
|
|
if (gst_atomic_queue_length (self->available_mems) == 0)
|
|
g_cond_wait (&self->buffer_cond, GST_OBJECT_GET_LOCK (self));
|
|
|
|
mem = gst_atomic_queue_pop (self->available_mems);
|
|
GST_OBJECT_UNLOCK (self);
|
|
|
|
gst_object_ref (mem->allocator);
|
|
surface = gst_va_memory_get_surface (mem);
|
|
gst_buffer_append_memory (buffer, mem);
|
|
|
|
GST_TRACE_OBJECT (self, "Prepared surface %#x in buffer %p", surface, buffer);
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
void
|
|
gst_va_allocator_flush (GstAllocator * allocator)
|
|
{
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (allocator);
|
|
|
|
GST_OBJECT_LOCK (self);
|
|
_available_mems_flush (self->display, self->available_mems,
|
|
&self->surface_count);
|
|
g_cond_signal (&self->buffer_cond);
|
|
GST_OBJECT_UNLOCK (self);
|
|
}
|
|
|
|
static gboolean
|
|
gst_va_allocator_try (GstAllocator * allocator)
|
|
{
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (allocator);
|
|
|
|
self->fourcc = 0;
|
|
self->rt_format = 0;
|
|
self->use_derived = FALSE;
|
|
self->img_format = GST_VIDEO_INFO_FORMAT (&self->info);
|
|
|
|
self->surface_format =
|
|
gst_va_video_surface_format_from_image_format (self->img_format,
|
|
self->surface_formats);
|
|
if (self->surface_format == GST_VIDEO_FORMAT_UNKNOWN) {
|
|
/* try a surface without fourcc but rt_format only */
|
|
self->fourcc = 0;
|
|
self->rt_format = gst_va_chroma_from_video_format (self->img_format);
|
|
} else {
|
|
self->fourcc = gst_va_fourcc_from_video_format (self->surface_format);
|
|
self->rt_format = gst_va_chroma_from_video_format (self->surface_format);
|
|
}
|
|
|
|
if (self->rt_format == 0) {
|
|
GST_ERROR_OBJECT (allocator, "Unsupported image format: %s",
|
|
gst_video_format_to_string (self->img_format));
|
|
return FALSE;
|
|
}
|
|
|
|
if (!_update_image_info (self)) {
|
|
GST_ERROR_OBJECT (allocator, "Failed to update allocator info");
|
|
return FALSE;
|
|
}
|
|
|
|
GST_INFO_OBJECT (self,
|
|
"va allocator info, surface format: %s, image format: %s, "
|
|
"use derived: %s, rt format: 0x%x, fourcc: %" GST_FOURCC_FORMAT,
|
|
(self->surface_format == GST_VIDEO_FORMAT_UNKNOWN) ? "unknown"
|
|
: gst_video_format_to_string (self->surface_format),
|
|
gst_video_format_to_string (self->img_format),
|
|
self->use_derived ? "true" : "false", self->rt_format,
|
|
GST_FOURCC_ARGS (self->fourcc));
|
|
return TRUE;
|
|
}
|
|
|
|
gboolean
|
|
gst_va_allocator_set_format (GstAllocator * allocator, GstVideoInfo * info,
|
|
guint usage_hint)
|
|
{
|
|
GstVaAllocator *self;
|
|
gboolean ret;
|
|
|
|
g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), FALSE);
|
|
g_return_val_if_fail (info, FALSE);
|
|
|
|
self = GST_VA_ALLOCATOR (allocator);
|
|
|
|
if (self->surface_count != 0) {
|
|
if (GST_VIDEO_INFO_FORMAT (info) == GST_VIDEO_INFO_FORMAT (&self->info)
|
|
&& GST_VIDEO_INFO_WIDTH (info) == GST_VIDEO_INFO_WIDTH (&self->info)
|
|
&& GST_VIDEO_INFO_HEIGHT (info) == GST_VIDEO_INFO_HEIGHT (&self->info)
|
|
&& usage_hint == self->usage_hint) {
|
|
*info = self->info; /* update callee info (offset & stride) */
|
|
return TRUE;
|
|
}
|
|
return FALSE;
|
|
}
|
|
|
|
self->usage_hint = usage_hint;
|
|
self->info = *info;
|
|
|
|
ret = gst_va_allocator_try (allocator);
|
|
if (ret)
|
|
*info = self->info;
|
|
|
|
return ret;
|
|
}
|
|
|
|
gboolean
|
|
gst_va_allocator_get_format (GstAllocator * allocator, GstVideoInfo * info,
|
|
guint * usage_hint)
|
|
{
|
|
GstVaAllocator *self = GST_VA_ALLOCATOR (allocator);
|
|
|
|
if (GST_VIDEO_INFO_FORMAT (&self->info) == GST_VIDEO_FORMAT_UNKNOWN)
|
|
return FALSE;
|
|
|
|
if (info)
|
|
*info = self->info;
|
|
if (usage_hint)
|
|
*usage_hint = self->usage_hint;
|
|
|
|
return TRUE;
|
|
}
|
|
|
|
/*============ Utilities =====================================================*/
|
|
|
|
VASurfaceID
|
|
gst_va_memory_get_surface (GstMemory * mem)
|
|
{
|
|
VASurfaceID surface = VA_INVALID_ID;
|
|
|
|
if (!mem->allocator)
|
|
return VA_INVALID_ID;
|
|
|
|
if (GST_IS_DMABUF_ALLOCATOR (mem->allocator)) {
|
|
GstVaBufferSurface *buf;
|
|
|
|
buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem),
|
|
gst_va_buffer_surface_quark ());
|
|
if (buf)
|
|
surface = buf->surface;
|
|
} else if (GST_IS_VA_ALLOCATOR (mem->allocator)) {
|
|
GstVaMemory *va_mem = (GstVaMemory *) mem;
|
|
surface = va_mem->surface;
|
|
}
|
|
|
|
return surface;
|
|
}
|
|
|
|
VASurfaceID
|
|
gst_va_buffer_get_surface (GstBuffer * buffer)
|
|
{
|
|
GstMemory *mem;
|
|
|
|
mem = gst_buffer_peek_memory (buffer, 0);
|
|
if (!mem)
|
|
return VA_INVALID_ID;
|
|
|
|
return gst_va_memory_get_surface (mem);
|
|
}
|