/* GStreamer * Copyright (C) 2020 Igalia, S.L. * Author: Víctor Jáquez * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Library General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Library General Public License for more details. * * You should have received a copy of the GNU Library General Public * License along with this library; if not, write to the * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, * Boston, MA 02110-1301, USA. */ /** * SECTION:gstvaallocator * @title: VA allocators * @short_description: VA allocators * @sources: * - gstvaallocator.h * * There are two types of VA allocators: * * * #GstVaAllocator * * #GstVaDmabufAllocator */ #ifdef HAVE_CONFIG_H #include "config.h" #endif #include "gstvaallocator.h" #include #include #include "gstvasurfacecopy.h" #include "gstvavideoformat.h" #include "vasurfaceimage.h" #define GST_CAT_DEFAULT gst_va_memory_debug GST_DEBUG_CATEGORY (gst_va_memory_debug); static void _init_debug_category (void) { #ifndef GST_DISABLE_GST_DEBUG static gsize _init = 0; if (g_once_init_enter (&_init)) { GST_DEBUG_CATEGORY_INIT (gst_va_memory_debug, "vamemory", 0, "VA memory"); g_once_init_leave (&_init, 1); } #endif } /*=========================== Quarks for GstMemory ===========================*/ static GQuark gst_va_buffer_surface_quark (void) { static gsize surface_quark = 0; if (g_once_init_enter (&surface_quark)) { GQuark quark = g_quark_from_string ("GstVaBufferSurface"); g_once_init_leave (&surface_quark, quark); } return surface_quark; } static GQuark gst_va_drm_mod_quark (void) { static gsize drm_mod_quark = 0; if (g_once_init_enter (&drm_mod_quark)) { GQuark quark = g_quark_from_string ("DRMModifier"); g_once_init_leave (&drm_mod_quark, quark); } return drm_mod_quark; } static GQuark gst_va_buffer_aux_surface_quark (void) { static gsize surface_quark = 0; if (g_once_init_enter (&surface_quark)) { GQuark quark = g_quark_from_string ("GstVaBufferAuxSurface"); g_once_init_leave (&surface_quark, quark); } return surface_quark; } /*========================= GstVaBufferSurface ===============================*/ typedef struct _GstVaBufferSurface GstVaBufferSurface; struct _GstVaBufferSurface { GstVaDisplay *display; VASurfaceID surface; guint n_mems; GstMemory *mems[GST_VIDEO_MAX_PLANES]; gint ref_count; gint ref_mems_count; }; static void gst_va_buffer_surface_unref (gpointer data) { GstVaBufferSurface *buf = data; g_return_if_fail (buf && GST_IS_VA_DISPLAY (buf->display)); if (g_atomic_int_dec_and_test (&buf->ref_count)) { GST_LOG_OBJECT (buf->display, "Destroying surface %#x", buf->surface); va_destroy_surfaces (buf->display, &buf->surface, 1); gst_clear_object (&buf->display); g_slice_free (GstVaBufferSurface, buf); } } static GstVaBufferSurface * gst_va_buffer_surface_new (VASurfaceID surface, GstVideoFormat format, gint width, gint height) { GstVaBufferSurface *buf = g_slice_new (GstVaBufferSurface); g_atomic_int_set (&buf->ref_count, 0); g_atomic_int_set (&buf->ref_mems_count, 0); buf->surface = surface; buf->display = NULL; buf->n_mems = 0; return buf; } /*=========================== GstVaMemoryPool ================================*/ /* queue for disposed surfaces */ typedef struct _GstVaMemoryPool GstVaMemoryPool; struct _GstVaMemoryPool { GstAtomicQueue *queue; gint surface_count; GMutex lock; }; #define GST_VA_MEMORY_POOL_CAST(obj) ((GstVaMemoryPool *)obj) #define GST_VA_MEMORY_POOL_LOCK(obj) g_mutex_lock (&GST_VA_MEMORY_POOL_CAST(obj)->lock) #define GST_VA_MEMORY_POOL_UNLOCK(obj) g_mutex_unlock (&GST_VA_MEMORY_POOL_CAST(obj)->lock) static void gst_va_memory_pool_init (GstVaMemoryPool * self) { self->queue = gst_atomic_queue_new (2); g_mutex_init (&self->lock); self->surface_count = 0; } static void gst_va_memory_pool_finalize (GstVaMemoryPool * self) { g_mutex_clear (&self->lock); gst_atomic_queue_unref (self->queue); } static void gst_va_memory_pool_flush_unlocked (GstVaMemoryPool * self, GstVaDisplay * display) { GstMemory *mem; GstVaBufferSurface *buf; while ((mem = gst_atomic_queue_pop (self->queue))) { /* destroy the surface */ buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_surface_quark ()); if (buf) { if (g_atomic_int_dec_and_test (&buf->ref_count)) { GST_LOG ("Destroying surface %#x", buf->surface); va_destroy_surfaces (display, &buf->surface, 1); self->surface_count -= 1; /* GstVaDmabufAllocator */ g_slice_free (GstVaBufferSurface, buf); } } else { self->surface_count -= 1; /* GstVaAllocator */ } GST_MINI_OBJECT_CAST (mem)->dispose = NULL; /* when mem are pushed available queue its allocator is unref, * then now it is required to ref the allocator here because * memory's finalize will unref it again */ gst_object_ref (mem->allocator); gst_memory_unref (mem); } } static void gst_va_memory_pool_flush (GstVaMemoryPool * self, GstVaDisplay * display) { GST_VA_MEMORY_POOL_LOCK (self); gst_va_memory_pool_flush_unlocked (self, display); GST_VA_MEMORY_POOL_UNLOCK (self); } static inline void gst_va_memory_pool_push (GstVaMemoryPool * self, GstMemory * mem) { gst_atomic_queue_push (self->queue, gst_memory_ref (mem)); } static inline GstMemory * gst_va_memory_pool_pop (GstVaMemoryPool * self) { return gst_atomic_queue_pop (self->queue); } static inline GstMemory * gst_va_memory_pool_peek (GstVaMemoryPool * self) { return gst_atomic_queue_peek (self->queue); } static inline guint gst_va_memory_pool_surface_count (GstVaMemoryPool * self) { return g_atomic_int_get (&self->surface_count); } static inline void gst_va_memory_pool_surface_inc (GstVaMemoryPool * self) { g_atomic_int_inc (&self->surface_count); } /*=========================== GstVaDmabufAllocator ===========================*/ /** * GstVaDmabufAllocator: * * A pooled memory allocator backed by the DMABufs exported from a * VASurfaceID. Also it is possible to import DMAbufs into a * VASurfaceID. * * Since: 1.22 */ typedef struct _GstVaDmabufAllocator GstVaDmabufAllocator; typedef struct _GstVaDmabufAllocatorClass GstVaDmabufAllocatorClass; struct _GstVaDmabufAllocator { GstDmaBufAllocator parent; GstVaDisplay *display; GstMemoryMapFunction parent_map; GstMemoryCopyFunction parent_copy; GstVideoInfo info; guint usage_hint; GstVaSurfaceCopy *copy; GstVaMemoryPool pool; }; struct _GstVaDmabufAllocatorClass { GstDmaBufAllocatorClass parent_class; }; #define gst_va_dmabuf_allocator_parent_class dmabuf_parent_class G_DEFINE_TYPE_WITH_CODE (GstVaDmabufAllocator, gst_va_dmabuf_allocator, GST_TYPE_DMABUF_ALLOCATOR, _init_debug_category ()); static GstVaSurfaceCopy * _ensure_surface_copy (GstVaSurfaceCopy ** old, GstVaDisplay * display, GstVideoInfo * info) { GstVaSurfaceCopy *surface_copy; surface_copy = g_atomic_pointer_get (old); if (!surface_copy) { surface_copy = gst_va_surface_copy_new (display, info); /* others create a new one and set it before us */ if (surface_copy && !g_atomic_pointer_compare_and_exchange (old, NULL, surface_copy)) { gst_va_surface_copy_free (surface_copy); surface_copy = g_atomic_pointer_get (old); } } return surface_copy; } /* If a buffer contains multiple memories (dmabuf objects) its very * difficult to provide a realiable way to fast-copy single memories: * While VA API sees surfaces with dependant dmabufs, GStreamer only * copies dmabufs in isolation; trying to solve it while keeping a * reference of the copied buffer and dmabuf index is very fragile. */ static GstMemory * gst_va_dmabuf_mem_copy (GstMemory * gmem, gssize offset, gssize size) { GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (gmem->allocator); GstVaBufferSurface *buf; guint64 *drm_mod; gsize mem_size; buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (gmem), gst_va_buffer_surface_quark ()); drm_mod = gst_mini_object_get_qdata (GST_MINI_OBJECT (gmem), gst_va_drm_mod_quark ()); /* 0 is DRM_FORMAT_MOD_LINEAR, we do not include its header now. */ if (buf->n_mems > 1 && *drm_mod != 0) { GST_ERROR_OBJECT (self, "Failed to copy multi-dmabuf because non-linear " "modifier: %#" G_GINT64_MODIFIER "x.", *drm_mod); return NULL; } /* check if it's full memory copy */ mem_size = gst_memory_get_sizes (gmem, NULL, NULL); if (size == -1) size = mem_size > offset ? mem_size - offset : 0; /* @XXX: if one-memory buffer it's possible to copy */ if (offset == 0 && size == mem_size && buf->n_mems == 1) { GstVaBufferSurface *buf_copy = NULL; GstMemory *copy; GstVaSurfaceCopy *copy_func; GST_VA_MEMORY_POOL_LOCK (&self->pool); copy = gst_va_memory_pool_pop (&self->pool); GST_VA_MEMORY_POOL_UNLOCK (&self->pool); if (copy) { gst_object_ref (copy->allocator); buf_copy = gst_mini_object_get_qdata (GST_MINI_OBJECT (copy), gst_va_buffer_surface_quark ()); g_assert (g_atomic_int_get (&buf_copy->ref_mems_count) == 0); g_atomic_int_add (&buf_copy->ref_mems_count, 1); } else { GstBuffer *buffer = gst_buffer_new (); if (!gst_va_dmabuf_allocator_setup_buffer (gmem->allocator, buffer)) { GST_WARNING_OBJECT (self, "Failed to create a new dmabuf memory"); return NULL; } copy = gst_buffer_get_memory (buffer, 0); gst_buffer_unref (buffer); buf_copy = gst_mini_object_get_qdata (GST_MINI_OBJECT (copy), gst_va_buffer_surface_quark ()); } g_assert (buf_copy->n_mems == 1); copy_func = _ensure_surface_copy (&self->copy, self->display, &self->info); if (copy_func && gst_va_surface_copy (copy_func, buf_copy->surface, buf->surface)) return copy; gst_memory_unref (copy); /* try system memory */ } if (*drm_mod != 0) { GST_ERROR_OBJECT (self, "Failed to copy dmabuf because non-linear " "modifier: %#" G_GINT64_MODIFIER "x.", *drm_mod); return NULL; } /* fallback to system memory */ return self->parent_copy (gmem, offset, size); } static gpointer gst_va_dmabuf_mem_map (GstMemory * gmem, gsize maxsize, GstMapFlags flags) { GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (gmem->allocator); VASurfaceID surface = gst_va_memory_get_surface (gmem); guint64 *drm_mod; drm_mod = gst_mini_object_get_qdata (GST_MINI_OBJECT (gmem), gst_va_drm_mod_quark ()); /* 0 is DRM_FORMAT_MOD_LINEAR, we do not include its header now. */ if (*drm_mod != 0) { GST_ERROR_OBJECT (self, "Failed to map the dmabuf because the modifier " "is: %#" G_GINT64_MODIFIER "x, which is not linear.", *drm_mod); return NULL; } va_sync_surface (self->display, surface); return self->parent_map (gmem, maxsize, flags); } static void gst_va_dmabuf_allocator_finalize (GObject * object) { GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (object); g_clear_pointer (&self->copy, gst_va_surface_copy_free); gst_va_memory_pool_finalize (&self->pool); gst_clear_object (&self->display); G_OBJECT_CLASS (dmabuf_parent_class)->finalize (object); } static void gst_va_dmabuf_allocator_dispose (GObject * object) { GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (object); gst_va_memory_pool_flush_unlocked (&self->pool, self->display); if (gst_va_memory_pool_surface_count (&self->pool) != 0) { GST_WARNING_OBJECT (self, "Surfaces leaked: %d", gst_va_memory_pool_surface_count (&self->pool)); } G_OBJECT_CLASS (dmabuf_parent_class)->dispose (object); } static void gst_va_dmabuf_allocator_class_init (GstVaDmabufAllocatorClass * klass) { GObjectClass *object_class = G_OBJECT_CLASS (klass); object_class->dispose = gst_va_dmabuf_allocator_dispose; object_class->finalize = gst_va_dmabuf_allocator_finalize; } static void gst_va_dmabuf_allocator_init (GstVaDmabufAllocator * self) { GstAllocator *allocator = GST_ALLOCATOR (self); self->parent_map = allocator->mem_map; allocator->mem_map = gst_va_dmabuf_mem_map; self->parent_copy = allocator->mem_copy; allocator->mem_copy = gst_va_dmabuf_mem_copy; gst_va_memory_pool_init (&self->pool); } /** * gst_va_dmabuf_allocator_new: * @display: a #GstVaDisplay * * Instanciate a new pooled allocator backed with both DMABuf and * VASurfaceID. * * Returns: a new allocated #GstAllocator * * Since: 1.22 */ GstAllocator * gst_va_dmabuf_allocator_new (GstVaDisplay * display) { GstVaDmabufAllocator *self; g_return_val_if_fail (GST_IS_VA_DISPLAY (display), NULL); self = g_object_new (GST_TYPE_VA_DMABUF_ALLOCATOR, NULL); self->display = gst_object_ref (display); gst_object_ref_sink (self); return GST_ALLOCATOR (self); } static inline goffset _get_fd_size (gint fd) { return lseek (fd, 0, SEEK_END); } static gboolean gst_va_dmabuf_memory_release (GstMiniObject * mini_object) { GstMemory *mem = GST_MEMORY_CAST (mini_object); GstVaBufferSurface *buf; GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (mem->allocator); guint i; buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_surface_quark ()); if (!buf) return TRUE; /* free this unknown buffer */ /* if this is the last reference to the GstVaBufferSurface, iterates * its array of memories to push them into the queue with thread * safetly. */ GST_VA_MEMORY_POOL_LOCK (&self->pool); if (g_atomic_int_dec_and_test (&buf->ref_mems_count)) { for (i = 0; i < buf->n_mems; i++) { GST_LOG_OBJECT (self, "releasing %p: dmabuf %d, va surface %#x", buf->mems[i], gst_dmabuf_memory_get_fd (buf->mems[i]), buf->surface); gst_va_memory_pool_push (&self->pool, buf->mems[i]); } } GST_VA_MEMORY_POOL_UNLOCK (&self->pool); /* note: if ref_mem_count doesn't reach zero, that memory will * "float" until it's pushed back into the pool by the last va * buffer surface ref */ /* Keep last in case we are holding on the last allocator ref */ gst_object_unref (mem->allocator); /* don't call mini_object's free */ return FALSE; } /* Creates an exported VASurfaceID and adds it as @buffer's memories * qdata * * If @info is not NULL, a dummy (non-pooled) buffer is created to * update offsets and strides, and it has to be unrefed immediately. */ static gboolean gst_va_dmabuf_allocator_setup_buffer_full (GstAllocator * allocator, GstBuffer * buffer, GstVideoInfo * info) { GstVaBufferSurface *buf; GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator); GstVideoFormat format; VADRMPRIMESurfaceDescriptor desc = { 0, }; VASurfaceAttribExternalBuffers *extbuf = NULL, ext_buf; VASurfaceID surface; guint32 i, fourcc, rt_format, export_flags; GDestroyNotify buffer_destroy = NULL; gsize object_offset[4]; g_return_val_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator), FALSE); format = GST_VIDEO_INFO_FORMAT (&self->info); fourcc = gst_va_fourcc_from_video_format (format); rt_format = gst_va_chroma_from_video_format (format); if (fourcc == 0 || rt_format == 0) { GST_ERROR_OBJECT (allocator, "Unsupported format: %s", gst_video_format_to_string (GST_VIDEO_INFO_FORMAT (&self->info))); return FALSE; } /* HACK(victor): disable tiling for i965 driver for RGB formats */ if (GST_VA_DISPLAY_IS_IMPLEMENTATION (self->display, INTEL_I965) && GST_VIDEO_INFO_IS_RGB (&self->info)) { /* *INDENT-OFF* */ ext_buf = (VASurfaceAttribExternalBuffers) { .width = GST_VIDEO_INFO_WIDTH (&self->info), .height = GST_VIDEO_INFO_HEIGHT (&self->info), .num_planes = GST_VIDEO_INFO_N_PLANES (&self->info), .pixel_format = fourcc, }; /* *INDENT-ON* */ extbuf = &ext_buf; } if (!va_create_surfaces (self->display, rt_format, fourcc, GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info), self->usage_hint, extbuf, &surface, 1)) return FALSE; /* workaround for missing layered dmabuf formats in i965 */ if (GST_VA_DISPLAY_IS_IMPLEMENTATION (self->display, INTEL_I965) && (fourcc == VA_FOURCC_YUY2 || fourcc == VA_FOURCC_UYVY)) { /* These are not representable as separate planes */ export_flags = VA_EXPORT_SURFACE_COMPOSED_LAYERS; } else { /* Each layer will contain exactly one plane. For example, an NV12 * surface will be exported as two layers */ export_flags = VA_EXPORT_SURFACE_SEPARATE_LAYERS; } export_flags |= VA_EXPORT_SURFACE_READ_WRITE; if (!va_export_surface_to_dmabuf (self->display, surface, export_flags, &desc)) goto failed; g_assert (GST_VIDEO_INFO_N_PLANES (&self->info) == desc.num_layers); if (fourcc != desc.fourcc) { GST_ERROR ("Unsupported fourcc: %" GST_FOURCC_FORMAT, GST_FOURCC_ARGS (desc.fourcc)); goto failed; } if (desc.num_objects == 0) { GST_ERROR ("Failed to export surface to dmabuf"); goto failed; } buf = gst_va_buffer_surface_new (surface, format, desc.width, desc.height); if (G_UNLIKELY (info)) { *info = self->info; GST_VIDEO_INFO_SIZE (info) = 0; } buf->n_mems = desc.num_objects; for (i = 0; i < desc.num_objects; i++) { gint fd = desc.objects[i].fd; /* don't rely on prime descriptor reported size since gallium drivers report * different values */ gsize size = _get_fd_size (fd); GstMemory *mem = gst_dmabuf_allocator_alloc (allocator, fd, size); guint64 *drm_mod = g_new (guint64, 1); if (size != desc.objects[i].size) { GST_WARNING_OBJECT (self, "driver bug: fd size (%" G_GSIZE_FORMAT ") differs from object descriptor size (%" G_GUINT32_FORMAT ")", size, desc.objects[i].size); } object_offset[i] = gst_buffer_get_size (buffer); gst_buffer_append_memory (buffer, mem); buf->mems[i] = mem; if (G_LIKELY (!info)) { GST_MINI_OBJECT (mem)->dispose = gst_va_dmabuf_memory_release; g_atomic_int_add (&buf->ref_mems_count, 1); } else { /* if no @info, surface will be destroyed as soon as buffer is * destroyed (e.g. gst_va_dmabuf_allocator_try()) */ buf->display = gst_object_ref (self->display); buffer_destroy = gst_va_buffer_surface_unref; } g_atomic_int_add (&buf->ref_count, 1); gst_mini_object_set_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_surface_quark (), buf, buffer_destroy); *drm_mod = desc.objects[i].drm_format_modifier; gst_mini_object_set_qdata (GST_MINI_OBJECT (mem), gst_va_drm_mod_quark (), drm_mod, g_free); if (G_UNLIKELY (info)) GST_VIDEO_INFO_PLANE_OFFSET (info, i) = GST_VIDEO_INFO_SIZE (info); GST_LOG_OBJECT (self, "buffer %p: new dmabuf %d / surface %#x [%dx%d] " "size %" G_GSIZE_FORMAT " drm mod %#" G_GINT64_MODIFIER "x", buffer, fd, surface, GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info), size, *drm_mod); } if (G_UNLIKELY (info)) { GST_VIDEO_INFO_SIZE (info) = gst_buffer_get_size (buffer); for (i = 0; i < desc.num_layers; i++) { g_assert (desc.layers[i].num_planes == 1); GST_VIDEO_INFO_PLANE_OFFSET (info, i) = object_offset[desc.layers[i].object_index[0]] + desc.layers[i].offset[0]; GST_VIDEO_INFO_PLANE_STRIDE (info, i) = desc.layers[i].pitch[0]; } } else { gst_va_memory_pool_surface_inc (&self->pool); } return TRUE; failed: { va_destroy_surfaces (self->display, &surface, 1); return FALSE; } } /** * gst_va_dmabuf_allocator_setup_buffer: * @allocator: a #GstAllocator * @buffer: an empty #GstBuffer * * This funciton creates a new VASurfaceID and exposes its DMABufs, * later it populates the @buffer with those DMABufs. * * Return: %TRUE if @buffer is populated correctly; %FALSE otherwise. * * Since: 1.22 */ gboolean gst_va_dmabuf_allocator_setup_buffer (GstAllocator * allocator, GstBuffer * buffer) { return gst_va_dmabuf_allocator_setup_buffer_full (allocator, buffer, NULL); } static VASurfaceID gst_va_dmabuf_allocator_prepare_buffer_unlocked (GstVaDmabufAllocator * self, GstBuffer * buffer) { GstMemory *mems[GST_VIDEO_MAX_PLANES] = { 0, }; GstVaBufferSurface *buf; gint i, j, idx; mems[0] = gst_va_memory_pool_pop (&self->pool); if (!mems[0]) return VA_INVALID_ID; buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mems[0]), gst_va_buffer_surface_quark ()); if (!buf) return VA_INVALID_ID; if (buf->surface == VA_INVALID_ID) return VA_INVALID_ID; for (idx = 1; idx < buf->n_mems; idx++) { /* grab next memory from queue */ { GstMemory *mem; GstVaBufferSurface *pbuf; mem = gst_va_memory_pool_peek (&self->pool); if (!mem) return VA_INVALID_ID; pbuf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_surface_quark ()); if (!pbuf) return VA_INVALID_ID; if (pbuf->surface != buf->surface) { GST_WARNING_OBJECT (self, "expecting memory with surface %#x but got %#x: " "possible memory interweaving", buf->surface, pbuf->surface); return VA_INVALID_ID; } } mems[idx] = gst_va_memory_pool_pop (&self->pool); }; /* append memories */ for (i = 0; i < buf->n_mems; i++) { gboolean found = FALSE; /* find next memory to append */ for (j = 0; j < idx; j++) { if (buf->mems[i] == mems[j]) { found = TRUE; break; } } /* if not found, free all the popped memories and bail */ if (!found) { if (!buf->display) buf->display = gst_object_ref (self->display); for (j = 0; j < idx; j++) { gst_object_ref (buf->mems[j]->allocator); GST_MINI_OBJECT (mems[j])->dispose = NULL; gst_memory_unref (mems[j]); } return VA_INVALID_ID; } g_atomic_int_add (&buf->ref_mems_count, 1); gst_object_ref (buf->mems[i]->allocator); gst_buffer_append_memory (buffer, buf->mems[i]); GST_LOG ("bufer %p: memory %p - dmabuf %d / surface %#x", buffer, buf->mems[i], gst_dmabuf_memory_get_fd (buf->mems[i]), gst_va_memory_get_surface (buf->mems[i])); } return buf->surface; } /** * gst_va_dmabuf_allocator_prepare_buffer: * @allocator: a #GstAllocator * @buffer: an empty #GstBuffer * * This method will populate @buffer with pooled VASurfaceID/DMABuf * memories. It doesn't allocate new VASurfacesID. * * Returns: %TRUE if @buffer was populated correctly; %FALSE * otherwise. * * Since: 1.22 */ gboolean gst_va_dmabuf_allocator_prepare_buffer (GstAllocator * allocator, GstBuffer * buffer) { GstVaDmabufAllocator *self; VASurfaceID surface; g_return_val_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator), FALSE); self = GST_VA_DMABUF_ALLOCATOR (allocator); GST_VA_MEMORY_POOL_LOCK (&self->pool); surface = gst_va_dmabuf_allocator_prepare_buffer_unlocked (self, buffer); GST_VA_MEMORY_POOL_UNLOCK (&self->pool); return (surface != VA_INVALID_ID); } /** * gst_va_dmabuf_allocator_flush: * @allocator: a #GstAllocator * * Removes all the memories in @allocator's pool. * * Since: 1.22 */ void gst_va_dmabuf_allocator_flush (GstAllocator * allocator) { GstVaDmabufAllocator *self; g_return_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator)); self = GST_VA_DMABUF_ALLOCATOR (allocator); gst_va_memory_pool_flush (&self->pool, self->display); } /** * gst_va_dmabuf_allocator_try: * @allocator: a #GstAllocator * * Try to allocate a test buffer in order to verify that the * allocator's configuration is valid. * * Returns: %TRUE if the configuration is valid; %FALSE otherwise. * * Since: 1.22 */ static gboolean gst_va_dmabuf_allocator_try (GstAllocator * allocator) { GstBuffer *buffer; GstVaDmabufAllocator *self; GstVideoInfo info; gboolean ret; g_return_val_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator), FALSE); self = GST_VA_DMABUF_ALLOCATOR (allocator); info = self->info; buffer = gst_buffer_new (); ret = gst_va_dmabuf_allocator_setup_buffer_full (allocator, buffer, &info); gst_buffer_unref (buffer); if (ret) self->info = info; return ret; } /** * gst_va_dmabuf_allocator_set_format: * @allocator: a #GstAllocator * @info: a #GstVideoInfo * @usage_hint: VA usage hint * * Sets the configuration defined by @info and @usage_hint for * @allocator, and it tries the configuration, if @allocator has not * allocated memories yet. * * If @allocator has memory allocated already, and frame size and * format in @info are the same as currently configured in @allocator, * the rest of @info parameters are updated internally. * * Returns: %TRUE if the configuration is valid or updated; %FALSE if * configuration is not valid or not updated. * * Since: 1.22 */ gboolean gst_va_dmabuf_allocator_set_format (GstAllocator * allocator, GstVideoInfo * info, guint usage_hint) { GstVaDmabufAllocator *self; gboolean ret; g_return_val_if_fail (GST_IS_VA_DMABUF_ALLOCATOR (allocator), FALSE); g_return_val_if_fail (info, FALSE); self = GST_VA_DMABUF_ALLOCATOR (allocator); if (gst_va_memory_pool_surface_count (&self->pool) != 0) { if (GST_VIDEO_INFO_FORMAT (info) == GST_VIDEO_INFO_FORMAT (&self->info) && GST_VIDEO_INFO_WIDTH (info) == GST_VIDEO_INFO_WIDTH (&self->info) && GST_VIDEO_INFO_HEIGHT (info) == GST_VIDEO_INFO_HEIGHT (&self->info) && usage_hint == self->usage_hint) { *info = self->info; /* update callee info (offset & stride) */ return TRUE; } return FALSE; } self->usage_hint = usage_hint; self->info = *info; g_clear_pointer (&self->copy, gst_va_surface_copy_free); ret = gst_va_dmabuf_allocator_try (allocator); if (ret) *info = self->info; return ret; } /** * gst_va_dmabuf_allocator_get_format: * @allocator: a #GstAllocator * @info: (out) (optional): a #GstVideoInfo * @usage_hint: (out) (optional): VA usage hint * * Gets current internal configuration of @allocator. * * Returns: %TRUE if @allocator is already configured; %FALSE * otherwise. * * Since: 1.22 */ gboolean gst_va_dmabuf_allocator_get_format (GstAllocator * allocator, GstVideoInfo * info, guint * usage_hint) { GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (allocator); if (GST_VIDEO_INFO_FORMAT (&self->info) == GST_VIDEO_FORMAT_UNKNOWN) return FALSE; if (info) *info = self->info; if (usage_hint) *usage_hint = self->usage_hint; return TRUE; } /** * gst_va_dmabuf_memories_setup: * @display: a #GstVaDisplay * @info: a #GstVideoInfo * @n_planes: number of planes * @mem: (array fixed-size=4) (element-type GstMemory): Memories. One * per plane. * @fds: (array length=n_planes) (element-type uintptr_t): array of * DMABuf file descriptors. * @offset: (array fixed-size=4) (element-type gsize): array of memory * offsets. * @usage_hint: VA usage hint. * * It imports the array of @mem, representing a single frame, into a * VASurfaceID and it's attached into every @mem. * * Returns: %TRUE if frame is imported correctly into a VASurfaceID; * %FALSE otherwise. * * Since: 1.22 */ /* XXX: use a surface pool to control the created surfaces */ /* XXX: remove n_planes argument and use GST_VIDEO_INFO_N_PLANES (info) */ gboolean gst_va_dmabuf_memories_setup (GstVaDisplay * display, GstVideoInfo * info, guint n_planes, GstMemory * mem[GST_VIDEO_MAX_PLANES], uintptr_t * fds, gsize offset[GST_VIDEO_MAX_PLANES], guint usage_hint) { GstVideoFormat format; GstVaBufferSurface *buf; /* *INDENT-OFF* */ VASurfaceAttribExternalBuffers ext_buf = { .width = GST_VIDEO_INFO_WIDTH (info), .height = GST_VIDEO_INFO_HEIGHT (info), .data_size = GST_VIDEO_INFO_SIZE (info), .num_planes = GST_VIDEO_INFO_N_PLANES (info), .buffers = fds, .num_buffers = GST_VIDEO_INFO_N_PLANES (info), }; /* *INDENT-ON* */ VASurfaceID surface; guint32 fourcc, rt_format; guint i; gboolean ret; g_return_val_if_fail (GST_IS_VA_DISPLAY (display), FALSE); g_return_val_if_fail (n_planes > 0 && n_planes <= GST_VIDEO_MAX_PLANES, FALSE); format = GST_VIDEO_INFO_FORMAT (info); if (format == GST_VIDEO_FORMAT_UNKNOWN) return FALSE; rt_format = gst_va_chroma_from_video_format (format); if (rt_format == 0) return FALSE; fourcc = gst_va_fourcc_from_video_format (format); if (fourcc == 0) return FALSE; ext_buf.pixel_format = fourcc; for (i = 0; i < n_planes; i++) { ext_buf.pitches[i] = GST_VIDEO_INFO_PLANE_STRIDE (info, i); ext_buf.offsets[i] = offset[i]; } ret = va_create_surfaces (display, rt_format, ext_buf.pixel_format, ext_buf.width, ext_buf.height, usage_hint, &ext_buf, &surface, 1); if (!ret) return FALSE; GST_LOG_OBJECT (display, "Created surface %#x [%dx%d]", surface, ext_buf.width, ext_buf.height); buf = gst_va_buffer_surface_new (surface, rt_format, ext_buf.width, ext_buf.height); buf->display = gst_object_ref (display); buf->n_mems = n_planes; memcpy (buf->mems, mem, sizeof (buf->mems)); for (i = 0; i < n_planes; i++) { g_atomic_int_add (&buf->ref_count, 1); gst_mini_object_set_qdata (GST_MINI_OBJECT (mem[i]), gst_va_buffer_surface_quark (), buf, gst_va_buffer_surface_unref); GST_INFO_OBJECT (display, "setting surface %#x to dmabuf fd %d", buf->surface, gst_dmabuf_memory_get_fd (mem[i])); } return TRUE; } /*===================== GstVaAllocator / GstVaMemory =========================*/ /** * GstVaAllocator: * * A pooled memory allocator backed by VASurfaceID. * * Since: 1.22 */ typedef struct _GstVaAllocator GstVaAllocator; typedef struct _GstVaAllocatorClass GstVaAllocatorClass; struct _GstVaAllocator { GstAllocator parent; GstVaDisplay *display; GstVaFeature feat_use_derived; gboolean use_derived; GArray *surface_formats; GstVideoFormat surface_format; GstVideoFormat img_format; guint32 fourcc; guint32 rt_format; GstVideoInfo derived_info; GstVideoInfo info; guint usage_hint; guint32 hacks; GstVaSurfaceCopy *copy; GstVaMemoryPool pool; }; struct _GstVaAllocatorClass { GstAllocatorClass parent_class; }; typedef struct _GstVaMemory GstVaMemory; struct _GstVaMemory { GstMemory mem; VASurfaceID surface; GstVideoFormat surface_format; VAImage image; gpointer mapped_data; GstMapFlags prev_mapflags; gint map_count; gboolean is_derived; gboolean is_dirty; GMutex lock; }; G_DEFINE_TYPE_WITH_CODE (GstVaAllocator, gst_va_allocator, GST_TYPE_ALLOCATOR, _init_debug_category ()); static gboolean _va_unmap (GstVaMemory * mem); static void gst_va_allocator_finalize (GObject * object) { GstVaAllocator *self = GST_VA_ALLOCATOR (object); g_clear_pointer (&self->copy, gst_va_surface_copy_free); gst_va_memory_pool_finalize (&self->pool); g_clear_pointer (&self->surface_formats, g_array_unref); gst_clear_object (&self->display); G_OBJECT_CLASS (gst_va_allocator_parent_class)->finalize (object); } static void gst_va_allocator_dispose (GObject * object) { GstVaAllocator *self = GST_VA_ALLOCATOR (object); gst_va_memory_pool_flush_unlocked (&self->pool, self->display); if (gst_va_memory_pool_surface_count (&self->pool) != 0) { GST_WARNING_OBJECT (self, "Surfaces leaked: %d", gst_va_memory_pool_surface_count (&self->pool)); } G_OBJECT_CLASS (gst_va_allocator_parent_class)->dispose (object); } static void _va_free (GstAllocator * allocator, GstMemory * mem) { GstVaAllocator *self = GST_VA_ALLOCATOR (allocator); GstVaMemory *va_mem = (GstVaMemory *) mem; if (va_mem->mapped_data) { g_warning (G_STRLOC ":%s: Freeing memory %p still mapped", G_STRFUNC, va_mem); _va_unmap (va_mem); } if (va_mem->surface != VA_INVALID_ID && mem->parent == NULL) { GST_LOG_OBJECT (self, "Destroying surface %#x", va_mem->surface); va_destroy_surfaces (self->display, &va_mem->surface, 1); } g_mutex_clear (&va_mem->lock); g_slice_free (GstVaMemory, va_mem); } static void gst_va_allocator_class_init (GstVaAllocatorClass * klass) { GstAllocatorClass *allocator_class = GST_ALLOCATOR_CLASS (klass); GObjectClass *object_class = G_OBJECT_CLASS (klass); object_class->dispose = gst_va_allocator_dispose; object_class->finalize = gst_va_allocator_finalize; allocator_class->free = _va_free; } static inline void _clean_mem (GstVaMemory * mem) { memset (&mem->image, 0, sizeof (mem->image)); mem->image.image_id = VA_INVALID_ID; mem->image.buf = VA_INVALID_ID; mem->is_derived = TRUE; mem->is_dirty = FALSE; mem->prev_mapflags = 0; mem->mapped_data = NULL; } static void _reset_mem (GstVaMemory * mem, GstAllocator * allocator, gsize size) { _clean_mem (mem); g_atomic_int_set (&mem->map_count, 0); g_mutex_init (&mem->lock); gst_memory_init (GST_MEMORY_CAST (mem), 0, allocator, NULL, size, 0 /* align */ , 0 /* offset */ , size); } static inline void _update_info (GstVideoInfo * info, const VAImage * image) { guint i; for (i = 0; i < image->num_planes; i++) { GST_VIDEO_INFO_PLANE_OFFSET (info, i) = image->offsets[i]; GST_VIDEO_INFO_PLANE_STRIDE (info, i) = image->pitches[i]; } GST_VIDEO_INFO_SIZE (info) = image->data_size; } static inline gboolean _update_image_info (GstVaAllocator * va_allocator) { VASurfaceID surface; VAImage image = {.image_id = VA_INVALID_ID, }; /* Create a test surface first */ if (!va_create_surfaces (va_allocator->display, va_allocator->rt_format, va_allocator->fourcc, GST_VIDEO_INFO_WIDTH (&va_allocator->info), GST_VIDEO_INFO_HEIGHT (&va_allocator->info), va_allocator->usage_hint, NULL, &surface, 1)) { GST_ERROR_OBJECT (va_allocator, "Failed to create a test surface"); return FALSE; } GST_DEBUG_OBJECT (va_allocator, "Created surface %#x [%dx%d]", surface, GST_VIDEO_INFO_WIDTH (&va_allocator->info), GST_VIDEO_INFO_HEIGHT (&va_allocator->info)); /* Try derived first, but different formats can never derive */ if (va_allocator->feat_use_derived != GST_VA_FEATURE_DISABLED && va_allocator->surface_format == va_allocator->img_format) { if (va_get_derive_image (va_allocator->display, surface, &image)) { va_allocator->use_derived = TRUE; va_allocator->derived_info = va_allocator->info; _update_info (&va_allocator->derived_info, &image); va_destroy_image (va_allocator->display, image.image_id); } image.image_id = VA_INVALID_ID; /* reset it */ } if (va_allocator->feat_use_derived == GST_VA_FEATURE_ENABLED && !va_allocator->use_derived) { GST_WARNING_OBJECT (va_allocator, "Derived images are disabled."); va_allocator->feat_use_derived = GST_VA_FEATURE_DISABLED; } /* Then we try to create a image. */ if (!va_create_image (va_allocator->display, va_allocator->img_format, GST_VIDEO_INFO_WIDTH (&va_allocator->info), GST_VIDEO_INFO_HEIGHT (&va_allocator->info), &image)) { va_destroy_surfaces (va_allocator->display, &surface, 1); return FALSE; } _update_info (&va_allocator->info, &image); va_destroy_image (va_allocator->display, image.image_id); va_destroy_surfaces (va_allocator->display, &surface, 1); return TRUE; } static gpointer _va_map_unlocked (GstVaMemory * mem, GstMapFlags flags) { GstAllocator *allocator = GST_MEMORY_CAST (mem)->allocator; GstVideoInfo *info; GstVaAllocator *va_allocator; GstVaDisplay *display; gboolean use_derived; g_return_val_if_fail (mem->surface != VA_INVALID_ID, NULL); g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), NULL); if (g_atomic_int_get (&mem->map_count) > 0) { if (!(mem->prev_mapflags & flags) || !mem->mapped_data) return NULL; else goto success; } va_allocator = GST_VA_ALLOCATOR (allocator); display = va_allocator->display; if (flags & GST_MAP_WRITE) { mem->is_dirty = TRUE; } else { /* GST_MAP_READ only */ mem->is_dirty = FALSE; } if (flags & GST_MAP_VA) { mem->mapped_data = &mem->surface; goto success; } if (va_allocator->feat_use_derived == GST_VA_FEATURE_ENABLED) { use_derived = TRUE; } else if (va_allocator->feat_use_derived == GST_VA_FEATURE_DISABLED) { use_derived = FALSE; } else { switch (gst_va_display_get_implementation (display)) { case GST_VA_IMPLEMENTATION_INTEL_IHD: /* On Gen7+ Intel graphics the memory is mappable but not * cached, so normal memcpy() access is very slow to read, but * it's ok for writing. So let's assume that users won't prefer * direct-mapped memory if they request read access. */ use_derived = va_allocator->use_derived && !(flags & GST_MAP_READ); break; case GST_VA_IMPLEMENTATION_INTEL_I965: /* YUV derived images are tiled, so writing them is also * problematic */ use_derived = va_allocator->use_derived && !((flags & GST_MAP_READ) || ((flags & GST_MAP_WRITE) && GST_VIDEO_INFO_IS_YUV (&va_allocator->derived_info))); break; case GST_VA_IMPLEMENTATION_MESA_GALLIUM: /* Reading RGB derived images, with non-standard resolutions, * looks like tiled too. TODO(victor): fill a bug in Mesa. */ use_derived = va_allocator->use_derived && !((flags & GST_MAP_READ) && GST_VIDEO_INFO_IS_RGB (&va_allocator->derived_info)); break; default: use_derived = va_allocator->use_derived; break; } } if (use_derived) info = &va_allocator->derived_info; else info = &va_allocator->info; if (!va_ensure_image (display, mem->surface, info, &mem->image, use_derived)) return NULL; mem->is_derived = use_derived; if (!mem->is_derived) { if (!va_get_image (display, mem->surface, &mem->image)) goto fail; } if (!va_map_buffer (display, mem->image.buf, &mem->mapped_data)) goto fail; success: { mem->prev_mapflags = flags; g_atomic_int_add (&mem->map_count, 1); return mem->mapped_data; } fail: { va_destroy_image (display, mem->image.image_id); _clean_mem (mem); return NULL; } } static gpointer _va_map (GstVaMemory * mem, gsize maxsize, GstMapFlags flags) { gpointer data; g_mutex_lock (&mem->lock); data = _va_map_unlocked (mem, flags); g_mutex_unlock (&mem->lock); return data; } static gboolean _va_unmap_unlocked (GstVaMemory * mem) { GstAllocator *allocator = GST_MEMORY_CAST (mem)->allocator; GstVaDisplay *display; gboolean ret = TRUE; if (!g_atomic_int_dec_and_test (&mem->map_count)) return TRUE; if (mem->prev_mapflags & GST_MAP_VA) goto bail; display = GST_VA_ALLOCATOR (allocator)->display; if (mem->image.image_id != VA_INVALID_ID) { if (mem->is_dirty && !mem->is_derived) { ret = va_put_image (display, mem->surface, &mem->image); mem->is_dirty = FALSE; } /* XXX(victor): if is derived and is dirty, create another surface * an replace it in mem */ } ret &= va_unmap_buffer (display, mem->image.buf); ret &= va_destroy_image (display, mem->image.image_id); bail: _clean_mem (mem); return ret; } static gboolean _va_unmap (GstVaMemory * mem) { gboolean ret; g_mutex_lock (&mem->lock); ret = _va_unmap_unlocked (mem); g_mutex_unlock (&mem->lock); return ret; } static GstMemory * _va_share (GstMemory * mem, gssize offset, gssize size) { GstVaMemory *vamem = (GstVaMemory *) mem; GstVaMemory *sub; GstMemory *parent; GST_DEBUG ("%p: share %" G_GSSIZE_FORMAT ", %" G_GSIZE_FORMAT, mem, offset, size); /* find real parent */ if ((parent = vamem->mem.parent) == NULL) parent = (GstMemory *) vamem; if (size == -1) size = mem->maxsize - offset; sub = g_slice_new (GstVaMemory); /* the shared memory is alwyas readonly */ gst_memory_init (GST_MEMORY_CAST (sub), GST_MINI_OBJECT_FLAGS (parent) | GST_MINI_OBJECT_FLAG_LOCK_READONLY, vamem->mem.allocator, parent, vamem->mem.maxsize, vamem->mem.align, vamem->mem.offset + offset, size); sub->surface = vamem->surface; sub->surface_format = vamem->surface_format; _clean_mem (sub); g_atomic_int_set (&sub->map_count, 0); g_mutex_init (&sub->lock); return GST_MEMORY_CAST (sub); } /* XXX(victor): deep copy implementation. */ static GstMemory * _va_copy (GstMemory * mem, gssize offset, gssize size) { GstMemory *copy; GstMapInfo sinfo, dinfo; GstVaAllocator *va_allocator = GST_VA_ALLOCATOR (mem->allocator); GstVaMemory *va_copy, *va_mem = (GstVaMemory *) mem; gsize mem_size; GST_DEBUG ("%p: copy %" G_GSSIZE_FORMAT ", %" G_GSIZE_FORMAT, mem, offset, size); { GST_VA_MEMORY_POOL_LOCK (&va_allocator->pool); copy = gst_va_memory_pool_pop (&va_allocator->pool); GST_VA_MEMORY_POOL_UNLOCK (&va_allocator->pool); if (!copy) { copy = gst_va_allocator_alloc (mem->allocator); if (!copy) { GST_WARNING ("failed to allocate new memory"); return NULL; } } else { gst_object_ref (mem->allocator); } } va_copy = (GstVaMemory *) copy; mem_size = gst_memory_get_sizes (mem, NULL, NULL); if (size == -1) size = mem_size > offset ? mem_size - offset : 0; if (offset == 0 && size == mem_size) { GstVaSurfaceCopy *copy_func; copy_func = _ensure_surface_copy (&va_allocator->copy, va_allocator->display, &va_allocator->info); if (copy_func && gst_va_surface_copy (copy_func, va_copy->surface, va_mem->surface)) return copy; } if (!gst_memory_map (mem, &sinfo, GST_MAP_READ)) { GST_WARNING ("failed to map memory to copy"); return NULL; } if (!gst_memory_map (copy, &dinfo, GST_MAP_WRITE)) { GST_WARNING ("could not write map memory %p", copy); gst_allocator_free (mem->allocator, copy); gst_memory_unmap (mem, &sinfo); return NULL; } memcpy (dinfo.data, sinfo.data + offset, size); gst_memory_unmap (copy, &dinfo); gst_memory_unmap (mem, &sinfo); return copy; } static void gst_va_allocator_init (GstVaAllocator * self) { GstAllocator *allocator = GST_ALLOCATOR (self); allocator->mem_type = GST_ALLOCATOR_VASURFACE; allocator->mem_map = (GstMemoryMapFunction) _va_map; allocator->mem_unmap = (GstMemoryUnmapFunction) _va_unmap; allocator->mem_share = _va_share; allocator->mem_copy = _va_copy; gst_va_memory_pool_init (&self->pool); self->feat_use_derived = GST_VA_FEATURE_AUTO; GST_OBJECT_FLAG_SET (self, GST_ALLOCATOR_FLAG_CUSTOM_ALLOC); } static gboolean gst_va_memory_release (GstMiniObject * mini_object) { GstMemory *mem = GST_MEMORY_CAST (mini_object); GstVaAllocator *self = GST_VA_ALLOCATOR (mem->allocator); GST_LOG ("releasing %p: surface %#x", mem, gst_va_memory_get_surface (mem)); gst_va_memory_pool_push (&self->pool, mem); /* Keep last in case we are holding on the last allocator ref */ gst_object_unref (mem->allocator); /* don't call mini_object's free */ return FALSE; } /** * gst_va_allocator_alloc: * @allocator: a #GstAllocator * * Allocate a new VASurfaceID backed #GstMemory. * * Returns: a #GstMemory backed with a VASurfaceID; %NULL, otherwise. * * Since: 1.22 */ GstMemory * gst_va_allocator_alloc (GstAllocator * allocator) { GstVaAllocator *self; GstVaMemory *mem; VASurfaceID surface; g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), NULL); self = GST_VA_ALLOCATOR (allocator); if (self->rt_format == 0) { GST_ERROR_OBJECT (self, "Unknown fourcc or chroma format"); return NULL; } if (!va_create_surfaces (self->display, self->rt_format, self->fourcc, GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info), self->usage_hint, NULL, &surface, 1)) return NULL; mem = g_slice_new (GstVaMemory); mem->surface = surface; mem->surface_format = self->surface_format; _reset_mem (mem, allocator, GST_VIDEO_INFO_SIZE (&self->info)); GST_MINI_OBJECT (mem)->dispose = gst_va_memory_release; gst_va_memory_pool_surface_inc (&self->pool); GST_LOG_OBJECT (self, "Created surface %#x [%dx%d]", mem->surface, GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info)); return GST_MEMORY_CAST (mem); } /** * gst_va_allocator_new: * @display: a #GstVaDisplay * @surface_formats: (element-type guint) (transfer full): a #GArray * of valid #GstVideoFormat for surfaces in current VA context. * * Instanciate a new pooled #GstAllocator backed by VASurfaceID. * * Returns: a #GstVaDisplay * * Since: 1.22 */ GstAllocator * gst_va_allocator_new (GstVaDisplay * display, GArray * surface_formats) { GstVaAllocator *self; g_return_val_if_fail (GST_IS_VA_DISPLAY (display), NULL); self = g_object_new (GST_TYPE_VA_ALLOCATOR, NULL); self->display = gst_object_ref (display); self->surface_formats = surface_formats; gst_object_ref_sink (self); return GST_ALLOCATOR (self); } /** * gst_va_allocator_setup_buffer: * @allocator: a #GstAllocator * @buffer: a #GstBuffer * * Populates an empty @buffer with a VASuface backed #GstMemory. * * Returns: %TRUE if @buffer is populated; %FALSE otherwise. * * Since: 1.22 */ gboolean gst_va_allocator_setup_buffer (GstAllocator * allocator, GstBuffer * buffer) { GstMemory *mem = gst_va_allocator_alloc (allocator); if (!mem) return FALSE; gst_buffer_append_memory (buffer, mem); return TRUE; } static VASurfaceID gst_va_allocator_prepare_buffer_unlocked (GstVaAllocator * self, GstBuffer * buffer) { GstMemory *mem; VASurfaceID surface; mem = gst_va_memory_pool_pop (&self->pool); if (!mem) return VA_INVALID_ID; gst_object_ref (mem->allocator); surface = gst_va_memory_get_surface (mem); gst_buffer_append_memory (buffer, mem); GST_LOG ("buffer %p: memory %p - surface %#x", buffer, mem, surface); return surface; } /** * gst_va_allocator_prepare_buffer: * @allocator: a #GstAllocator * @buffer: an empty #GstBuffer * * This method will populate @buffer with pooled VASurfaceID * memories. It doesn't allocate new VASurfacesID. * * Returns: %TRUE if @buffer was populated correctly; %FALSE * otherwise. * * Since: 1.22 */ gboolean gst_va_allocator_prepare_buffer (GstAllocator * allocator, GstBuffer * buffer) { GstVaAllocator *self; VASurfaceID surface; g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), FALSE); self = GST_VA_ALLOCATOR (allocator); GST_VA_MEMORY_POOL_LOCK (&self->pool); surface = gst_va_allocator_prepare_buffer_unlocked (self, buffer); GST_VA_MEMORY_POOL_UNLOCK (&self->pool); return (surface != VA_INVALID_ID); } /** * gst_va_allocator_flush: * @allocator: a #GstAllocator * * Removes all the memories in @allocator's pool. * * Since: 1.22 */ void gst_va_allocator_flush (GstAllocator * allocator) { GstVaAllocator *self; g_return_if_fail (GST_IS_VA_ALLOCATOR (allocator)); self = GST_VA_ALLOCATOR (allocator); gst_va_memory_pool_flush (&self->pool, self->display); } /** * gst_va_allocator_try: * @allocator: a #GstAllocator * * Try to allocate a test buffer in order to verify that the * allocator's configuration is valid. * * Returns: %TRUE if the configuration is valid; %FALSE otherwise. * * Since: 1.22 */ static gboolean gst_va_allocator_try (GstAllocator * allocator) { GstVaAllocator *self; g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), FALSE); self = GST_VA_ALLOCATOR (allocator); self->fourcc = 0; self->rt_format = 0; self->use_derived = FALSE; self->img_format = GST_VIDEO_INFO_FORMAT (&self->info); self->surface_format = gst_va_video_surface_format_from_image_format (self->img_format, self->surface_formats); if (self->surface_format == GST_VIDEO_FORMAT_UNKNOWN) { /* try a surface without fourcc but rt_format only */ self->fourcc = 0; self->rt_format = gst_va_chroma_from_video_format (self->img_format); } else { if (G_LIKELY (!(self->hacks & GST_VA_HACK_SURFACE_NO_FOURCC))) self->fourcc = gst_va_fourcc_from_video_format (self->surface_format); self->rt_format = gst_va_chroma_from_video_format (self->surface_format); } if (self->rt_format == 0) { GST_ERROR_OBJECT (allocator, "Unsupported image format: %s", gst_video_format_to_string (self->img_format)); return FALSE; } if (!_update_image_info (self)) { GST_ERROR_OBJECT (allocator, "Failed to update allocator info"); return FALSE; } GST_INFO_OBJECT (self, "va allocator info, surface format: %s, image format: %s, " "use derived: %s, rt format: 0x%x, fourcc: %" GST_FOURCC_FORMAT, (self->surface_format == GST_VIDEO_FORMAT_UNKNOWN) ? "unknown" : gst_video_format_to_string (self->surface_format), gst_video_format_to_string (self->img_format), self->use_derived ? "true" : "false", self->rt_format, GST_FOURCC_ARGS (self->fourcc)); return TRUE; } /** * gst_va_allocator_set_format: * @allocator: a #GstAllocator * @info: (inout): a #GstVideoInfo * @usage_hint: VA usage hint * @use_derived: a #GstVaFeature * * Sets the configuration defined by @info, @usage_hint and * @use_derived for @allocator, and it tries the configuration, if * @allocator has not allocated memories yet. * * If @allocator has memory allocated already, and frame size and * format in @info are the same as currently configured in @allocator, * the rest of @info parameters are updated internally. * * Returns: %TRUE if the configuration is valid or updated; %FALSE if * configuration is not valid or not updated. * * Since: 1.22 */ gboolean gst_va_allocator_set_format (GstAllocator * allocator, GstVideoInfo * info, guint usage_hint, GstVaFeature use_derived) { GstVaAllocator *self; gboolean ret; g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), FALSE); g_return_val_if_fail (info, FALSE); self = GST_VA_ALLOCATOR (allocator); if (gst_va_memory_pool_surface_count (&self->pool) != 0) { if (GST_VIDEO_INFO_FORMAT (info) == GST_VIDEO_INFO_FORMAT (&self->info) && GST_VIDEO_INFO_WIDTH (info) == GST_VIDEO_INFO_WIDTH (&self->info) && GST_VIDEO_INFO_HEIGHT (info) == GST_VIDEO_INFO_HEIGHT (&self->info) && usage_hint == self->usage_hint && use_derived == self->feat_use_derived) { *info = self->info; /* update callee info (offset & stride) */ return TRUE; } return FALSE; } self->usage_hint = usage_hint; self->feat_use_derived = use_derived; self->info = *info; g_clear_pointer (&self->copy, gst_va_surface_copy_free); ret = gst_va_allocator_try (allocator); if (ret) *info = self->info; return ret; } /** * gst_va_allocator_get_format: * @allocator: a #GstAllocator * @info: (out) (optional): a #GstVideoInfo * @usage_hint: (out) (optional): VA usage hint * @use_derived: (out) (optional): a #GstVaFeature if derived images * are used for buffer mapping. * * Gets current internal configuration of @allocator. * * Returns: %TRUE if @allocator is already configured; %FALSE * otherwise. * * Since: 1.22 */ gboolean gst_va_allocator_get_format (GstAllocator * allocator, GstVideoInfo * info, guint * usage_hint, GstVaFeature * use_derived) { GstVaAllocator *self; g_return_val_if_fail (GST_IS_VA_ALLOCATOR (allocator), FALSE); self = GST_VA_ALLOCATOR (allocator); if (GST_VIDEO_INFO_FORMAT (&self->info) == GST_VIDEO_FORMAT_UNKNOWN) return FALSE; if (info) *info = self->info; if (usage_hint) *usage_hint = self->usage_hint; if (use_derived) *use_derived = self->feat_use_derived; return TRUE; } /** * gst_va_allocator_set_hacks: (skip) * @allocator: a #GstAllocator * @hacks: hacks id to set * * Internal method to set allocator specific logic changes. * * Since: 1.22 */ void gst_va_allocator_set_hacks (GstAllocator * allocator, guint32 hacks) { GstVaAllocator *self; g_return_if_fail (GST_IS_VA_ALLOCATOR (allocator)); self = GST_VA_ALLOCATOR (allocator); self->hacks = hacks; } /** * gst_va_allocator_peek_display: * @allocator: a #GstAllocator * * Returns: (type #GstVaDisplay) (transfer none): the display which this * @allocator belongs to. The reference of the display is unchanged. * * Since: 1.22 */ GstVaDisplay * gst_va_allocator_peek_display (GstAllocator * allocator) { if (!allocator) return NULL; if (GST_IS_VA_DMABUF_ALLOCATOR (allocator)) { return GST_VA_DMABUF_ALLOCATOR (allocator)->display; } else if (GST_IS_VA_ALLOCATOR (allocator)) { return GST_VA_ALLOCATOR (allocator)->display; } return NULL; } /*============ Utilities =====================================================*/ /** * gst_va_memory_get_surface: * @mem: a #GstMemory * * Returns: (type guint) (transfer none): the VASurfaceID in @mem. * * Since: 1.22 */ VASurfaceID gst_va_memory_get_surface (GstMemory * mem) { VASurfaceID surface = VA_INVALID_ID; if (!mem->allocator) return VA_INVALID_ID; if (GST_IS_DMABUF_ALLOCATOR (mem->allocator)) { GstVaBufferSurface *buf; buf = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_surface_quark ()); if (buf) surface = buf->surface; } else if (GST_IS_VA_ALLOCATOR (mem->allocator)) { GstVaMemory *va_mem = (GstVaMemory *) mem; surface = va_mem->surface; } return surface; } /** * gst_va_memory_peek_display: * @mem: a #GstMemory * * Returns: (type #GstVaDisplay) (transfer none): the display which * this @mem belongs to. The reference of the display is unchanged. * * Since: 1.22 */ GstVaDisplay * gst_va_memory_peek_display (GstMemory * mem) { GstAllocator *allocator; if (!mem) return NULL; allocator = GST_MEMORY_CAST (mem)->allocator; /* no allocator, not VA kind memory. */ if (!allocator) return NULL; return gst_va_allocator_peek_display (allocator); } /** * gst_va_buffer_get_surface: * @buffer: a #GstBuffer * * Returns: (type guint) (transfer none): the VASurfaceID in @buffer. * * Since: 1.22 */ VASurfaceID gst_va_buffer_get_surface (GstBuffer * buffer) { GstMemory *mem; mem = gst_buffer_peek_memory (buffer, 0); if (!mem) return VA_INVALID_ID; return gst_va_memory_get_surface (mem); } /** * gst_va_buffer_create_aux_surface: * @buffer: a #GstBuffer * * Creates a new VASurfaceID with @buffer's allocator and attached it * to it. * * *This method is used only by plugin's internal VA decoder.* * * Returns: %TRUE if the new VASurfaceID is attached to @buffer * correctly; %FALSE, otherwise. * * Since: 1.22 */ gboolean gst_va_buffer_create_aux_surface (GstBuffer * buffer) { GstMemory *mem; VASurfaceID surface = VA_INVALID_ID; GstVaDisplay *display = NULL; GstVideoFormat format; gint width, height; GstVaBufferSurface *surface_buffer; mem = gst_buffer_peek_memory (buffer, 0); if (!mem) return FALSE; /* Already created. */ surface_buffer = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_aux_surface_quark ()); if (surface_buffer) return TRUE; if (!mem->allocator) return FALSE; if (GST_IS_VA_DMABUF_ALLOCATOR (mem->allocator)) { GstVaDmabufAllocator *self = GST_VA_DMABUF_ALLOCATOR (mem->allocator); guint32 fourcc, rt_format; format = GST_VIDEO_INFO_FORMAT (&self->info); fourcc = gst_va_fourcc_from_video_format (format); rt_format = gst_va_chroma_from_video_format (format); if (fourcc == 0 || rt_format == 0) { GST_ERROR_OBJECT (self, "Unsupported format: %s", gst_video_format_to_string (GST_VIDEO_INFO_FORMAT (&self->info))); return FALSE; } display = self->display; width = GST_VIDEO_INFO_WIDTH (&self->info); height = GST_VIDEO_INFO_HEIGHT (&self->info); if (!va_create_surfaces (self->display, rt_format, fourcc, GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info), self->usage_hint, NULL, &surface, 1)) return FALSE; } else if (GST_IS_VA_ALLOCATOR (mem->allocator)) { GstVaAllocator *self = GST_VA_ALLOCATOR (mem->allocator); if (self->rt_format == 0) { GST_ERROR_OBJECT (self, "Unknown fourcc or chroma format"); return FALSE; } display = self->display; width = GST_VIDEO_INFO_WIDTH (&self->info); height = GST_VIDEO_INFO_HEIGHT (&self->info); format = GST_VIDEO_INFO_FORMAT (&self->info); if (!va_create_surfaces (self->display, self->rt_format, self->fourcc, GST_VIDEO_INFO_WIDTH (&self->info), GST_VIDEO_INFO_HEIGHT (&self->info), self->usage_hint, NULL, &surface, 1)) return FALSE; } else { g_assert_not_reached (); } if (!display || surface == VA_INVALID_ID) return FALSE; surface_buffer = gst_va_buffer_surface_new (surface, format, width, height); surface_buffer->display = gst_object_ref (display); g_atomic_int_add (&surface_buffer->ref_count, 1); gst_mini_object_set_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_aux_surface_quark (), surface_buffer, gst_va_buffer_surface_unref); return TRUE; } /** * gst_va_buffer_get_aux_surface: * @buffer: a #GstBuffer * * Returns: (type guint) (transfer none): the VASurfaceID attached to * @buffer. * * Since: 1.22 */ VASurfaceID gst_va_buffer_get_aux_surface (GstBuffer * buffer) { GstVaBufferSurface *surface_buffer; GstMemory *mem; mem = gst_buffer_peek_memory (buffer, 0); if (!mem) return VA_INVALID_ID; surface_buffer = gst_mini_object_get_qdata (GST_MINI_OBJECT (mem), gst_va_buffer_aux_surface_quark ()); if (!surface_buffer) return VA_INVALID_ID; /* No one increments it, and its lifetime is the same with the gstmemory itself */ g_assert (g_atomic_int_get (&surface_buffer->ref_count) == 1); return surface_buffer->surface; }