/* * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include #include #include #include "refstruct.h" #include "libavutil/avassert.h" #include "libavutil/error.h" #include "libavutil/macros.h" #include "libavutil/mem.h" #include "libavutil/mem_internal.h" #include "libavutil/thread.h" #ifndef REFSTRUCT_CHECKED #ifndef ASSERT_LEVEL #define ASSERT_LEVEL 0 #endif #define REFSTRUCT_CHECKED (ASSERT_LEVEL >= 1) #endif #if REFSTRUCT_CHECKED #define ff_assert(cond) av_assert0(cond) #else #define ff_assert(cond) ((void)0) #endif #define REFSTRUCT_COOKIE AV_NE((uint64_t)MKBETAG('R', 'e', 'f', 'S') << 32 | MKBETAG('t', 'r', 'u', 'c'), \ MKTAG('R', 'e', 'f', 'S') | (uint64_t)MKTAG('t', 'r', 'u', 'c') << 32) #if __STDC_VERSION__ >= 201112L && !defined(_MSC_VER) #define REFCOUNT_OFFSET FFALIGN(sizeof(RefCount), FFMAX(ALIGN_64, _Alignof(max_align_t))) #else #define REFCOUNT_OFFSET FFALIGN(sizeof(RefCount), ALIGN_64) #endif typedef struct RefCount { /** * An uintptr_t is big enough to hold the address of every reference, * so no overflow can happen when incrementing the refcount as long as * the user does not throw away references. */ atomic_uintptr_t refcount; FFRefStructOpaque opaque; void (*free_cb)(FFRefStructOpaque opaque, void *obj); void (*free)(void *ref); #if REFSTRUCT_CHECKED uint64_t cookie; #endif } RefCount; static RefCount *get_refcount(void *obj) { RefCount *ref = (RefCount*)((char*)obj - REFCOUNT_OFFSET); ff_assert(ref->cookie == REFSTRUCT_COOKIE); return ref; } static const RefCount *cget_refcount(const void *obj) { const RefCount *ref = (const RefCount*)((const char*)obj - REFCOUNT_OFFSET); ff_assert(ref->cookie == REFSTRUCT_COOKIE); return ref; } static void *get_userdata(void *buf) { return (char*)buf + REFCOUNT_OFFSET; } static void refcount_init(RefCount *ref, FFRefStructOpaque opaque, void (*free_cb)(FFRefStructOpaque opaque, void *obj)) { atomic_init(&ref->refcount, 1); ref->opaque = opaque; ref->free_cb = free_cb; ref->free = av_free; #if REFSTRUCT_CHECKED ref->cookie = REFSTRUCT_COOKIE; #endif } void *ff_refstruct_alloc_ext_c(size_t size, unsigned flags, FFRefStructOpaque opaque, void (*free_cb)(FFRefStructOpaque opaque, void *obj)) { void *buf, *obj; if (size > SIZE_MAX - REFCOUNT_OFFSET) return NULL; buf = av_malloc(size + REFCOUNT_OFFSET); if (!buf) return NULL; refcount_init(buf, opaque, free_cb); obj = get_userdata(buf); if (!(flags & FF_REFSTRUCT_FLAG_NO_ZEROING)) memset(obj, 0, size); return obj; } void ff_refstruct_unref(void *objp) { void *obj; RefCount *ref; memcpy(&obj, objp, sizeof(obj)); if (!obj) return; memcpy(objp, &(void *){ NULL }, sizeof(obj)); ref = get_refcount(obj); if (atomic_fetch_sub_explicit(&ref->refcount, 1, memory_order_acq_rel) == 1) { if (ref->free_cb) ref->free_cb(ref->opaque, obj); ref->free(ref); } return; } void *ff_refstruct_ref(void *obj) { RefCount *ref = get_refcount(obj); atomic_fetch_add_explicit(&ref->refcount, 1, memory_order_relaxed); return obj; } const void *ff_refstruct_ref_c(const void *obj) { /* Casting const away here is fine, as it is only supposed * to apply to the user's data and not our bookkeeping data. */ RefCount *ref = get_refcount((void*)obj); atomic_fetch_add_explicit(&ref->refcount, 1, memory_order_relaxed); return obj; } void ff_refstruct_replace(void *dstp, const void *src) { const void *dst; memcpy(&dst, dstp, sizeof(dst)); if (src == dst) return; ff_refstruct_unref(dstp); if (src) { dst = ff_refstruct_ref_c(src); memcpy(dstp, &dst, sizeof(dst)); } } int ff_refstruct_exclusive(const void *obj) { const RefCount *ref = cget_refcount(obj); /* Casting const away here is safe, because it is a load. * It is necessary because atomic_load_explicit() does not * accept const atomics in C11 (see also N1807). */ return atomic_load_explicit((atomic_uintptr_t*)&ref->refcount, memory_order_acquire) == 1; } struct FFRefStructPool { size_t size; FFRefStructOpaque opaque; int (*init_cb)(FFRefStructOpaque opaque, void *obj); void (*reset_cb)(FFRefStructOpaque opaque, void *obj); void (*free_entry_cb)(FFRefStructOpaque opaque, void *obj); void (*free_cb)(FFRefStructOpaque opaque); int uninited; unsigned entry_flags; unsigned pool_flags; /** The number of outstanding entries not in available_entries. */ atomic_uintptr_t refcount; /** * This is a linked list of available entries; * the RefCount's opaque pointer is used as next pointer * for available entries. * While the entries are in use, the opaque is a pointer * to the corresponding FFRefStructPool. */ RefCount *available_entries; AVMutex mutex; }; static void pool_free(FFRefStructPool *pool) { ff_mutex_destroy(&pool->mutex); if (pool->free_cb) pool->free_cb(pool->opaque); av_free(get_refcount(pool)); } static void pool_free_entry(FFRefStructPool *pool, RefCount *ref) { if (pool->free_entry_cb) pool->free_entry_cb(pool->opaque, get_userdata(ref)); av_free(ref); } static void pool_return_entry(void *ref_) { RefCount *ref = ref_; FFRefStructPool *pool = ref->opaque.nc; ff_mutex_lock(&pool->mutex); if (!pool->uninited) { ref->opaque.nc = pool->available_entries; pool->available_entries = ref; ref = NULL; } ff_mutex_unlock(&pool->mutex); if (ref) pool_free_entry(pool, ref); if (atomic_fetch_sub_explicit(&pool->refcount, 1, memory_order_acq_rel) == 1) pool_free(pool); } static void pool_reset_entry(FFRefStructOpaque opaque, void *entry) { FFRefStructPool *pool = opaque.nc; pool->reset_cb(pool->opaque, entry); } static int refstruct_pool_get_ext(void *datap, FFRefStructPool *pool) { void *ret = NULL; memcpy(datap, &(void *){ NULL }, sizeof(void*)); ff_mutex_lock(&pool->mutex); ff_assert(!pool->uninited); if (pool->available_entries) { RefCount *ref = pool->available_entries; ret = get_userdata(ref); pool->available_entries = ref->opaque.nc; ref->opaque.nc = pool; atomic_init(&ref->refcount, 1); } ff_mutex_unlock(&pool->mutex); if (!ret) { RefCount *ref; ret = ff_refstruct_alloc_ext(pool->size, pool->entry_flags, pool, pool->reset_cb ? pool_reset_entry : NULL); if (!ret) return AVERROR(ENOMEM); ref = get_refcount(ret); ref->free = pool_return_entry; if (pool->init_cb) { int err = pool->init_cb(pool->opaque, ret); if (err < 0) { if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_RESET_ON_INIT_ERROR) pool->reset_cb(pool->opaque, ret); if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_FREE_ON_INIT_ERROR) pool->free_entry_cb(pool->opaque, ret); av_free(ref); return err; } } } atomic_fetch_add_explicit(&pool->refcount, 1, memory_order_relaxed); if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_ZERO_EVERY_TIME) memset(ret, 0, pool->size); memcpy(datap, &ret, sizeof(ret)); return 0; } void *ff_refstruct_pool_get(FFRefStructPool *pool) { void *ret; refstruct_pool_get_ext(&ret, pool); return ret; } /** * Hint: The content of pool_unref() and refstruct_pool_uninit() * could currently be merged; they are only separate functions * in case we would ever introduce weak references. */ static void pool_unref(void *ref) { FFRefStructPool *pool = get_userdata(ref); if (atomic_fetch_sub_explicit(&pool->refcount, 1, memory_order_acq_rel) == 1) pool_free(pool); } static void refstruct_pool_uninit(FFRefStructOpaque unused, void *obj) { FFRefStructPool *pool = obj; RefCount *entry; ff_mutex_lock(&pool->mutex); ff_assert(!pool->uninited); pool->uninited = 1; entry = pool->available_entries; pool->available_entries = NULL; ff_mutex_unlock(&pool->mutex); while (entry) { void *next = entry->opaque.nc; pool_free_entry(pool, entry); entry = next; } } FFRefStructPool *ff_refstruct_pool_alloc(size_t size, unsigned flags) { return ff_refstruct_pool_alloc_ext(size, flags, NULL, NULL, NULL, NULL, NULL); } FFRefStructPool *ff_refstruct_pool_alloc_ext_c(size_t size, unsigned flags, FFRefStructOpaque opaque, int (*init_cb)(FFRefStructOpaque opaque, void *obj), void (*reset_cb)(FFRefStructOpaque opaque, void *obj), void (*free_entry_cb)(FFRefStructOpaque opaque, void *obj), void (*free_cb)(FFRefStructOpaque opaque)) { FFRefStructPool *pool = ff_refstruct_alloc_ext(sizeof(*pool), 0, NULL, refstruct_pool_uninit); int err; if (!pool) return NULL; get_refcount(pool)->free = pool_unref; pool->size = size; pool->opaque = opaque; pool->init_cb = init_cb; pool->reset_cb = reset_cb; pool->free_entry_cb = free_entry_cb; pool->free_cb = free_cb; #define COMMON_FLAGS FF_REFSTRUCT_POOL_FLAG_NO_ZEROING pool->entry_flags = flags & COMMON_FLAGS; // Filter out nonsense combinations to avoid checks later. if (!pool->reset_cb) flags &= ~FF_REFSTRUCT_POOL_FLAG_RESET_ON_INIT_ERROR; if (!pool->free_entry_cb) flags &= ~FF_REFSTRUCT_POOL_FLAG_FREE_ON_INIT_ERROR; pool->pool_flags = flags; if (flags & FF_REFSTRUCT_POOL_FLAG_ZERO_EVERY_TIME) { // We will zero the buffer before every use, so zeroing // upon allocating the buffer is unnecessary. pool->entry_flags |= FF_REFSTRUCT_FLAG_NO_ZEROING; } atomic_init(&pool->refcount, 1); err = ff_mutex_init(&pool->mutex, NULL); if (err) { // Don't call ff_refstruct_uninit() on pool, as it hasn't been properly // set up and is just a POD right now. av_free(get_refcount(pool)); return NULL; } return pool; }