mirror of
https://github.com/Swiftgram/Telegram-iOS.git
synced 2025-12-22 22:25:57 +00:00
Update ffmpeg to 7.1.1
This commit is contained in:
@@ -0,0 +1,386 @@
|
||||
/*
|
||||
* This file is part of FFmpeg.
|
||||
*
|
||||
* FFmpeg is free software; you can redistribute it and/or
|
||||
* modify it under the terms of the GNU Lesser General Public
|
||||
* License as published by the Free Software Foundation; either
|
||||
* version 2.1 of the License, or (at your option) any later version.
|
||||
*
|
||||
* FFmpeg is distributed in the hope that it will be useful,
|
||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||||
* Lesser General Public License for more details.
|
||||
*
|
||||
* You should have received a copy of the GNU Lesser General Public
|
||||
* License along with FFmpeg; if not, write to the Free Software
|
||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
||||
*/
|
||||
|
||||
#include <stdatomic.h>
|
||||
#include <stdint.h>
|
||||
#include <string.h>
|
||||
|
||||
#include "refstruct.h"
|
||||
|
||||
#include "libavutil/avassert.h"
|
||||
#include "libavutil/error.h"
|
||||
#include "libavutil/macros.h"
|
||||
#include "libavutil/mem.h"
|
||||
#include "libavutil/mem_internal.h"
|
||||
#include "libavutil/thread.h"
|
||||
|
||||
#ifndef REFSTRUCT_CHECKED
|
||||
#ifndef ASSERT_LEVEL
|
||||
#define ASSERT_LEVEL 0
|
||||
#endif
|
||||
#define REFSTRUCT_CHECKED (ASSERT_LEVEL >= 1)
|
||||
#endif
|
||||
|
||||
#if REFSTRUCT_CHECKED
|
||||
#define ff_assert(cond) av_assert0(cond)
|
||||
#else
|
||||
#define ff_assert(cond) ((void)0)
|
||||
#endif
|
||||
|
||||
#define REFSTRUCT_COOKIE AV_NE((uint64_t)MKBETAG('R', 'e', 'f', 'S') << 32 | MKBETAG('t', 'r', 'u', 'c'), \
|
||||
MKTAG('R', 'e', 'f', 'S') | (uint64_t)MKTAG('t', 'r', 'u', 'c') << 32)
|
||||
|
||||
#if __STDC_VERSION__ >= 201112L && !defined(_MSC_VER)
|
||||
#define REFCOUNT_OFFSET FFALIGN(sizeof(RefCount), FFMAX(ALIGN_64, _Alignof(max_align_t)))
|
||||
#else
|
||||
#define REFCOUNT_OFFSET FFALIGN(sizeof(RefCount), ALIGN_64)
|
||||
#endif
|
||||
|
||||
typedef struct RefCount {
|
||||
/**
|
||||
* An uintptr_t is big enough to hold the address of every reference,
|
||||
* so no overflow can happen when incrementing the refcount as long as
|
||||
* the user does not throw away references.
|
||||
*/
|
||||
atomic_uintptr_t refcount;
|
||||
FFRefStructOpaque opaque;
|
||||
void (*free_cb)(FFRefStructOpaque opaque, void *obj);
|
||||
void (*free)(void *ref);
|
||||
|
||||
#if REFSTRUCT_CHECKED
|
||||
uint64_t cookie;
|
||||
#endif
|
||||
} RefCount;
|
||||
|
||||
static RefCount *get_refcount(void *obj)
|
||||
{
|
||||
RefCount *ref = (RefCount*)((char*)obj - REFCOUNT_OFFSET);
|
||||
ff_assert(ref->cookie == REFSTRUCT_COOKIE);
|
||||
return ref;
|
||||
}
|
||||
|
||||
static const RefCount *cget_refcount(const void *obj)
|
||||
{
|
||||
const RefCount *ref = (const RefCount*)((const char*)obj - REFCOUNT_OFFSET);
|
||||
ff_assert(ref->cookie == REFSTRUCT_COOKIE);
|
||||
return ref;
|
||||
}
|
||||
|
||||
static void *get_userdata(void *buf)
|
||||
{
|
||||
return (char*)buf + REFCOUNT_OFFSET;
|
||||
}
|
||||
|
||||
static void refcount_init(RefCount *ref, FFRefStructOpaque opaque,
|
||||
void (*free_cb)(FFRefStructOpaque opaque, void *obj))
|
||||
{
|
||||
atomic_init(&ref->refcount, 1);
|
||||
ref->opaque = opaque;
|
||||
ref->free_cb = free_cb;
|
||||
ref->free = av_free;
|
||||
|
||||
#if REFSTRUCT_CHECKED
|
||||
ref->cookie = REFSTRUCT_COOKIE;
|
||||
#endif
|
||||
}
|
||||
|
||||
void *ff_refstruct_alloc_ext_c(size_t size, unsigned flags, FFRefStructOpaque opaque,
|
||||
void (*free_cb)(FFRefStructOpaque opaque, void *obj))
|
||||
{
|
||||
void *buf, *obj;
|
||||
|
||||
if (size > SIZE_MAX - REFCOUNT_OFFSET)
|
||||
return NULL;
|
||||
buf = av_malloc(size + REFCOUNT_OFFSET);
|
||||
if (!buf)
|
||||
return NULL;
|
||||
refcount_init(buf, opaque, free_cb);
|
||||
obj = get_userdata(buf);
|
||||
if (!(flags & FF_REFSTRUCT_FLAG_NO_ZEROING))
|
||||
memset(obj, 0, size);
|
||||
|
||||
return obj;
|
||||
}
|
||||
|
||||
void ff_refstruct_unref(void *objp)
|
||||
{
|
||||
void *obj;
|
||||
RefCount *ref;
|
||||
|
||||
memcpy(&obj, objp, sizeof(obj));
|
||||
if (!obj)
|
||||
return;
|
||||
memcpy(objp, &(void *){ NULL }, sizeof(obj));
|
||||
|
||||
ref = get_refcount(obj);
|
||||
if (atomic_fetch_sub_explicit(&ref->refcount, 1, memory_order_acq_rel) == 1) {
|
||||
if (ref->free_cb)
|
||||
ref->free_cb(ref->opaque, obj);
|
||||
ref->free(ref);
|
||||
}
|
||||
|
||||
return;
|
||||
}
|
||||
|
||||
void *ff_refstruct_ref(void *obj)
|
||||
{
|
||||
RefCount *ref = get_refcount(obj);
|
||||
|
||||
atomic_fetch_add_explicit(&ref->refcount, 1, memory_order_relaxed);
|
||||
|
||||
return obj;
|
||||
}
|
||||
|
||||
const void *ff_refstruct_ref_c(const void *obj)
|
||||
{
|
||||
/* Casting const away here is fine, as it is only supposed
|
||||
* to apply to the user's data and not our bookkeeping data. */
|
||||
RefCount *ref = get_refcount((void*)obj);
|
||||
|
||||
atomic_fetch_add_explicit(&ref->refcount, 1, memory_order_relaxed);
|
||||
|
||||
return obj;
|
||||
}
|
||||
|
||||
void ff_refstruct_replace(void *dstp, const void *src)
|
||||
{
|
||||
const void *dst;
|
||||
memcpy(&dst, dstp, sizeof(dst));
|
||||
|
||||
if (src == dst)
|
||||
return;
|
||||
ff_refstruct_unref(dstp);
|
||||
if (src) {
|
||||
dst = ff_refstruct_ref_c(src);
|
||||
memcpy(dstp, &dst, sizeof(dst));
|
||||
}
|
||||
}
|
||||
|
||||
int ff_refstruct_exclusive(const void *obj)
|
||||
{
|
||||
const RefCount *ref = cget_refcount(obj);
|
||||
/* Casting const away here is safe, because it is a load.
|
||||
* It is necessary because atomic_load_explicit() does not
|
||||
* accept const atomics in C11 (see also N1807). */
|
||||
return atomic_load_explicit((atomic_uintptr_t*)&ref->refcount, memory_order_acquire) == 1;
|
||||
}
|
||||
|
||||
struct FFRefStructPool {
|
||||
size_t size;
|
||||
FFRefStructOpaque opaque;
|
||||
int (*init_cb)(FFRefStructOpaque opaque, void *obj);
|
||||
void (*reset_cb)(FFRefStructOpaque opaque, void *obj);
|
||||
void (*free_entry_cb)(FFRefStructOpaque opaque, void *obj);
|
||||
void (*free_cb)(FFRefStructOpaque opaque);
|
||||
|
||||
int uninited;
|
||||
unsigned entry_flags;
|
||||
unsigned pool_flags;
|
||||
|
||||
/** The number of outstanding entries not in available_entries. */
|
||||
atomic_uintptr_t refcount;
|
||||
/**
|
||||
* This is a linked list of available entries;
|
||||
* the RefCount's opaque pointer is used as next pointer
|
||||
* for available entries.
|
||||
* While the entries are in use, the opaque is a pointer
|
||||
* to the corresponding FFRefStructPool.
|
||||
*/
|
||||
RefCount *available_entries;
|
||||
AVMutex mutex;
|
||||
};
|
||||
|
||||
static void pool_free(FFRefStructPool *pool)
|
||||
{
|
||||
ff_mutex_destroy(&pool->mutex);
|
||||
if (pool->free_cb)
|
||||
pool->free_cb(pool->opaque);
|
||||
av_free(get_refcount(pool));
|
||||
}
|
||||
|
||||
static void pool_free_entry(FFRefStructPool *pool, RefCount *ref)
|
||||
{
|
||||
if (pool->free_entry_cb)
|
||||
pool->free_entry_cb(pool->opaque, get_userdata(ref));
|
||||
av_free(ref);
|
||||
}
|
||||
|
||||
static void pool_return_entry(void *ref_)
|
||||
{
|
||||
RefCount *ref = ref_;
|
||||
FFRefStructPool *pool = ref->opaque.nc;
|
||||
|
||||
ff_mutex_lock(&pool->mutex);
|
||||
if (!pool->uninited) {
|
||||
ref->opaque.nc = pool->available_entries;
|
||||
pool->available_entries = ref;
|
||||
ref = NULL;
|
||||
}
|
||||
ff_mutex_unlock(&pool->mutex);
|
||||
|
||||
if (ref)
|
||||
pool_free_entry(pool, ref);
|
||||
|
||||
if (atomic_fetch_sub_explicit(&pool->refcount, 1, memory_order_acq_rel) == 1)
|
||||
pool_free(pool);
|
||||
}
|
||||
|
||||
static void pool_reset_entry(FFRefStructOpaque opaque, void *entry)
|
||||
{
|
||||
FFRefStructPool *pool = opaque.nc;
|
||||
|
||||
pool->reset_cb(pool->opaque, entry);
|
||||
}
|
||||
|
||||
static int refstruct_pool_get_ext(void *datap, FFRefStructPool *pool)
|
||||
{
|
||||
void *ret = NULL;
|
||||
|
||||
memcpy(datap, &(void *){ NULL }, sizeof(void*));
|
||||
|
||||
ff_mutex_lock(&pool->mutex);
|
||||
ff_assert(!pool->uninited);
|
||||
if (pool->available_entries) {
|
||||
RefCount *ref = pool->available_entries;
|
||||
ret = get_userdata(ref);
|
||||
pool->available_entries = ref->opaque.nc;
|
||||
ref->opaque.nc = pool;
|
||||
atomic_init(&ref->refcount, 1);
|
||||
}
|
||||
ff_mutex_unlock(&pool->mutex);
|
||||
|
||||
if (!ret) {
|
||||
RefCount *ref;
|
||||
ret = ff_refstruct_alloc_ext(pool->size, pool->entry_flags, pool,
|
||||
pool->reset_cb ? pool_reset_entry : NULL);
|
||||
if (!ret)
|
||||
return AVERROR(ENOMEM);
|
||||
ref = get_refcount(ret);
|
||||
ref->free = pool_return_entry;
|
||||
if (pool->init_cb) {
|
||||
int err = pool->init_cb(pool->opaque, ret);
|
||||
if (err < 0) {
|
||||
if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_RESET_ON_INIT_ERROR)
|
||||
pool->reset_cb(pool->opaque, ret);
|
||||
if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_FREE_ON_INIT_ERROR)
|
||||
pool->free_entry_cb(pool->opaque, ret);
|
||||
av_free(ref);
|
||||
return err;
|
||||
}
|
||||
}
|
||||
}
|
||||
atomic_fetch_add_explicit(&pool->refcount, 1, memory_order_relaxed);
|
||||
|
||||
if (pool->pool_flags & FF_REFSTRUCT_POOL_FLAG_ZERO_EVERY_TIME)
|
||||
memset(ret, 0, pool->size);
|
||||
|
||||
memcpy(datap, &ret, sizeof(ret));
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void *ff_refstruct_pool_get(FFRefStructPool *pool)
|
||||
{
|
||||
void *ret;
|
||||
refstruct_pool_get_ext(&ret, pool);
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
* Hint: The content of pool_unref() and refstruct_pool_uninit()
|
||||
* could currently be merged; they are only separate functions
|
||||
* in case we would ever introduce weak references.
|
||||
*/
|
||||
static void pool_unref(void *ref)
|
||||
{
|
||||
FFRefStructPool *pool = get_userdata(ref);
|
||||
if (atomic_fetch_sub_explicit(&pool->refcount, 1, memory_order_acq_rel) == 1)
|
||||
pool_free(pool);
|
||||
}
|
||||
|
||||
static void refstruct_pool_uninit(FFRefStructOpaque unused, void *obj)
|
||||
{
|
||||
FFRefStructPool *pool = obj;
|
||||
RefCount *entry;
|
||||
|
||||
ff_mutex_lock(&pool->mutex);
|
||||
ff_assert(!pool->uninited);
|
||||
pool->uninited = 1;
|
||||
entry = pool->available_entries;
|
||||
pool->available_entries = NULL;
|
||||
ff_mutex_unlock(&pool->mutex);
|
||||
|
||||
while (entry) {
|
||||
void *next = entry->opaque.nc;
|
||||
pool_free_entry(pool, entry);
|
||||
entry = next;
|
||||
}
|
||||
}
|
||||
|
||||
FFRefStructPool *ff_refstruct_pool_alloc(size_t size, unsigned flags)
|
||||
{
|
||||
return ff_refstruct_pool_alloc_ext(size, flags, NULL, NULL, NULL, NULL, NULL);
|
||||
}
|
||||
|
||||
FFRefStructPool *ff_refstruct_pool_alloc_ext_c(size_t size, unsigned flags,
|
||||
FFRefStructOpaque opaque,
|
||||
int (*init_cb)(FFRefStructOpaque opaque, void *obj),
|
||||
void (*reset_cb)(FFRefStructOpaque opaque, void *obj),
|
||||
void (*free_entry_cb)(FFRefStructOpaque opaque, void *obj),
|
||||
void (*free_cb)(FFRefStructOpaque opaque))
|
||||
{
|
||||
FFRefStructPool *pool = ff_refstruct_alloc_ext(sizeof(*pool), 0, NULL,
|
||||
refstruct_pool_uninit);
|
||||
int err;
|
||||
|
||||
if (!pool)
|
||||
return NULL;
|
||||
get_refcount(pool)->free = pool_unref;
|
||||
|
||||
pool->size = size;
|
||||
pool->opaque = opaque;
|
||||
pool->init_cb = init_cb;
|
||||
pool->reset_cb = reset_cb;
|
||||
pool->free_entry_cb = free_entry_cb;
|
||||
pool->free_cb = free_cb;
|
||||
#define COMMON_FLAGS FF_REFSTRUCT_POOL_FLAG_NO_ZEROING
|
||||
pool->entry_flags = flags & COMMON_FLAGS;
|
||||
// Filter out nonsense combinations to avoid checks later.
|
||||
if (!pool->reset_cb)
|
||||
flags &= ~FF_REFSTRUCT_POOL_FLAG_RESET_ON_INIT_ERROR;
|
||||
if (!pool->free_entry_cb)
|
||||
flags &= ~FF_REFSTRUCT_POOL_FLAG_FREE_ON_INIT_ERROR;
|
||||
pool->pool_flags = flags;
|
||||
|
||||
if (flags & FF_REFSTRUCT_POOL_FLAG_ZERO_EVERY_TIME) {
|
||||
// We will zero the buffer before every use, so zeroing
|
||||
// upon allocating the buffer is unnecessary.
|
||||
pool->entry_flags |= FF_REFSTRUCT_FLAG_NO_ZEROING;
|
||||
}
|
||||
|
||||
atomic_init(&pool->refcount, 1);
|
||||
|
||||
err = ff_mutex_init(&pool->mutex, NULL);
|
||||
if (err) {
|
||||
// Don't call ff_refstruct_uninit() on pool, as it hasn't been properly
|
||||
// set up and is just a POD right now.
|
||||
av_free(get_refcount(pool));
|
||||
return NULL;
|
||||
}
|
||||
return pool;
|
||||
}
|
||||
Reference in New Issue
Block a user