Merge pull request #1565 from mmichal10/percpu-refcnt

Percpu refcnt
This commit is contained in:
Robert Baldyga 2025-02-27 15:14:22 +01:00 committed by GitHub
commit 30d0cd0df0
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
4 changed files with 456 additions and 2 deletions

View File

@ -1,6 +1,6 @@
/*
* Copyright(c) 2012-2022 Intel Corporation
* Copyright(c) 2024 Huawei Technologies
* Copyright(c) 2024-2025 Huawei Technologies
* SPDX-License-Identifier: BSD-3-Clause
*/
@ -407,6 +407,11 @@ static inline u64 env_atomic64_inc_return(env_atomic64 *a)
return atomic64_inc_return(a);
}
static inline u64 env_atomic64_dec_return(env_atomic64 *a)
{
return atomic64_dec_return(a);
}
static inline u64 env_atomic64_cmpxchg(atomic64_t *a, u64 old, u64 new)
{
return atomic64_cmpxchg(a, old, new);

View File

@ -0,0 +1,345 @@
/*
* Copyright(c) 2019-2021 Intel Corporation
* Copyright(c) 2023-2025 Huawei Technologies Co., Ltd.
* SPDX-License-Identifier: BSD-3-Clause
*/
#include "ocf_env_refcnt.h"
#include "ocf/ocf_err.h"
#include "ocf_env.h"
#define ENV_REFCNT_CB_ARMING 1
#define ENV_REFCNT_CB_ARMED 2
static void _env_refcnt_do_on_cpus_cb(struct work_struct *work)
{
struct notify_cpu_work *ctx =
container_of(work, struct notify_cpu_work, work);
ctx->cb(ctx->priv);
env_atomic_dec(&ctx->rc->notify.to_notify);
wake_up(&ctx->rc->notify.notify_wait_queue);
}
static void _env_refcnt_do_on_cpus(struct env_refcnt *rc,
env_refcnt_do_on_cpu_cb_t cb, void *priv)
{
int cpu_no;
struct notify_cpu_work *work;
ENV_BUG_ON(env_atomic_read(&rc->notify.to_notify));
for_each_online_cpu(cpu_no) {
work = rc->notify.notify_work_items[cpu_no];
env_atomic_inc(&rc->notify.to_notify);
work->cb = cb;
work->rc = rc;
work->priv = priv;
INIT_WORK(&work->work, _env_refcnt_do_on_cpus_cb);
queue_work_on(cpu_no, rc->notify.notify_work_queue,
&work->work);
}
wait_event(rc->notify.notify_wait_queue,
!env_atomic_read(&rc->notify.to_notify));
}
static void _env_refcnt_init_pcpu(void *ctx)
{
struct env_refcnt *rc = ctx;
struct env_refcnt_pcpu *pcpu = this_cpu_ptr(rc->pcpu);
pcpu->freeze = false;
env_atomic64_set(&pcpu->counter, 0);
}
int env_refcnt_init(struct env_refcnt *rc, const char *name, size_t name_len)
{
int cpu_no, result;
env_memset(rc, sizeof(*rc), 0);
env_strncpy(rc->name, sizeof(rc->name), name, name_len);
rc->pcpu = alloc_percpu(struct env_refcnt_pcpu);
if (!rc->pcpu)
return -OCF_ERR_NO_MEM;
init_waitqueue_head(&rc->notify.notify_wait_queue);
rc->notify.notify_work_queue = alloc_workqueue("refcnt_%s", 0,
0, rc->name);
if (!rc->notify.notify_work_queue) {
result = -OCF_ERR_NO_MEM;
goto cleanup_pcpu;
}
rc->notify.notify_work_items = env_vzalloc(
sizeof(*rc->notify.notify_work_items) * num_online_cpus());
if (!rc->notify.notify_work_items) {
result = -OCF_ERR_NO_MEM;
goto cleanup_wq;
}
for_each_online_cpu(cpu_no) {
rc->notify.notify_work_items[cpu_no] = env_vmalloc(
sizeof(*rc->notify.notify_work_items[cpu_no]));
if (!rc->notify.notify_work_items[cpu_no]) {
result = -OCF_ERR_NO_MEM;
goto cleanup_work;
}
}
result = env_spinlock_init(&rc->freeze.lock);
if (result)
goto cleanup_work;
_env_refcnt_do_on_cpus(rc, _env_refcnt_init_pcpu, rc);
rc->callback.pfn = NULL;
rc->callback.priv = NULL;
return 0;
cleanup_work:
for_each_online_cpu(cpu_no) {
if (rc->notify.notify_work_items[cpu_no]) {
env_vfree(rc->notify.notify_work_items[cpu_no]);
rc->notify.notify_work_items[cpu_no] = NULL;
}
}
env_vfree(rc->notify.notify_work_items);
rc->notify.notify_work_items = NULL;
cleanup_wq:
destroy_workqueue(rc->notify.notify_work_queue);
rc->notify.notify_work_queue = NULL;
cleanup_pcpu:
free_percpu(rc->pcpu);
rc->pcpu = NULL;
return result;
}
void env_refcnt_deinit(struct env_refcnt *rc)
{
int cpu_no;
env_spinlock_destroy(&rc->freeze.lock);
ENV_BUG_ON(env_atomic_read(&rc->notify.to_notify));
for_each_online_cpu(cpu_no) {
if (rc->notify.notify_work_items[cpu_no]) {
env_vfree(rc->notify.notify_work_items[cpu_no]);
rc->notify.notify_work_items[cpu_no] = NULL;
}
}
env_vfree(rc->notify.notify_work_items);
rc->notify.notify_work_items = NULL;
destroy_workqueue(rc->notify.notify_work_queue);
rc->notify.notify_work_queue = NULL;
free_percpu(rc->pcpu);
rc->pcpu = NULL;
}
static inline void _env_refcnt_call_freeze_cb(struct env_refcnt *rc)
{
bool fire;
fire = (env_atomic_cmpxchg(&rc->callback.armed, ENV_REFCNT_CB_ARMED, 0)
== ENV_REFCNT_CB_ARMED);
smp_mb();
if (fire)
rc->callback.pfn(rc->callback.priv);
}
void env_refcnt_dec(struct env_refcnt *rc)
{
struct env_refcnt_pcpu *pcpu;
bool freeze;
int64_t countdown = 0;
bool callback;
unsigned long flags;
pcpu = get_cpu_ptr(rc->pcpu);
freeze = pcpu->freeze;
if (!freeze)
env_atomic64_dec(&pcpu->counter);
put_cpu_ptr(pcpu);
if (freeze) {
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
countdown = env_atomic64_dec_return(&rc->freeze.countdown);
callback = !rc->freeze.initializing && countdown == 0;
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
if (callback)
_env_refcnt_call_freeze_cb(rc);
}
}
bool env_refcnt_inc(struct env_refcnt *rc)
{
struct env_refcnt_pcpu *pcpu;
bool freeze;
pcpu = get_cpu_ptr(rc->pcpu);
freeze = pcpu->freeze;
if (!freeze)
env_atomic64_inc(&pcpu->counter);
put_cpu_ptr(pcpu);
return !freeze;
}
struct env_refcnt_freeze_ctx {
struct env_refcnt *rc;
env_atomic64 sum;
};
static void _env_refcnt_freeze_pcpu(void *_ctx)
{
struct env_refcnt_freeze_ctx *ctx = _ctx;
struct env_refcnt_pcpu *pcpu = this_cpu_ptr(ctx->rc->pcpu);
pcpu->freeze = true;
env_atomic64_add(env_atomic64_read(&pcpu->counter), &ctx->sum);
}
void env_refcnt_freeze(struct env_refcnt *rc)
{
struct env_refcnt_freeze_ctx ctx;
int freeze_cnt;
bool callback;
unsigned long flags;
ctx.rc = rc;
env_atomic64_set(&ctx.sum, 0);
/* initiate freeze */
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
freeze_cnt = ++(rc->freeze.counter);
if (freeze_cnt > 1) {
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
return;
}
rc->freeze.initializing = true;
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
/* notify CPUs about freeze */
_env_refcnt_do_on_cpus(rc, _env_refcnt_freeze_pcpu, &ctx);
/* update countdown */
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
env_atomic64_add(env_atomic64_read(&ctx.sum), &rc->freeze.countdown);
rc->freeze.initializing = false;
callback = (env_atomic64_read(&rc->freeze.countdown) == 0);
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
/* if countdown finished trigger callback */
if (callback)
_env_refcnt_call_freeze_cb(rc);
}
void env_refcnt_register_zero_cb(struct env_refcnt *rc, env_refcnt_cb_t cb,
void *priv)
{
bool callback;
bool concurrent_arming;
unsigned long flags;
concurrent_arming = (env_atomic_inc_return(&rc->callback.armed)
> ENV_REFCNT_CB_ARMING);
ENV_BUG_ON(concurrent_arming);
/* arm callback */
rc->callback.pfn = cb;
rc->callback.priv = priv;
smp_wmb();
env_atomic_set(&rc->callback.armed, ENV_REFCNT_CB_ARMED);
/* fire callback in case countdown finished */
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
callback = (
env_atomic64_read(&rc->freeze.countdown) == 0 &&
!rc->freeze.initializing
);
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
if (callback)
_env_refcnt_call_freeze_cb(rc);
}
static void _env_refcnt_unfreeze_pcpu(void *_ctx)
{
struct env_refcnt_freeze_ctx *ctx = _ctx;
struct env_refcnt_pcpu *pcpu = this_cpu_ptr(ctx->rc->pcpu);
ENV_BUG_ON(!pcpu->freeze);
env_atomic64_set(&pcpu->counter, 0);
pcpu->freeze = false;
}
void env_refcnt_unfreeze(struct env_refcnt *rc)
{
struct env_refcnt_freeze_ctx ctx;
int freeze_cnt;
unsigned long flags;
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
freeze_cnt = --(rc->freeze.counter);
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
ENV_BUG_ON(freeze_cnt < 0);
if (freeze_cnt > 0)
return;
ENV_BUG_ON(env_atomic64_read(&rc->freeze.countdown));
/* disarm callback */
env_atomic_set(&rc->callback.armed, 0);
smp_wmb();
/* notify CPUs about unfreeze */
ctx.rc = rc;
_env_refcnt_do_on_cpus(rc, _env_refcnt_unfreeze_pcpu, &ctx);
}
bool env_refcnt_frozen(struct env_refcnt *rc)
{
bool frozen;
unsigned long flags;
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
frozen = !!rc->freeze.counter;
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
return frozen;
}
bool env_refcnt_zeroed(struct env_refcnt *rc)
{
bool frozen;
bool initializing;
int64_t countdown;
unsigned long flags;
env_spinlock_lock_irqsave(&rc->freeze.lock, flags);
frozen = !!rc->freeze.counter;
initializing = rc->freeze.initializing;
countdown = env_atomic64_read(&rc->freeze.countdown);
env_spinlock_unlock_irqrestore(&rc->freeze.lock, flags);
return frozen && !initializing && countdown == 0;
}

View File

@ -0,0 +1,104 @@
/*
* Copyright(c) 2019-2021 Intel Corporation
* Copyright(c) 2023-2025 Huawei Technologies Co., Ltd.
* SPDX-License-Identifier: BSD-3-Clause
*/
#ifndef __OCF_ENV_REFCNT_H__
#define __OCF_ENV_REFCNT_H__
#include "ocf_env.h"
typedef void (*env_refcnt_cb_t)(void *priv);
struct env_refcnt_pcpu {
env_atomic64 counter;
bool freeze;
};
typedef void (*env_refcnt_do_on_cpu_cb_t)(void *priv);
struct notify_cpu_work {
struct work_struct work;
/* function to call on each cpu */
env_refcnt_do_on_cpu_cb_t cb;
/* priv passed to cb */
void *priv;
/* refcnt instance */
struct env_refcnt *rc;
};
struct env_refcnt {
struct env_refcnt_pcpu __percpu *pcpu __aligned(64);
struct {
/* freeze counter */
int counter;
/* global counter used instead of per-CPU ones after
* freeze
*/
env_atomic64 countdown;
/* freeze initializing - freeze was requested but not all
* CPUs were notified.
*/
bool initializing;
env_spinlock lock;
} freeze;
struct {
struct notify_cpu_work **notify_work_items;
env_atomic to_notify;
wait_queue_head_t notify_wait_queue;
struct workqueue_struct *notify_work_queue;
} notify;
struct {
env_atomic armed;
env_refcnt_cb_t pfn;
void *priv;
} callback;
char name[32];
};
/* Initialize reference counter */
int env_refcnt_init(struct env_refcnt *rc, const char *name, size_t name_len);
void env_refcnt_deinit(struct env_refcnt *rc);
/* Try to increment counter. Returns counter value (> 0) if successful, 0
* if counter is frozen
*/
bool env_refcnt_inc(struct env_refcnt *rc);
/* Decrement reference counter */
void env_refcnt_dec(struct env_refcnt *rc);
/* Disallow incrementing of underlying counter - attempts to increment counter
* will be failing until env_refcnt_unfreeze is called.
* It's ok to call freeze multiple times, in which case counter is frozen
* until all freeze calls are offset by a corresponding unfreeze.
*/
void env_refcnt_freeze(struct env_refcnt *rc);
/* Cancel the effect of single env_refcnt_freeze call */
void env_refcnt_unfreeze(struct env_refcnt *rc);
bool env_refcnt_frozen(struct env_refcnt *rc);
bool env_refcnt_zeroed(struct env_refcnt *rc);
/* Register callback to be called when reference counter drops to 0.
* Must be called after counter is frozen.
* Cannot be called until previously regsitered callback had fired.
*/
void env_refcnt_register_zero_cb(struct env_refcnt *rc, env_refcnt_cb_t cb,
void *priv);
#endif // __OCF_ENV_REFCNT_H__

2
ocf

@ -1 +1 @@
Subproject commit 6ad1007e6fa47ff6468eaba0b304d77334c34d4a
Subproject commit a12496320c1852b71ae91b3688e6f1d4697c7fd4