open-cas-linux/modules/cas_cache/exp_obj.c
Tamar Levovitz a202d8c9f5 Add support for Linux kernel up to 6.5
Signed-off-by: Tamar Levovitz <tamar.levovitz@huawei.com>
Signed-off-by: Robert Baldyga <robert.baldyga@huawei.com>
2024-05-14 08:47:21 +02:00

621 lines
13 KiB
C

/*
* Copyright(c) 2012-2022 Intel Corporation
* Copyright(c) 2024 Huawei Technologies
* SPDX-License-Identifier: BSD-3-Clause
*/
#include <linux/module.h>
#include <linux/blkdev.h>
#include <linux/slab.h>
#include <linux/string.h>
#include <linux/blkpg.h>
#include <linux/blk-mq.h>
#include "disk.h"
#include "exp_obj.h"
#include "linux_kernel_version.h"
#include "cas_cache.h"
#include "debug.h"
#define CAS_DEV_MINORS 16
#define KMEM_CACHE_MIN_SIZE sizeof(void *)
static inline int bd_claim_by_disk(struct block_device *bdev, void *holder,
struct gendisk *disk)
{
return bd_link_disk_holder(bdev, disk);
}
static inline void bd_release_from_disk(struct block_device *bdev,
struct gendisk *disk)
{
return bd_unlink_disk_holder(bdev, disk);
}
#if LINUX_VERSION_CODE >= KERNEL_VERSION(5, 18, 0)
#define KRETURN(x) return
#define MAKE_RQ_RET_TYPE void
#elif LINUX_VERSION_CODE >= KERNEL_VERSION(4, 3, 0)
#define KRETURN(x) ({ return (x); })
#define MAKE_RQ_RET_TYPE blk_qc_t
#elif LINUX_VERSION_CODE >= KERNEL_VERSION(3, 2, 0)
#define KRETURN(x) return
#define MAKE_RQ_RET_TYPE void
#else
#define KRETURN(x) ({ return (x); })
#define MAKE_RQ_RET_TYPE int
#endif
int __init cas_init_exp_objs(void)
{
CAS_DEBUG_TRACE();
cas_module.disk_major = register_blkdev(cas_module.disk_major,
"cas");
if (cas_module.disk_major <= 0) {
CAS_DEBUG_ERROR("Cannot allocate major number");
return -EINVAL;
}
CAS_DEBUG_PARAM("Allocated major number: %d", cas_module.disk_major);
cas_module.exp_obj_cache = kmem_cache_create("cas_exp_obj",
sizeof(struct cas_exp_obj), 0, 0, NULL);
if (!cas_module.exp_obj_cache) {
unregister_blkdev(cas_module.disk_major, "cas");
return -ENOMEM;
}
return 0;
}
void cas_deinit_exp_objs(void)
{
CAS_DEBUG_TRACE();
kmem_cache_destroy(cas_module.exp_obj_cache);
unregister_blkdev(cas_module.disk_major, "cas");
}
static MAKE_RQ_RET_TYPE _cas_exp_obj_submit_bio(struct bio *bio)
{
struct cas_disk *dsk;
struct cas_exp_obj *exp_obj;
BUG_ON(!bio);
dsk = CAS_BIO_GET_GENDISK(bio)->private_data;
exp_obj = dsk->exp_obj;
exp_obj->ops->submit_bio(dsk, bio, exp_obj->private);
KRETURN(0);
}
static MAKE_RQ_RET_TYPE _cas_exp_obj_make_rq_fn(struct request_queue *q,
struct bio *bio)
{
_cas_exp_obj_submit_bio(bio);
cas_blk_queue_exit(q);
KRETURN(0);
}
static int _cas_del_partitions(struct cas_disk *dsk)
{
struct block_device *bd = cas_disk_get_blkdev(dsk);
struct file *bd_file;
unsigned long __user usr_bpart;
unsigned long __user usr_barg;
struct blkpg_partition bpart;
struct blkpg_ioctl_arg barg;
int result = 0;
int part_no;
bd_file = filp_open(dsk->path, 0, 0);
if (IS_ERR(bd_file))
return PTR_ERR(bd_file);
usr_bpart = cas_vm_mmap(NULL, 0, sizeof(bpart));
if (IS_ERR((void *)usr_bpart)) {
result = PTR_ERR((void *)usr_bpart);
goto out_map_bpart;
}
usr_barg = cas_vm_mmap(NULL, 0, sizeof(barg));
if (IS_ERR((void *)usr_barg)) {
result = PTR_ERR((void *)usr_barg);
goto out_map_barg;
}
memset(&bpart, 0, sizeof(bpart));
memset(&barg, 0, sizeof(barg));
barg.data = (void __user *)usr_bpart;
barg.op = BLKPG_DEL_PARTITION;
result = copy_to_user((void __user *)usr_barg, &barg, sizeof(barg));
if (result) {
result = -EINVAL;
goto out_copy;
}
while ((part_no = cas_bd_get_next_part(bd))) {
bpart.pno = part_no;
result = copy_to_user((void __user *)usr_bpart, &bpart,
sizeof(bpart));
if (result) {
result = -EINVAL;
break;
}
result = cas_vfs_ioctl(bd_file, BLKPG, usr_barg);
if (result == 0) {
printk(KERN_INFO "Partition %d on %s hidden\n",
part_no, bd->bd_disk->disk_name);
} else {
printk(KERN_ERR "Error(%d) hiding the partition %d on %s\n",
result, part_no, bd->bd_disk->disk_name);
break;
}
}
out_copy:
cas_vm_munmap(usr_barg, sizeof(barg));
out_map_barg:
cas_vm_munmap(usr_bpart, sizeof(bpart));
out_map_bpart:
filp_close(bd_file, NULL);
return result;
}
static int _cas_exp_obj_hide_parts(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj = dsk->exp_obj;
struct block_device *bd = cas_disk_get_blkdev(dsk);
struct gendisk *gdsk = cas_disk_get_gendisk(dsk);
if (bd != cas_bdev_whole(bd))
/* It is partition, no more job required */
return 0;
if (GET_DISK_MAX_PARTS(dsk->bd->bd_disk) > 1) {
if (_cas_del_partitions(dsk)) {
printk(KERN_ERR "Error deleting a partition on thedevice %s\n",
gdsk->disk_name);
/* Try restore previous partitions by rescaning */
cas_reread_partitions(bd);
return -EINVAL;
}
}
/* Save original flags and minors */
exp_obj->gd_flags = gdsk->flags & _CAS_GENHD_FLAGS;
exp_obj->gd_minors = gdsk->minors;
/* Setup disk of bottom device as not partitioned device */
gdsk->flags &= ~_CAS_GENHD_FLAGS;
gdsk->minors = 1;
/* Rescan partitions */
cas_reread_partitions(bd);
return 0;
}
static int _cas_exp_obj_allocate_minors(int count)
{
int minor = -1;
if (cas_module.next_minor + count <= (1 << MINORBITS)) {
minor = cas_module.next_minor;
cas_module.next_minor += count;
}
return minor;
}
static int _cas_exp_obj_set_dev_t(struct cas_disk *dsk, struct gendisk *gd)
{
struct cas_exp_obj *exp_obj = dsk->exp_obj;
int flags;
int minors = GET_DISK_MAX_PARTS(cas_disk_get_gendisk(dsk));
struct block_device *bdev;
bdev = cas_disk_get_blkdev(dsk);
BUG_ON(!bdev);
if (cas_bdev_whole(bdev) != bdev) {
minors = 1;
flags = 0;
} else {
if (_cas_exp_obj_hide_parts(dsk))
return -EINVAL;
flags = exp_obj->gd_flags;
}
gd->first_minor = _cas_exp_obj_allocate_minors(minors);
if (gd->first_minor < 0) {
CAS_DEBUG_DISK_ERROR(dsk, "Cannot allocate %d minors", minors);
return -EINVAL;
}
gd->minors = minors;
gd->major = cas_module.disk_major;
gd->flags |= flags;
return 0;
}
static void _cas_exp_obj_clear_dev_t(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj = dsk->exp_obj;
struct block_device *bdev = cas_disk_get_blkdev(dsk);
struct gendisk *gdsk = cas_disk_get_gendisk(dsk);
if (cas_bdev_whole(bdev) == bdev) {
/* Restore previous configuration of bottom disk */
gdsk->minors = exp_obj->gd_minors;
gdsk->flags |= exp_obj->gd_flags;
cas_reread_partitions(bdev);
}
}
CAS_BDEV_OPEN(_cas_exp_obj_open, struct gendisk *gd)
{
struct cas_disk *dsk = gd->private_data;
struct cas_exp_obj *exp_obj = dsk->exp_obj;
int result = -ENAVAIL;
mutex_lock(&exp_obj->openers_lock);
if (!exp_obj->claimed) {
if (unlikely(exp_obj->openers == UINT_MAX)) {
result = -EBUSY;
} else {
exp_obj->openers++;
result = 0;
}
}
mutex_unlock(&dsk->exp_obj->openers_lock);
return result;
}
CAS_BDEV_CLOSE(_cas_exp_obj_close, struct gendisk *gd)
{
struct cas_disk *dsk = gd->private_data;
struct cas_exp_obj *exp_obj = dsk->exp_obj;
BUG_ON(exp_obj->openers == 0);
mutex_lock(&exp_obj->openers_lock);
exp_obj->openers--;
mutex_unlock(&exp_obj->openers_lock);
}
static const struct block_device_operations _cas_exp_obj_ops = {
.owner = THIS_MODULE,
.open = CAS_REFER_BDEV_OPEN_CALLBACK(_cas_exp_obj_open),
.release = CAS_REFER_BDEV_CLOSE_CALLBACK(_cas_exp_obj_close),
CAS_SET_SUBMIT_BIO(_cas_exp_obj_submit_bio)
};
static int cas_exp_obj_alloc(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj;
BUG_ON(!dsk);
BUG_ON(dsk->exp_obj);
CAS_DEBUG_DISK_TRACE(dsk);
exp_obj = kmem_cache_zalloc(cas_module.exp_obj_cache, GFP_KERNEL);
if (!exp_obj) {
CAS_DEBUG_ERROR("Cannot allocate memory");
return -ENOMEM;
}
dsk->exp_obj = exp_obj;
return 0;
}
static void cas_exp_obj_free(struct cas_disk *dsk)
{
if (!dsk->exp_obj)
return;
kmem_cache_free(cas_module.exp_obj_cache, dsk->exp_obj);
dsk->exp_obj = NULL;
}
static CAS_BLK_STATUS_T _cas_exp_obj_queue_rq(struct blk_mq_hw_ctx *hctx,
const struct blk_mq_queue_data *bd)
{
return CAS_BLK_STS_NOTSUPP;
}
static struct blk_mq_ops cas_mq_ops = {
.queue_rq = _cas_exp_obj_queue_rq,
#ifdef CAS_BLK_MQ_OPS_MAP_QUEUE
.map_queue = blk_mq_map_queue,
#endif
};
static void _cas_init_queues(struct cas_disk *dsk)
{
struct request_queue *q = dsk->exp_obj->queue;
struct blk_mq_hw_ctx *hctx;
unsigned long i;
queue_for_each_hw_ctx(q, hctx, i) {
if (!hctx->nr_ctx || !hctx->tags)
continue;
hctx->driver_data = dsk;
}
}
static int _cas_init_tag_set(struct cas_disk *dsk, struct blk_mq_tag_set *set)
{
BUG_ON(!dsk);
BUG_ON(!set);
set->ops = &cas_mq_ops;
set->nr_hw_queues = num_online_cpus();
set->numa_node = NUMA_NO_NODE;
/*TODO: Should we inherit qd from core device? */
set->queue_depth = CAS_BLKDEV_DEFAULT_RQ;
set->cmd_size = 0;
set->flags = BLK_MQ_F_SHOULD_MERGE | CAS_BLK_MQ_F_STACKING | CAS_BLK_MQ_F_BLOCKING;
set->driver_data = dsk;
return blk_mq_alloc_tag_set(set);
}
static int _cas_exp_obj_check_path(const char *dev_name)
{
struct file *exported;
char *path;
int result;
path = kmalloc(PATH_MAX, GFP_KERNEL);
if (!path)
return -ENOMEM;
snprintf(path, PATH_MAX, "/dev/%s", dev_name);
exported = filp_open(path, O_RDONLY, 0);
if (!IS_ERR_OR_NULL(exported)) {
filp_close(exported, NULL);
result = -EEXIST;
} else {
/* failed to open file - it is safe to assume,
* it does not exist
*/
result = 0;
}
kfree(path);
return result;
}
int cas_exp_obj_create(struct cas_disk *dsk, const char *dev_name,
struct module *owner, struct cas_exp_obj_ops *ops, void *priv)
{
struct cas_exp_obj *exp_obj;
struct request_queue *queue;
struct gendisk *gd;
int result = 0;
BUG_ON(!owner);
BUG_ON(!dsk);
BUG_ON(!ops);
BUG_ON(dsk->exp_obj);
CAS_DEBUG_DISK_TRACE(dsk);
if (strlen(dev_name) >= DISK_NAME_LEN)
return -EINVAL;
result = _cas_exp_obj_check_path(dev_name);
if (result == -EEXIST) {
printk(KERN_ERR "Could not activate exported object, "
"because file /dev/%s exists.\n", dev_name);
}
if (result)
return result;
result = cas_exp_obj_alloc(dsk);
if (result)
return result;
exp_obj = dsk->exp_obj;
mutex_init(&exp_obj->openers_lock);
exp_obj->dev_name = kstrdup(dev_name, GFP_KERNEL);
if (!exp_obj->dev_name) {
result = -ENOMEM;
goto error_kstrdup;
}
if (!try_module_get(owner)) {
CAS_DEBUG_DISK_ERROR(dsk, "Cannot get reference to module");
result = -ENAVAIL;
goto error_module_get;
}
exp_obj->owner = owner;
exp_obj->ops = ops;
result = _cas_init_tag_set(dsk, &exp_obj->tag_set);
if (result) {
goto error_init_tag_set;
}
result = cas_alloc_mq_disk(&gd, &queue, &exp_obj->tag_set);
if (result) {
goto error_alloc_mq_disk;
}
exp_obj->gd = gd;
result = _cas_exp_obj_set_dev_t(dsk, gd);
if (result)
goto error_exp_obj_set_dev_t;
BUG_ON(queue->queuedata);
queue->queuedata = dsk;
exp_obj->queue = queue;
exp_obj->private = priv;
_cas_init_queues(dsk);
gd->fops = &_cas_exp_obj_ops;
gd->private_data = dsk;
strscpy(gd->disk_name, exp_obj->dev_name, sizeof(gd->disk_name));
cas_blk_queue_make_request(queue, _cas_exp_obj_make_rq_fn);
if (exp_obj->ops->set_geometry) {
result = exp_obj->ops->set_geometry(dsk, exp_obj->private);
if (result)
goto error_set_geometry;
}
if (cas_add_disk(gd))
goto error_add_disk;
result = bd_claim_by_disk(dsk->bd, dsk, gd);
if (result)
goto error_bd_claim;
return 0;
error_bd_claim:
del_gendisk(dsk->exp_obj->gd);
error_add_disk:
error_set_geometry:
exp_obj->private = NULL;
_cas_exp_obj_clear_dev_t(dsk);
error_exp_obj_set_dev_t:
cas_cleanup_mq_disk(gd);
exp_obj->gd = NULL;
error_alloc_mq_disk:
blk_mq_free_tag_set(&exp_obj->tag_set);
error_init_tag_set:
module_put(owner);
dsk->exp_obj->owner = NULL;
error_module_get:
kfree(exp_obj->dev_name);
error_kstrdup:
cas_exp_obj_free(dsk);
return result;
}
int cas_exp_obj_destroy(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj;
BUG_ON(!dsk);
if (!dsk->exp_obj)
return -ENODEV;
CAS_DEBUG_DISK_TRACE(dsk);
exp_obj = dsk->exp_obj;
bd_release_from_disk(dsk->bd, exp_obj->gd);
_cas_exp_obj_clear_dev_t(dsk);
del_gendisk(exp_obj->gd);
if (exp_obj->queue)
cas_cleanup_queue(exp_obj->queue);
blk_mq_free_tag_set(&exp_obj->tag_set);
put_disk(exp_obj->gd);
return 0;
}
void cas_exp_obj_cleanup(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj;
struct module *owner;
CAS_DEBUG_DISK_TRACE(dsk);
exp_obj = dsk->exp_obj;
if (!exp_obj)
return;
owner = exp_obj->owner;
kfree(exp_obj->dev_name);
cas_exp_obj_free(dsk);
if (owner)
module_put(owner);
}
int cas_exp_obj_lock(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj;
int result = -EBUSY;
BUG_ON(!dsk);
BUG_ON(!dsk->exp_obj);
CAS_DEBUG_DISK_TRACE(dsk);
exp_obj = dsk->exp_obj;
mutex_lock(&exp_obj->openers_lock);
if (exp_obj->openers == 0) {
exp_obj->claimed = true;
result = 0;
}
mutex_unlock(&exp_obj->openers_lock);
return result;
}
int cas_exp_obj_unlock(struct cas_disk *dsk)
{
struct cas_exp_obj *exp_obj;
BUG_ON(!dsk);
CAS_DEBUG_DISK_TRACE(dsk);
exp_obj = dsk->exp_obj;
mutex_lock(&exp_obj->openers_lock);
exp_obj->claimed = false;
mutex_unlock(&exp_obj->openers_lock);
return 0;
}
struct request_queue *cas_exp_obj_get_queue(struct cas_disk *dsk)
{
BUG_ON(!dsk);
BUG_ON(!dsk->exp_obj);
return dsk->exp_obj->queue;
}
struct gendisk *cas_exp_obj_get_gendisk(struct cas_disk *dsk)
{
BUG_ON(!dsk);
BUG_ON(!dsk->exp_obj);
return dsk->exp_obj->gd;
}