From 93b27b78947787b715eb88b00a77e1774afa3280 Mon Sep 17 00:00:00 2001 From: Michal Mielewczyk Date: Wed, 1 Jun 2022 14:25:53 +0200 Subject: [PATCH] tests: test for support of different io sizes Signed-off-by: Michal Mielewczyk --- .../512b/test_different_io_sizes_support.py | 93 +++++++++++++++++++ 1 file changed, 93 insertions(+) create mode 100644 test/functional/tests/512b/test_different_io_sizes_support.py diff --git a/test/functional/tests/512b/test_different_io_sizes_support.py b/test/functional/tests/512b/test_different_io_sizes_support.py new file mode 100644 index 0000000..fca44bb --- /dev/null +++ b/test/functional/tests/512b/test_different_io_sizes_support.py @@ -0,0 +1,93 @@ +# +# Copyright(c) 2022 Intel Corporation +# SPDX-License-Identifier: BSD-3-Clause +# + +import pytest +import datetime +import os + +from core.test_run import TestRun +from api.cas import casadm +from storage_devices.disk import DiskType, DiskTypeSet +from api.cas.cache_config import CacheMode +from test_tools import fs_utils +from test_tools.disk_utils import Filesystem +from test_utils.filesystem.directory import Directory +from test_utils.filesystem.file import File +from test_utils.size import Size, Unit +from test_tools.fio.fio import Fio +from test_tools.fio.fio_param import ReadWrite, IoEngine + + +mountpoint = "/tmp/diff_io_size_support_test" + + +@pytest.mark.parametrizex("cache_mode", [CacheMode.WB, CacheMode.WT]) +@pytest.mark.require_disk("cache", DiskTypeSet([DiskType.optane, DiskType.nand])) +@pytest.mark.require_disk("core", DiskTypeSet([DiskType.hdd, DiskType.nand])) +def test_support_different_io_size(cache_mode): + """ + title: OpenCAS supports different IO sizes + description: OpenCAS supports IO of size in rage from 512b to 128K + pass_criteria: + - No IO errors + """ + with TestRun.step("Prepare devices"): + cache_disk = TestRun.disks["cache"] + core_disk = TestRun.disks["core"] + cache_disk.create_partitions([Size(1, Unit.GibiByte)]) + cache_disk = cache_disk.partitions[0] + core_disk.create_partitions([Size(50, Unit.GibiByte)]) + core_disk = core_disk.partitions[0] + + with TestRun.step("Start cache"): + cache = casadm.start_cache(cache_dev=cache_disk, cache_mode=cache_mode, force=True) + core = cache.add_core(core_disk) + + with TestRun.step("Load the default ioclass config file"): + cache.load_io_class("/etc/opencas/ioclass-config.csv") + + with TestRun.step("Create a filesystem on the core device and mount it"): + TestRun.executor.run(f"rm -rf {mountpoint}") + fs_utils.create_directory(path=mountpoint) + core.create_filesystem(Filesystem.xfs) + core.mount(mountpoint) + + with TestRun.step("Run fio with block sizes: 512, 1k, 4k, 5k, 8k, 16k, 32k, 64 and 128k"): + bs_list = [ + Size(512, Unit.Byte), + Size(1, Unit.KibiByte), + Size(4, Unit.KibiByte), + Size(5, Unit.KibiByte), + Size(8, Unit.KibiByte), + Size(16, Unit.KibiByte), + Size(32, Unit.KibiByte), + Size(64, Unit.KibiByte), + Size(128, Unit.KibiByte), + ] + + fio = Fio().create_command() + fio.io_engine(IoEngine.libaio) + fio.time_based() + fio.do_verify() + fio.direct() + fio.read_write(ReadWrite.randwrite) + fio.run_time(datetime.timedelta(seconds=1200)) + fio.io_depth(16) + fio.verify_pattern(0xABCD) + + for i, bs in enumerate(bs_list): + fio_job = fio.add_job() + fio_job.target(os.path.join(mountpoint, str(bs.value))) + fio_job.block_size(bs) + fio_job.file_size(Size((i + 1) * 200, Unit.MebiByte)) + fio_output = fio.run() + + fio_errors = fio_output[0].total_errors() + if fio_errors != 0: + TestRun.fail(f"fio errors: {fio_errors}, should equal 0") + + with TestRun.step("Cleanup"): + core.unmount() + TestRun.executor.run(f"rm -rf {mountpoint}")