tests: test for support of different io sizes

Signed-off-by: Michal Mielewczyk <michal.mielewczyk@intel.com>
This commit is contained in:
Michal Mielewczyk 2022-06-01 14:25:53 +02:00
parent 685b68516a
commit 93b27b7894

View File

@ -0,0 +1,93 @@
#
# Copyright(c) 2022 Intel Corporation
# SPDX-License-Identifier: BSD-3-Clause
#
import pytest
import datetime
import os
from core.test_run import TestRun
from api.cas import casadm
from storage_devices.disk import DiskType, DiskTypeSet
from api.cas.cache_config import CacheMode
from test_tools import fs_utils
from test_tools.disk_utils import Filesystem
from test_utils.filesystem.directory import Directory
from test_utils.filesystem.file import File
from test_utils.size import Size, Unit
from test_tools.fio.fio import Fio
from test_tools.fio.fio_param import ReadWrite, IoEngine
mountpoint = "/tmp/diff_io_size_support_test"
@pytest.mark.parametrizex("cache_mode", [CacheMode.WB, CacheMode.WT])
@pytest.mark.require_disk("cache", DiskTypeSet([DiskType.optane, DiskType.nand]))
@pytest.mark.require_disk("core", DiskTypeSet([DiskType.hdd, DiskType.nand]))
def test_support_different_io_size(cache_mode):
"""
title: OpenCAS supports different IO sizes
description: OpenCAS supports IO of size in rage from 512b to 128K
pass_criteria:
- No IO errors
"""
with TestRun.step("Prepare devices"):
cache_disk = TestRun.disks["cache"]
core_disk = TestRun.disks["core"]
cache_disk.create_partitions([Size(1, Unit.GibiByte)])
cache_disk = cache_disk.partitions[0]
core_disk.create_partitions([Size(50, Unit.GibiByte)])
core_disk = core_disk.partitions[0]
with TestRun.step("Start cache"):
cache = casadm.start_cache(cache_dev=cache_disk, cache_mode=cache_mode, force=True)
core = cache.add_core(core_disk)
with TestRun.step("Load the default ioclass config file"):
cache.load_io_class("/etc/opencas/ioclass-config.csv")
with TestRun.step("Create a filesystem on the core device and mount it"):
TestRun.executor.run(f"rm -rf {mountpoint}")
fs_utils.create_directory(path=mountpoint)
core.create_filesystem(Filesystem.xfs)
core.mount(mountpoint)
with TestRun.step("Run fio with block sizes: 512, 1k, 4k, 5k, 8k, 16k, 32k, 64 and 128k"):
bs_list = [
Size(512, Unit.Byte),
Size(1, Unit.KibiByte),
Size(4, Unit.KibiByte),
Size(5, Unit.KibiByte),
Size(8, Unit.KibiByte),
Size(16, Unit.KibiByte),
Size(32, Unit.KibiByte),
Size(64, Unit.KibiByte),
Size(128, Unit.KibiByte),
]
fio = Fio().create_command()
fio.io_engine(IoEngine.libaio)
fio.time_based()
fio.do_verify()
fio.direct()
fio.read_write(ReadWrite.randwrite)
fio.run_time(datetime.timedelta(seconds=1200))
fio.io_depth(16)
fio.verify_pattern(0xABCD)
for i, bs in enumerate(bs_list):
fio_job = fio.add_job()
fio_job.target(os.path.join(mountpoint, str(bs.value)))
fio_job.block_size(bs)
fio_job.file_size(Size((i + 1) * 200, Unit.MebiByte))
fio_output = fio.run()
fio_errors = fio_output[0].total_errors()
if fio_errors != 0:
TestRun.fail(f"fio errors: {fio_errors}, should equal 0")
with TestRun.step("Cleanup"):
core.unmount()
TestRun.executor.run(f"rm -rf {mountpoint}")