commit
82ce9342a2
@ -194,7 +194,7 @@ class Cache:
|
|||||||
def set_params_nhit(self, promotion_params_nhit: PromotionParametersNhit) -> Output:
|
def set_params_nhit(self, promotion_params_nhit: PromotionParametersNhit) -> Output:
|
||||||
return casadm.set_param_promotion_nhit(
|
return casadm.set_param_promotion_nhit(
|
||||||
self.cache_id,
|
self.cache_id,
|
||||||
threshold=promotion_params_nhit.threshold.get_value(),
|
threshold=promotion_params_nhit.threshold,
|
||||||
trigger=promotion_params_nhit.trigger
|
trigger=promotion_params_nhit.trigger
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -72,9 +72,9 @@ class CacheMode(Enum):
|
|||||||
|
|
||||||
|
|
||||||
class SeqCutOffPolicy(Enum):
|
class SeqCutOffPolicy(Enum):
|
||||||
full = 0
|
full = "full"
|
||||||
always = 1
|
always = "always"
|
||||||
never = 2
|
never = "never"
|
||||||
DEFAULT = full
|
DEFAULT = full
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
@ -85,6 +85,9 @@ class SeqCutOffPolicy(Enum):
|
|||||||
|
|
||||||
raise ValueError(f"{name} is not a valid sequential cut off name")
|
raise ValueError(f"{name} is not a valid sequential cut off name")
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return self.value
|
||||||
|
|
||||||
|
|
||||||
class MetadataMode(Enum):
|
class MetadataMode(Enum):
|
||||||
normal = "normal"
|
normal = "normal"
|
||||||
@ -240,7 +243,7 @@ class SeqCutOffParameters:
|
|||||||
|
|
||||||
|
|
||||||
class PromotionParametersNhit:
|
class PromotionParametersNhit:
|
||||||
def __init__(self, threshold: Size = None, trigger: int = None):
|
def __init__(self, threshold: int = None, trigger: int = None):
|
||||||
self.threshold = threshold
|
self.threshold = threshold
|
||||||
self.trigger = trigger
|
self.trigger = trigger
|
||||||
|
|
||||||
|
@ -68,6 +68,36 @@ def get_cores(cache_id: int) -> list:
|
|||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def get_inactive_cores(cache_id: int) -> list:
|
||||||
|
from api.cas.core import Core, CoreStatus
|
||||||
|
|
||||||
|
cores_dict = get_cas_devices_dict()["cores"].values()
|
||||||
|
|
||||||
|
def is_inactive(core):
|
||||||
|
return CoreStatus[core["status"].lower()] == CoreStatus.inactive
|
||||||
|
|
||||||
|
return [
|
||||||
|
Core(core["device_path"], core["cache_id"])
|
||||||
|
for core in cores_dict
|
||||||
|
if is_inactive(core) and core["cache_id"] == cache_id
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def get_detached_cores(cache_id: int) -> list:
|
||||||
|
from api.cas.core import Core, CoreStatus
|
||||||
|
|
||||||
|
cores_dict = get_cas_devices_dict()["cores"].values()
|
||||||
|
|
||||||
|
def is_detached(core):
|
||||||
|
return CoreStatus[core["status"].lower()] == CoreStatus.detached
|
||||||
|
|
||||||
|
return [
|
||||||
|
Core(core["device_path"], core["cache_id"])
|
||||||
|
for core in cores_dict
|
||||||
|
if is_detached(core) and core["cache_id"] == cache_id
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
def get_cas_devices_dict() -> dict:
|
def get_cas_devices_dict() -> dict:
|
||||||
device_list = list(csv.DictReader(casadm.list_caches(OutputFormat.csv).stdout.split("\n")))
|
device_list = list(csv.DictReader(casadm.list_caches(OutputFormat.csv).stdout.split("\n")))
|
||||||
devices = {"caches": {}, "cores": {}, "core_pool": {}}
|
devices = {"caches": {}, "cores": {}, "core_pool": {}}
|
||||||
@ -92,9 +122,7 @@ def get_cas_devices_dict() -> dict:
|
|||||||
]
|
]
|
||||||
if core_pool:
|
if core_pool:
|
||||||
params.append(("core_pool", device))
|
params.append(("core_pool", device))
|
||||||
devices["core_pool"][device["disk"]] = dict(
|
devices["core_pool"][device["disk"]] = dict([(key, value) for key, value in params])
|
||||||
[(key, value) for key, value in params]
|
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
devices["cores"][(cache_id, int(device["id"]))] = dict(
|
devices["cores"][(cache_id, int(device["id"]))] = dict(
|
||||||
[(key, value) for key, value in params]
|
[(key, value) for key, value in params]
|
||||||
@ -205,11 +233,14 @@ def get_io_class_list(cache_id: int) -> list:
|
|||||||
return ret
|
return ret
|
||||||
|
|
||||||
|
|
||||||
def get_core_info_by_path(core_disk_path) -> dict | None:
|
def get_core_info_for_cache_by_path(core_disk_path: str, target_cache_id: int) -> dict | None:
|
||||||
output = casadm.list_caches(OutputFormat.csv, by_id_path=True)
|
output = casadm.list_caches(OutputFormat.csv, by_id_path=True)
|
||||||
reader = csv.DictReader(io.StringIO(output.stdout))
|
reader = csv.DictReader(io.StringIO(output.stdout))
|
||||||
|
cache_id = -1
|
||||||
for row in reader:
|
for row in reader:
|
||||||
if row["type"] == "core" and row["disk"] == core_disk_path:
|
if row["type"] == "cache":
|
||||||
|
cache_id = int(row["id"])
|
||||||
|
if row["type"] == "core" and row["disk"] == core_disk_path and target_cache_id == cache_id:
|
||||||
return {
|
return {
|
||||||
"core_id": row["id"],
|
"core_id": row["id"],
|
||||||
"core_device": row["disk"],
|
"core_device": row["disk"],
|
||||||
|
@ -11,7 +11,7 @@ from enum import Enum
|
|||||||
from api.cas import casadm
|
from api.cas import casadm
|
||||||
from api.cas.cache_config import SeqCutOffParameters, SeqCutOffPolicy
|
from api.cas.cache_config import SeqCutOffParameters, SeqCutOffPolicy
|
||||||
from api.cas.casadm_params import StatsFilter
|
from api.cas.casadm_params import StatsFilter
|
||||||
from api.cas.casadm_parser import get_seq_cut_off_parameters, get_core_info_by_path
|
from api.cas.casadm_parser import get_seq_cut_off_parameters, get_core_info_for_cache_by_path
|
||||||
from api.cas.statistics import CoreStats, CoreIoClassStats
|
from api.cas.statistics import CoreStats, CoreIoClassStats
|
||||||
from core.test_run_utils import TestRun
|
from core.test_run_utils import TestRun
|
||||||
from storage_devices.device import Device
|
from storage_devices.device import Device
|
||||||
@ -35,18 +35,19 @@ class Core(Device):
|
|||||||
def __init__(self, core_device: str, cache_id: int):
|
def __init__(self, core_device: str, cache_id: int):
|
||||||
self.core_device = Device(core_device)
|
self.core_device = Device(core_device)
|
||||||
self.path = None
|
self.path = None
|
||||||
|
self.cache_id = cache_id
|
||||||
core_info = self.__get_core_info()
|
core_info = self.__get_core_info()
|
||||||
# "-" is special case for cores in core pool
|
# "-" is special case for cores in core pool
|
||||||
if core_info["core_id"] != "-":
|
if core_info["core_id"] != "-":
|
||||||
self.core_id = int(core_info["core_id"])
|
self.core_id = int(core_info["core_id"])
|
||||||
if core_info["exp_obj"] != "-":
|
if core_info["exp_obj"] != "-":
|
||||||
Device.__init__(self, core_info["exp_obj"])
|
Device.__init__(self, core_info["exp_obj"])
|
||||||
self.cache_id = cache_id
|
|
||||||
self.partitions = []
|
self.partitions = []
|
||||||
self.block_size = None
|
self.block_size = None
|
||||||
|
|
||||||
def __get_core_info(self):
|
def __get_core_info(self):
|
||||||
return get_core_info_by_path(self.core_device.path)
|
return get_core_info_for_cache_by_path(core_disk_path=self.core_device.path,
|
||||||
|
target_cache_id=self.cache_id)
|
||||||
|
|
||||||
def create_filesystem(self, fs_type: disk_utils.Filesystem, force=True, blocksize=None):
|
def create_filesystem(self, fs_type: disk_utils.Filesystem, force=True, blocksize=None):
|
||||||
super().create_filesystem(fs_type, force, blocksize)
|
super().create_filesystem(fs_type, force, blocksize)
|
||||||
|
@ -1,6 +1,6 @@
|
|||||||
#
|
#
|
||||||
# Copyright(c) 2020-2022 Intel Corporation
|
# Copyright(c) 2020-2022 Intel Corporation
|
||||||
# Copyright(c) 2024 Huawei Technologies Co., Ltd.
|
# Copyright(c) 2024 Huawei Technologies
|
||||||
# SPDX-License-Identifier: BSD-3-Clause
|
# SPDX-License-Identifier: BSD-3-Clause
|
||||||
#
|
#
|
||||||
|
|
||||||
@ -11,13 +11,13 @@ from api.cas.cache_config import CacheMode, CleaningPolicy, CacheModeTrait
|
|||||||
from api.cas.casadm_parser import wait_for_flushing
|
from api.cas.casadm_parser import wait_for_flushing
|
||||||
from core.test_run import TestRun
|
from core.test_run import TestRun
|
||||||
from storage_devices.disk import DiskType, DiskTypeSet, DiskTypeLowerThan
|
from storage_devices.disk import DiskType, DiskTypeSet, DiskTypeLowerThan
|
||||||
from test_tools.disk_utils import Filesystem
|
from test_tools import fs_utils
|
||||||
from test_tools.dd import Dd
|
from test_tools.dd import Dd
|
||||||
|
from test_tools.disk_utils import Filesystem
|
||||||
from test_utils import os_utils
|
from test_utils import os_utils
|
||||||
from test_utils.os_utils import Udev, DropCachesMode
|
from test_utils.os_utils import Udev, DropCachesMode
|
||||||
from test_utils.size import Size, Unit
|
from test_utils.size import Size, Unit
|
||||||
from tests.lazy_writes.recovery.recovery_tests_methods import compare_files
|
from tests.lazy_writes.recovery.recovery_tests_methods import compare_files
|
||||||
from test_tools import fs_utils
|
|
||||||
|
|
||||||
mount_point = "/mnt/cas"
|
mount_point = "/mnt/cas"
|
||||||
test_file_path = f"{mount_point}/test_file"
|
test_file_path = f"{mount_point}/test_file"
|
||||||
@ -197,17 +197,17 @@ def test_interrupt_cache_flush(cache_mode, filesystem):
|
|||||||
@pytest.mark.require_disk("core", DiskTypeLowerThan("cache"))
|
@pytest.mark.require_disk("core", DiskTypeLowerThan("cache"))
|
||||||
def test_interrupt_core_remove(cache_mode, filesystem):
|
def test_interrupt_core_remove(cache_mode, filesystem):
|
||||||
"""
|
"""
|
||||||
title: Test if OpenCAS works correctly after core's removing interruption.
|
title: Core removal interruption.
|
||||||
description: |
|
description: |
|
||||||
Negative test of the ability of OpenCAS to handle core's removing interruption.
|
Test for proper handling of 'core remove' operation interruption.
|
||||||
pass_criteria:
|
pass_criteria:
|
||||||
- No system crash.
|
- No system crash.
|
||||||
- Core would not be removed from cache after interruption.
|
- Core would not be removed from cache after interruption.
|
||||||
- Flushing would be stopped after interruption.
|
- Flushing would be stopped after interruption.
|
||||||
- Md5sum are correct during all test steps.
|
- Checksums are correct during all test steps.
|
||||||
- Dirty blocks quantity after interruption is lower but non-zero.
|
- Dirty blocks quantity after interruption is lower but non-zero.
|
||||||
"""
|
"""
|
||||||
with TestRun.step("Prepare cache and core."):
|
with TestRun.step("Prepare cache and core devices"):
|
||||||
cache_dev = TestRun.disks["cache"]
|
cache_dev = TestRun.disks["cache"]
|
||||||
cache_dev.create_partitions([cache_size])
|
cache_dev.create_partitions([cache_size])
|
||||||
cache_part = cache_dev.partitions[0]
|
cache_part = cache_dev.partitions[0]
|
||||||
@ -216,37 +216,36 @@ def test_interrupt_core_remove(cache_mode, filesystem):
|
|||||||
core_part = core_dev.partitions[0]
|
core_part = core_dev.partitions[0]
|
||||||
|
|
||||||
for _ in TestRun.iteration(
|
for _ in TestRun.iteration(
|
||||||
range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times."
|
range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times"
|
||||||
):
|
):
|
||||||
|
with TestRun.step("Start cache"):
|
||||||
with TestRun.step("Start cache."):
|
|
||||||
cache = casadm.start_cache(cache_part, cache_mode, force=True)
|
cache = casadm.start_cache(cache_part, cache_mode, force=True)
|
||||||
|
|
||||||
with TestRun.step("Set cleaning policy to NOP."):
|
with TestRun.step("Set cleaning policy to NOP"):
|
||||||
cache.set_cleaning_policy(CleaningPolicy.nop)
|
cache.set_cleaning_policy(CleaningPolicy.nop)
|
||||||
|
|
||||||
with TestRun.step(f"Add core device with {filesystem} filesystem and mount it."):
|
with TestRun.step(f"Add core device with {filesystem} filesystem and mount it"):
|
||||||
core_part.create_filesystem(filesystem)
|
core_part.create_filesystem(filesystem)
|
||||||
core = cache.add_core(core_part)
|
core = cache.add_core(core_part)
|
||||||
core.mount(mount_point)
|
core.mount(mount_point)
|
||||||
|
|
||||||
with TestRun.step(f"Create test file in mount point of exported object."):
|
with TestRun.step("Create test file in mount point of exported object"):
|
||||||
test_file = create_test_file()
|
test_file = create_test_file()
|
||||||
|
|
||||||
with TestRun.step("Check md5 sum of test file."):
|
with TestRun.step("Calculate checksum of test file"):
|
||||||
test_file_md5sum_before = test_file.md5sum()
|
test_file_crc32sum_before = test_file.crc32sum()
|
||||||
|
|
||||||
with TestRun.step(
|
with TestRun.step(
|
||||||
"Get number of dirty data on exported object before core removal interruption."
|
"Get number of dirty data on exported object before core removal interruption"
|
||||||
):
|
):
|
||||||
os_utils.sync()
|
os_utils.sync()
|
||||||
os_utils.drop_caches(DropCachesMode.ALL)
|
os_utils.drop_caches(DropCachesMode.ALL)
|
||||||
cache_dirty_blocks_before = cache.get_dirty_blocks()
|
cache_dirty_blocks_before = cache.get_dirty_blocks()
|
||||||
|
|
||||||
with TestRun.step("Unmount core."):
|
with TestRun.step("Unmount core"):
|
||||||
core.unmount()
|
core.unmount()
|
||||||
|
|
||||||
with TestRun.step("Start removing core device."):
|
with TestRun.step("Start removing core"):
|
||||||
flush_pid = TestRun.executor.run_in_background(
|
flush_pid = TestRun.executor.run_in_background(
|
||||||
cli.remove_core_cmd(str(cache.cache_id), str(core.core_id))
|
cli.remove_core_cmd(str(cache.cache_id), str(core.core_id))
|
||||||
)
|
)
|
||||||
@ -258,42 +257,39 @@ def test_interrupt_core_remove(cache_mode, filesystem):
|
|||||||
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
|
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
|
||||||
TestRun.executor.run(f"kill -s SIGINT {flush_pid}")
|
TestRun.executor.run(f"kill -s SIGINT {flush_pid}")
|
||||||
|
|
||||||
with TestRun.step("Check md5 sum of test file after interruption."):
|
|
||||||
cache.set_cache_mode(CacheMode.WO)
|
|
||||||
test_file_md5sum_interrupt = test_file.md5sum()
|
|
||||||
cache.set_cache_mode(cache_mode)
|
|
||||||
|
|
||||||
with TestRun.step(
|
with TestRun.step(
|
||||||
"Check number of dirty data on exported object after core removal interruption."
|
"Check number of dirty data on exported object after core removal interruption"
|
||||||
):
|
):
|
||||||
cache_dirty_blocks_after = cache.get_dirty_blocks()
|
cache_dirty_blocks_after = cache.get_dirty_blocks()
|
||||||
if cache_dirty_blocks_after >= cache_dirty_blocks_before:
|
if cache_dirty_blocks_after >= cache_dirty_blocks_before:
|
||||||
TestRun.LOGGER.error(
|
TestRun.LOGGER.error(
|
||||||
"Quantity of dirty lines after core removal interruption " "should be lower."
|
"Quantity of dirty lines after core removal interruption should be lower."
|
||||||
)
|
)
|
||||||
if int(cache_dirty_blocks_after) == 0:
|
if int(cache_dirty_blocks_after) == 0:
|
||||||
TestRun.LOGGER.error(
|
TestRun.LOGGER.error(
|
||||||
"Quantity of dirty lines after core removal interruption " "should not be zero."
|
"Quantity of dirty lines after core removal interruption should not be zero."
|
||||||
)
|
)
|
||||||
|
|
||||||
with TestRun.step("Remove core from cache."):
|
with TestRun.step("Mount core and verify test file checksum after interruption"):
|
||||||
core.remove_core()
|
core.mount(mount_point)
|
||||||
|
|
||||||
with TestRun.step("Stop cache."):
|
if test_file.crc32sum() != test_file_crc32sum_before:
|
||||||
|
TestRun.LOGGER.error("Checksum after interrupting core removal is different.")
|
||||||
|
|
||||||
|
with TestRun.step("Unmount core"):
|
||||||
|
core.unmount()
|
||||||
|
|
||||||
|
with TestRun.step("Stop cache"):
|
||||||
cache.stop()
|
cache.stop()
|
||||||
|
|
||||||
with TestRun.step("Mount core device."):
|
with TestRun.step("Mount core device"):
|
||||||
core_part.mount(mount_point)
|
core_part.mount(mount_point)
|
||||||
|
|
||||||
with TestRun.step("Check md5 sum of test file again."):
|
with TestRun.step("Verify checksum of test file again"):
|
||||||
if test_file_md5sum_before != test_file.md5sum():
|
if test_file.crc32sum() != test_file_crc32sum_before:
|
||||||
TestRun.LOGGER.error("Md5 sum before interrupting core removal is different.")
|
TestRun.LOGGER.error("Checksum after core removal is different.")
|
||||||
|
|
||||||
is_sum_diff_after_interrupt = test_file_md5sum_interrupt != test_file.md5sum()
|
with TestRun.step("Unmount core device"):
|
||||||
if is_sum_diff_after_interrupt:
|
|
||||||
TestRun.LOGGER.error("Md5 sum after interrupting core removal is different.")
|
|
||||||
|
|
||||||
with TestRun.step("Unmount core device."):
|
|
||||||
core_part.unmount()
|
core_part.unmount()
|
||||||
|
|
||||||
|
|
||||||
@ -315,8 +311,19 @@ def test_interrupt_cache_mode_switch_parametrized(cache_mode, stop_percentage):
|
|||||||
- Md5sum are correct during all test steps.
|
- Md5sum are correct during all test steps.
|
||||||
- Dirty blocks quantity after interruption is lower but non-zero.
|
- Dirty blocks quantity after interruption is lower but non-zero.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
|
test_file_size = Size(1, Unit.GibiByte)
|
||||||
|
test_file_path = "/mnt/cas/test_file"
|
||||||
|
|
||||||
with TestRun.step("Prepare cache and core."):
|
with TestRun.step("Prepare cache and core."):
|
||||||
cache_part, core_part = prepare()
|
cache_dev = TestRun.disks["cache"]
|
||||||
|
core_dev = TestRun.disks["core"]
|
||||||
|
|
||||||
|
cache_dev.create_partitions([cache_size])
|
||||||
|
core_dev.create_partitions([cache_size * 2])
|
||||||
|
|
||||||
|
cache_part = cache_dev.partitions[0]
|
||||||
|
core_part = core_dev.partitions[0]
|
||||||
|
|
||||||
with TestRun.step("Disable udev"):
|
with TestRun.step("Disable udev"):
|
||||||
Udev.disable()
|
Udev.disable()
|
||||||
@ -325,23 +332,22 @@ def test_interrupt_cache_mode_switch_parametrized(cache_mode, stop_percentage):
|
|||||||
range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times."
|
range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times."
|
||||||
):
|
):
|
||||||
|
|
||||||
with TestRun.step("Start cache."):
|
with TestRun.step("Start cache"):
|
||||||
cache = casadm.start_cache(cache_part, cache_mode, force=True)
|
cache = casadm.start_cache(cache_part, cache_mode, force=True)
|
||||||
|
|
||||||
with TestRun.step("Set cleaning policy to NOP."):
|
with TestRun.step("Set cleaning policy to NOP"):
|
||||||
cache.set_cleaning_policy(CleaningPolicy.nop)
|
cache.set_cleaning_policy(CleaningPolicy.nop)
|
||||||
|
|
||||||
with TestRun.step(f"Add core device."):
|
with TestRun.step(f"Add core device"):
|
||||||
core = cache.add_core(core_part)
|
core = cache.add_core(core_part)
|
||||||
|
|
||||||
with TestRun.step(f"Create test file in mount point of exported object."):
|
with TestRun.step(f"Create test file in mount point of exported object"):
|
||||||
test_file_size = Size(4, Unit.GibiByte)
|
|
||||||
test_file = fs_utils.create_random_test_file(test_file_path, test_file_size)
|
test_file = fs_utils.create_random_test_file(test_file_path, test_file_size)
|
||||||
|
|
||||||
with TestRun.step("Check md5 sum of test file."):
|
with TestRun.step("Calculate md5sum of test file"):
|
||||||
test_file_md5_before = test_file.md5sum()
|
test_file_md5_before = test_file.md5sum()
|
||||||
|
|
||||||
with TestRun.step("Export file to CAS"):
|
with TestRun.step("Copy test data to core"):
|
||||||
dd = (
|
dd = (
|
||||||
Dd()
|
Dd()
|
||||||
.block_size(test_file_size)
|
.block_size(test_file_size)
|
||||||
@ -351,49 +357,58 @@ def test_interrupt_cache_mode_switch_parametrized(cache_mode, stop_percentage):
|
|||||||
)
|
)
|
||||||
dd.run()
|
dd.run()
|
||||||
|
|
||||||
with TestRun.step("Get number of dirty data on exported object before interruption."):
|
with TestRun.step("Get number of dirty data on exported object before interruption"):
|
||||||
os_utils.sync()
|
os_utils.sync()
|
||||||
os_utils.drop_caches(DropCachesMode.ALL)
|
os_utils.drop_caches(DropCachesMode.ALL)
|
||||||
cache_dirty_blocks_before = cache.get_dirty_blocks()
|
cache_dirty_blocks_before = cache.get_dirty_blocks()
|
||||||
|
|
||||||
with TestRun.step("Start switching cache mode."):
|
with TestRun.step("Start switching cache mode"):
|
||||||
flush_pid = TestRun.executor.run_in_background(
|
flush_pid = TestRun.executor.run_in_background(
|
||||||
cli.set_cache_mode_cmd(
|
cli.set_cache_mode_cmd(
|
||||||
str(CacheMode.DEFAULT.name.lower()), str(cache.cache_id), "yes"
|
cache_mode=str(CacheMode.DEFAULT.name.lower()),
|
||||||
|
cache_id=str(cache.cache_id),
|
||||||
|
flush_cache="yes",
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
with TestRun.step("Send interruption signal."):
|
with TestRun.step("Kill flush process during cache flush operation"):
|
||||||
wait_for_flushing(cache, core)
|
wait_for_flushing(cache, core)
|
||||||
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
|
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
|
||||||
while percentage < stop_percentage:
|
while percentage < stop_percentage:
|
||||||
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
|
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
|
||||||
TestRun.executor.run(f"kill -s SIGINT {flush_pid}")
|
TestRun.executor.kill_process(flush_pid)
|
||||||
|
|
||||||
with TestRun.step("Check number of dirty data on exported object after interruption."):
|
with TestRun.step("Check number of dirty data on exported object after interruption"):
|
||||||
cache_dirty_blocks_after = cache.get_dirty_blocks()
|
cache_dirty_blocks_after = cache.get_dirty_blocks()
|
||||||
if cache_dirty_blocks_after >= cache_dirty_blocks_before:
|
if cache_dirty_blocks_after >= cache_dirty_blocks_before:
|
||||||
TestRun.LOGGER.error(
|
TestRun.LOGGER.error(
|
||||||
"Quantity of dirty lines after cache mode switching "
|
"Quantity of dirty lines after cache mode switching "
|
||||||
"interruption should be lower."
|
"interruption should be lower."
|
||||||
)
|
)
|
||||||
if int(cache_dirty_blocks_after) == 0:
|
if cache_dirty_blocks_after == Size.zero():
|
||||||
TestRun.LOGGER.error(
|
TestRun.LOGGER.error(
|
||||||
"Quantity of dirty lines after cache mode switching "
|
"Quantity of dirty lines after cache mode switching "
|
||||||
"interruption should not be zero."
|
"interruption should not be zero."
|
||||||
)
|
)
|
||||||
|
|
||||||
with TestRun.step("Check cache mode."):
|
with TestRun.step("Check cache mode"):
|
||||||
if cache.get_cache_mode() != cache_mode:
|
if cache.get_cache_mode() != cache_mode:
|
||||||
TestRun.LOGGER.error("Cache mode should remain the same.")
|
TestRun.LOGGER.error("Cache mode should remain the same.")
|
||||||
|
|
||||||
with TestRun.step("Stop cache."):
|
with TestRun.step("Stop cache"):
|
||||||
cache.stop()
|
cache.stop()
|
||||||
|
|
||||||
with TestRun.step("Check md5 sum of test file again."):
|
with TestRun.step("Copy test data from the exported object to a file"):
|
||||||
Dd().block_size(test_file_size).input(core.path).output(test_file.full_path).oflag(
|
dd = (
|
||||||
"direct"
|
Dd()
|
||||||
).run()
|
.block_size(test_file_size)
|
||||||
|
.input(core.path)
|
||||||
|
.output(test_file.full_path)
|
||||||
|
.oflag("direct")
|
||||||
|
)
|
||||||
|
dd.run()
|
||||||
|
|
||||||
|
with TestRun.step("Compare md5 sum of test files"):
|
||||||
target_file_md5 = test_file.md5sum()
|
target_file_md5 = test_file.md5sum()
|
||||||
compare_files(test_file_md5_before, target_file_md5)
|
compare_files(test_file_md5_before, target_file_md5)
|
||||||
|
|
||||||
@ -491,7 +506,7 @@ def create_test_file():
|
|||||||
bs = Size(512, Unit.KibiByte)
|
bs = Size(512, Unit.KibiByte)
|
||||||
cnt = int(cache_size.value / bs.value)
|
cnt = int(cache_size.value / bs.value)
|
||||||
test_file = File.create_file(test_file_path)
|
test_file = File.create_file(test_file_path)
|
||||||
dd = Dd().output(test_file_path).input("/dev/zero").block_size(bs).count(cnt)
|
dd = Dd().output(test_file_path).input("/dev/zero").block_size(bs).count(cnt).oflag("direct")
|
||||||
dd.run()
|
dd.run()
|
||||||
test_file.refresh_item()
|
test_file.refresh_item()
|
||||||
return test_file
|
return test_file
|
||||||
|
@ -12,7 +12,7 @@ from core.test_run import TestRun
|
|||||||
from test_utils.size import Size, Unit
|
from test_utils.size import Size, Unit
|
||||||
from api.cas.cache_config import CacheLineSize, CacheMode, CacheStatus
|
from api.cas.cache_config import CacheLineSize, CacheMode, CacheStatus
|
||||||
from api.cas.casadm_params import StatsFilter
|
from api.cas.casadm_params import StatsFilter
|
||||||
from api.cas.casadm_parser import get_core_info_by_path
|
from api.cas.casadm_parser import get_core_info_for_cache_by_path
|
||||||
from api.cas.core import CoreStatus, Core
|
from api.cas.core import CoreStatus, Core
|
||||||
from test_tools.dd import Dd
|
from test_tools.dd import Dd
|
||||||
from api.cas.cli import standby_activate_cmd
|
from api.cas.cli import standby_activate_cmd
|
||||||
@ -173,7 +173,11 @@ def test_activate_incomplete_cache():
|
|||||||
TestRun.fail(f"Expected one inactive core. Got {inactive_core_count}")
|
TestRun.fail(f"Expected one inactive core. Got {inactive_core_count}")
|
||||||
|
|
||||||
with TestRun.step("Check if core is in an appropriate state"):
|
with TestRun.step("Check if core is in an appropriate state"):
|
||||||
core_status = CoreStatus[get_core_info_by_path(core_dev_path)["status"].lower()]
|
core_status = CoreStatus[
|
||||||
|
get_core_info_for_cache_by_path(
|
||||||
|
core_disk_path=core_dev_path, target_cache_id=cache.cache_id
|
||||||
|
)["status"].lower()
|
||||||
|
]
|
||||||
if core_status != CoreStatus.inactive:
|
if core_status != CoreStatus.inactive:
|
||||||
TestRun.fail(
|
TestRun.fail(
|
||||||
"The core is in an invalid state. "
|
"The core is in an invalid state. "
|
||||||
|
@ -1,11 +1,13 @@
|
|||||||
#
|
#
|
||||||
# Copyright(c) 2020-2021 Intel Corporation
|
# Copyright(c) 2020-2021 Intel Corporation
|
||||||
|
# Copyright(c) 2024 Huawei Technologies Co., Ltd.
|
||||||
# SPDX-License-Identifier: BSD-3-Clause
|
# SPDX-License-Identifier: BSD-3-Clause
|
||||||
#
|
#
|
||||||
|
|
||||||
import math
|
import math
|
||||||
|
import posixpath
|
||||||
import pytest
|
import pytest
|
||||||
import os
|
|
||||||
from api.cas import casadm, cli_messages
|
from api.cas import casadm, cli_messages
|
||||||
from api.cas.cache_config import CacheLineSize
|
from api.cas.cache_config import CacheLineSize
|
||||||
from core.test_run import TestRun
|
from core.test_run import TestRun
|
||||||
@ -32,7 +34,7 @@ def test_device_capabilities():
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
core_device = TestRun.disks['core']
|
core_device = TestRun.disks['core']
|
||||||
max_io_size_path = os.path.join(disk_utils.get_sysfs_path(core_device.get_device_id()),
|
max_io_size_path = posixpath.join(disk_utils.get_sysfs_path(core_device.get_device_id()),
|
||||||
'queue/max_sectors_kb')
|
'queue/max_sectors_kb')
|
||||||
default_max_io_size = fs_utils.read_file(max_io_size_path)
|
default_max_io_size = fs_utils.read_file(max_io_size_path)
|
||||||
|
|
||||||
@ -148,7 +150,7 @@ def measure_capabilities(dev):
|
|||||||
dev_id = dev.parent_device.get_device_id() if isinstance(dev, Partition) \
|
dev_id = dev.parent_device.get_device_id() if isinstance(dev, Partition) \
|
||||||
else dev.get_device_id()
|
else dev.get_device_id()
|
||||||
for c in capabilities:
|
for c in capabilities:
|
||||||
path = os.path.join(disk_utils.get_sysfs_path(dev_id), 'queue', c)
|
path = posixpath.join(disk_utils.get_sysfs_path(dev_id), 'queue', c)
|
||||||
command = f"cat {path}"
|
command = f"cat {path}"
|
||||||
output = TestRun.executor.run(command)
|
output = TestRun.executor.run(command)
|
||||||
if output.exit_code == 0:
|
if output.exit_code == 0:
|
||||||
|
Loading…
Reference in New Issue
Block a user