Merge pull request #1573 from Kamoppl/kamilg/fix_bugs

Kamilg/fix bugs
This commit is contained in:
Katarzyna Treder 2024-10-30 14:16:16 +01:00 committed by GitHub
commit 82ce9342a2
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
7 changed files with 134 additions and 78 deletions

View File

@ -194,7 +194,7 @@ class Cache:
def set_params_nhit(self, promotion_params_nhit: PromotionParametersNhit) -> Output: def set_params_nhit(self, promotion_params_nhit: PromotionParametersNhit) -> Output:
return casadm.set_param_promotion_nhit( return casadm.set_param_promotion_nhit(
self.cache_id, self.cache_id,
threshold=promotion_params_nhit.threshold.get_value(), threshold=promotion_params_nhit.threshold,
trigger=promotion_params_nhit.trigger trigger=promotion_params_nhit.trigger
) )

View File

@ -72,9 +72,9 @@ class CacheMode(Enum):
class SeqCutOffPolicy(Enum): class SeqCutOffPolicy(Enum):
full = 0 full = "full"
always = 1 always = "always"
never = 2 never = "never"
DEFAULT = full DEFAULT = full
@classmethod @classmethod
@ -85,6 +85,9 @@ class SeqCutOffPolicy(Enum):
raise ValueError(f"{name} is not a valid sequential cut off name") raise ValueError(f"{name} is not a valid sequential cut off name")
def __str__(self):
return self.value
class MetadataMode(Enum): class MetadataMode(Enum):
normal = "normal" normal = "normal"
@ -240,7 +243,7 @@ class SeqCutOffParameters:
class PromotionParametersNhit: class PromotionParametersNhit:
def __init__(self, threshold: Size = None, trigger: int = None): def __init__(self, threshold: int = None, trigger: int = None):
self.threshold = threshold self.threshold = threshold
self.trigger = trigger self.trigger = trigger

View File

@ -68,6 +68,36 @@ def get_cores(cache_id: int) -> list:
] ]
def get_inactive_cores(cache_id: int) -> list:
from api.cas.core import Core, CoreStatus
cores_dict = get_cas_devices_dict()["cores"].values()
def is_inactive(core):
return CoreStatus[core["status"].lower()] == CoreStatus.inactive
return [
Core(core["device_path"], core["cache_id"])
for core in cores_dict
if is_inactive(core) and core["cache_id"] == cache_id
]
def get_detached_cores(cache_id: int) -> list:
from api.cas.core import Core, CoreStatus
cores_dict = get_cas_devices_dict()["cores"].values()
def is_detached(core):
return CoreStatus[core["status"].lower()] == CoreStatus.detached
return [
Core(core["device_path"], core["cache_id"])
for core in cores_dict
if is_detached(core) and core["cache_id"] == cache_id
]
def get_cas_devices_dict() -> dict: def get_cas_devices_dict() -> dict:
device_list = list(csv.DictReader(casadm.list_caches(OutputFormat.csv).stdout.split("\n"))) device_list = list(csv.DictReader(casadm.list_caches(OutputFormat.csv).stdout.split("\n")))
devices = {"caches": {}, "cores": {}, "core_pool": {}} devices = {"caches": {}, "cores": {}, "core_pool": {}}
@ -92,9 +122,7 @@ def get_cas_devices_dict() -> dict:
] ]
if core_pool: if core_pool:
params.append(("core_pool", device)) params.append(("core_pool", device))
devices["core_pool"][device["disk"]] = dict( devices["core_pool"][device["disk"]] = dict([(key, value) for key, value in params])
[(key, value) for key, value in params]
)
else: else:
devices["cores"][(cache_id, int(device["id"]))] = dict( devices["cores"][(cache_id, int(device["id"]))] = dict(
[(key, value) for key, value in params] [(key, value) for key, value in params]
@ -205,11 +233,14 @@ def get_io_class_list(cache_id: int) -> list:
return ret return ret
def get_core_info_by_path(core_disk_path) -> dict | None: def get_core_info_for_cache_by_path(core_disk_path: str, target_cache_id: int) -> dict | None:
output = casadm.list_caches(OutputFormat.csv, by_id_path=True) output = casadm.list_caches(OutputFormat.csv, by_id_path=True)
reader = csv.DictReader(io.StringIO(output.stdout)) reader = csv.DictReader(io.StringIO(output.stdout))
cache_id = -1
for row in reader: for row in reader:
if row["type"] == "core" and row["disk"] == core_disk_path: if row["type"] == "cache":
cache_id = int(row["id"])
if row["type"] == "core" and row["disk"] == core_disk_path and target_cache_id == cache_id:
return { return {
"core_id": row["id"], "core_id": row["id"],
"core_device": row["disk"], "core_device": row["disk"],

View File

@ -11,7 +11,7 @@ from enum import Enum
from api.cas import casadm from api.cas import casadm
from api.cas.cache_config import SeqCutOffParameters, SeqCutOffPolicy from api.cas.cache_config import SeqCutOffParameters, SeqCutOffPolicy
from api.cas.casadm_params import StatsFilter from api.cas.casadm_params import StatsFilter
from api.cas.casadm_parser import get_seq_cut_off_parameters, get_core_info_by_path from api.cas.casadm_parser import get_seq_cut_off_parameters, get_core_info_for_cache_by_path
from api.cas.statistics import CoreStats, CoreIoClassStats from api.cas.statistics import CoreStats, CoreIoClassStats
from core.test_run_utils import TestRun from core.test_run_utils import TestRun
from storage_devices.device import Device from storage_devices.device import Device
@ -35,18 +35,19 @@ class Core(Device):
def __init__(self, core_device: str, cache_id: int): def __init__(self, core_device: str, cache_id: int):
self.core_device = Device(core_device) self.core_device = Device(core_device)
self.path = None self.path = None
self.cache_id = cache_id
core_info = self.__get_core_info() core_info = self.__get_core_info()
# "-" is special case for cores in core pool # "-" is special case for cores in core pool
if core_info["core_id"] != "-": if core_info["core_id"] != "-":
self.core_id = int(core_info["core_id"]) self.core_id = int(core_info["core_id"])
if core_info["exp_obj"] != "-": if core_info["exp_obj"] != "-":
Device.__init__(self, core_info["exp_obj"]) Device.__init__(self, core_info["exp_obj"])
self.cache_id = cache_id
self.partitions = [] self.partitions = []
self.block_size = None self.block_size = None
def __get_core_info(self): def __get_core_info(self):
return get_core_info_by_path(self.core_device.path) return get_core_info_for_cache_by_path(core_disk_path=self.core_device.path,
target_cache_id=self.cache_id)
def create_filesystem(self, fs_type: disk_utils.Filesystem, force=True, blocksize=None): def create_filesystem(self, fs_type: disk_utils.Filesystem, force=True, blocksize=None):
super().create_filesystem(fs_type, force, blocksize) super().create_filesystem(fs_type, force, blocksize)

View File

@ -1,6 +1,6 @@
# #
# Copyright(c) 2020-2022 Intel Corporation # Copyright(c) 2020-2022 Intel Corporation
# Copyright(c) 2024 Huawei Technologies Co., Ltd. # Copyright(c) 2024 Huawei Technologies
# SPDX-License-Identifier: BSD-3-Clause # SPDX-License-Identifier: BSD-3-Clause
# #
@ -11,13 +11,13 @@ from api.cas.cache_config import CacheMode, CleaningPolicy, CacheModeTrait
from api.cas.casadm_parser import wait_for_flushing from api.cas.casadm_parser import wait_for_flushing
from core.test_run import TestRun from core.test_run import TestRun
from storage_devices.disk import DiskType, DiskTypeSet, DiskTypeLowerThan from storage_devices.disk import DiskType, DiskTypeSet, DiskTypeLowerThan
from test_tools.disk_utils import Filesystem from test_tools import fs_utils
from test_tools.dd import Dd from test_tools.dd import Dd
from test_tools.disk_utils import Filesystem
from test_utils import os_utils from test_utils import os_utils
from test_utils.os_utils import Udev, DropCachesMode from test_utils.os_utils import Udev, DropCachesMode
from test_utils.size import Size, Unit from test_utils.size import Size, Unit
from tests.lazy_writes.recovery.recovery_tests_methods import compare_files from tests.lazy_writes.recovery.recovery_tests_methods import compare_files
from test_tools import fs_utils
mount_point = "/mnt/cas" mount_point = "/mnt/cas"
test_file_path = f"{mount_point}/test_file" test_file_path = f"{mount_point}/test_file"
@ -197,17 +197,17 @@ def test_interrupt_cache_flush(cache_mode, filesystem):
@pytest.mark.require_disk("core", DiskTypeLowerThan("cache")) @pytest.mark.require_disk("core", DiskTypeLowerThan("cache"))
def test_interrupt_core_remove(cache_mode, filesystem): def test_interrupt_core_remove(cache_mode, filesystem):
""" """
title: Test if OpenCAS works correctly after core's removing interruption. title: Core removal interruption.
description: | description: |
Negative test of the ability of OpenCAS to handle core's removing interruption. Test for proper handling of 'core remove' operation interruption.
pass_criteria: pass_criteria:
- No system crash. - No system crash.
- Core would not be removed from cache after interruption. - Core would not be removed from cache after interruption.
- Flushing would be stopped after interruption. - Flushing would be stopped after interruption.
- Md5sum are correct during all test steps. - Checksums are correct during all test steps.
- Dirty blocks quantity after interruption is lower but non-zero. - Dirty blocks quantity after interruption is lower but non-zero.
""" """
with TestRun.step("Prepare cache and core."): with TestRun.step("Prepare cache and core devices"):
cache_dev = TestRun.disks["cache"] cache_dev = TestRun.disks["cache"]
cache_dev.create_partitions([cache_size]) cache_dev.create_partitions([cache_size])
cache_part = cache_dev.partitions[0] cache_part = cache_dev.partitions[0]
@ -216,37 +216,36 @@ def test_interrupt_core_remove(cache_mode, filesystem):
core_part = core_dev.partitions[0] core_part = core_dev.partitions[0]
for _ in TestRun.iteration( for _ in TestRun.iteration(
range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times." range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times"
): ):
with TestRun.step("Start cache"):
with TestRun.step("Start cache."):
cache = casadm.start_cache(cache_part, cache_mode, force=True) cache = casadm.start_cache(cache_part, cache_mode, force=True)
with TestRun.step("Set cleaning policy to NOP."): with TestRun.step("Set cleaning policy to NOP"):
cache.set_cleaning_policy(CleaningPolicy.nop) cache.set_cleaning_policy(CleaningPolicy.nop)
with TestRun.step(f"Add core device with {filesystem} filesystem and mount it."): with TestRun.step(f"Add core device with {filesystem} filesystem and mount it"):
core_part.create_filesystem(filesystem) core_part.create_filesystem(filesystem)
core = cache.add_core(core_part) core = cache.add_core(core_part)
core.mount(mount_point) core.mount(mount_point)
with TestRun.step(f"Create test file in mount point of exported object."): with TestRun.step("Create test file in mount point of exported object"):
test_file = create_test_file() test_file = create_test_file()
with TestRun.step("Check md5 sum of test file."): with TestRun.step("Calculate checksum of test file"):
test_file_md5sum_before = test_file.md5sum() test_file_crc32sum_before = test_file.crc32sum()
with TestRun.step( with TestRun.step(
"Get number of dirty data on exported object before core removal interruption." "Get number of dirty data on exported object before core removal interruption"
): ):
os_utils.sync() os_utils.sync()
os_utils.drop_caches(DropCachesMode.ALL) os_utils.drop_caches(DropCachesMode.ALL)
cache_dirty_blocks_before = cache.get_dirty_blocks() cache_dirty_blocks_before = cache.get_dirty_blocks()
with TestRun.step("Unmount core."): with TestRun.step("Unmount core"):
core.unmount() core.unmount()
with TestRun.step("Start removing core device."): with TestRun.step("Start removing core"):
flush_pid = TestRun.executor.run_in_background( flush_pid = TestRun.executor.run_in_background(
cli.remove_core_cmd(str(cache.cache_id), str(core.core_id)) cli.remove_core_cmd(str(cache.cache_id), str(core.core_id))
) )
@ -258,42 +257,39 @@ def test_interrupt_core_remove(cache_mode, filesystem):
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id) percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
TestRun.executor.run(f"kill -s SIGINT {flush_pid}") TestRun.executor.run(f"kill -s SIGINT {flush_pid}")
with TestRun.step("Check md5 sum of test file after interruption."):
cache.set_cache_mode(CacheMode.WO)
test_file_md5sum_interrupt = test_file.md5sum()
cache.set_cache_mode(cache_mode)
with TestRun.step( with TestRun.step(
"Check number of dirty data on exported object after core removal interruption." "Check number of dirty data on exported object after core removal interruption"
): ):
cache_dirty_blocks_after = cache.get_dirty_blocks() cache_dirty_blocks_after = cache.get_dirty_blocks()
if cache_dirty_blocks_after >= cache_dirty_blocks_before: if cache_dirty_blocks_after >= cache_dirty_blocks_before:
TestRun.LOGGER.error( TestRun.LOGGER.error(
"Quantity of dirty lines after core removal interruption " "should be lower." "Quantity of dirty lines after core removal interruption should be lower."
) )
if int(cache_dirty_blocks_after) == 0: if int(cache_dirty_blocks_after) == 0:
TestRun.LOGGER.error( TestRun.LOGGER.error(
"Quantity of dirty lines after core removal interruption " "should not be zero." "Quantity of dirty lines after core removal interruption should not be zero."
) )
with TestRun.step("Remove core from cache."): with TestRun.step("Mount core and verify test file checksum after interruption"):
core.remove_core() core.mount(mount_point)
with TestRun.step("Stop cache."): if test_file.crc32sum() != test_file_crc32sum_before:
TestRun.LOGGER.error("Checksum after interrupting core removal is different.")
with TestRun.step("Unmount core"):
core.unmount()
with TestRun.step("Stop cache"):
cache.stop() cache.stop()
with TestRun.step("Mount core device."): with TestRun.step("Mount core device"):
core_part.mount(mount_point) core_part.mount(mount_point)
with TestRun.step("Check md5 sum of test file again."): with TestRun.step("Verify checksum of test file again"):
if test_file_md5sum_before != test_file.md5sum(): if test_file.crc32sum() != test_file_crc32sum_before:
TestRun.LOGGER.error("Md5 sum before interrupting core removal is different.") TestRun.LOGGER.error("Checksum after core removal is different.")
is_sum_diff_after_interrupt = test_file_md5sum_interrupt != test_file.md5sum() with TestRun.step("Unmount core device"):
if is_sum_diff_after_interrupt:
TestRun.LOGGER.error("Md5 sum after interrupting core removal is different.")
with TestRun.step("Unmount core device."):
core_part.unmount() core_part.unmount()
@ -315,8 +311,19 @@ def test_interrupt_cache_mode_switch_parametrized(cache_mode, stop_percentage):
- Md5sum are correct during all test steps. - Md5sum are correct during all test steps.
- Dirty blocks quantity after interruption is lower but non-zero. - Dirty blocks quantity after interruption is lower but non-zero.
""" """
test_file_size = Size(1, Unit.GibiByte)
test_file_path = "/mnt/cas/test_file"
with TestRun.step("Prepare cache and core."): with TestRun.step("Prepare cache and core."):
cache_part, core_part = prepare() cache_dev = TestRun.disks["cache"]
core_dev = TestRun.disks["core"]
cache_dev.create_partitions([cache_size])
core_dev.create_partitions([cache_size * 2])
cache_part = cache_dev.partitions[0]
core_part = core_dev.partitions[0]
with TestRun.step("Disable udev"): with TestRun.step("Disable udev"):
Udev.disable() Udev.disable()
@ -325,23 +332,22 @@ def test_interrupt_cache_mode_switch_parametrized(cache_mode, stop_percentage):
range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times." range(iterations_per_config), f"Reload cache configuration {iterations_per_config} times."
): ):
with TestRun.step("Start cache."): with TestRun.step("Start cache"):
cache = casadm.start_cache(cache_part, cache_mode, force=True) cache = casadm.start_cache(cache_part, cache_mode, force=True)
with TestRun.step("Set cleaning policy to NOP."): with TestRun.step("Set cleaning policy to NOP"):
cache.set_cleaning_policy(CleaningPolicy.nop) cache.set_cleaning_policy(CleaningPolicy.nop)
with TestRun.step(f"Add core device."): with TestRun.step(f"Add core device"):
core = cache.add_core(core_part) core = cache.add_core(core_part)
with TestRun.step(f"Create test file in mount point of exported object."): with TestRun.step(f"Create test file in mount point of exported object"):
test_file_size = Size(4, Unit.GibiByte)
test_file = fs_utils.create_random_test_file(test_file_path, test_file_size) test_file = fs_utils.create_random_test_file(test_file_path, test_file_size)
with TestRun.step("Check md5 sum of test file."): with TestRun.step("Calculate md5sum of test file"):
test_file_md5_before = test_file.md5sum() test_file_md5_before = test_file.md5sum()
with TestRun.step("Export file to CAS"): with TestRun.step("Copy test data to core"):
dd = ( dd = (
Dd() Dd()
.block_size(test_file_size) .block_size(test_file_size)
@ -351,49 +357,58 @@ def test_interrupt_cache_mode_switch_parametrized(cache_mode, stop_percentage):
) )
dd.run() dd.run()
with TestRun.step("Get number of dirty data on exported object before interruption."): with TestRun.step("Get number of dirty data on exported object before interruption"):
os_utils.sync() os_utils.sync()
os_utils.drop_caches(DropCachesMode.ALL) os_utils.drop_caches(DropCachesMode.ALL)
cache_dirty_blocks_before = cache.get_dirty_blocks() cache_dirty_blocks_before = cache.get_dirty_blocks()
with TestRun.step("Start switching cache mode."): with TestRun.step("Start switching cache mode"):
flush_pid = TestRun.executor.run_in_background( flush_pid = TestRun.executor.run_in_background(
cli.set_cache_mode_cmd( cli.set_cache_mode_cmd(
str(CacheMode.DEFAULT.name.lower()), str(cache.cache_id), "yes" cache_mode=str(CacheMode.DEFAULT.name.lower()),
cache_id=str(cache.cache_id),
flush_cache="yes",
) )
) )
with TestRun.step("Send interruption signal."): with TestRun.step("Kill flush process during cache flush operation"):
wait_for_flushing(cache, core) wait_for_flushing(cache, core)
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id) percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
while percentage < stop_percentage: while percentage < stop_percentage:
percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id) percentage = casadm_parser.get_flushing_progress(cache.cache_id, core.core_id)
TestRun.executor.run(f"kill -s SIGINT {flush_pid}") TestRun.executor.kill_process(flush_pid)
with TestRun.step("Check number of dirty data on exported object after interruption."): with TestRun.step("Check number of dirty data on exported object after interruption"):
cache_dirty_blocks_after = cache.get_dirty_blocks() cache_dirty_blocks_after = cache.get_dirty_blocks()
if cache_dirty_blocks_after >= cache_dirty_blocks_before: if cache_dirty_blocks_after >= cache_dirty_blocks_before:
TestRun.LOGGER.error( TestRun.LOGGER.error(
"Quantity of dirty lines after cache mode switching " "Quantity of dirty lines after cache mode switching "
"interruption should be lower." "interruption should be lower."
) )
if int(cache_dirty_blocks_after) == 0: if cache_dirty_blocks_after == Size.zero():
TestRun.LOGGER.error( TestRun.LOGGER.error(
"Quantity of dirty lines after cache mode switching " "Quantity of dirty lines after cache mode switching "
"interruption should not be zero." "interruption should not be zero."
) )
with TestRun.step("Check cache mode."): with TestRun.step("Check cache mode"):
if cache.get_cache_mode() != cache_mode: if cache.get_cache_mode() != cache_mode:
TestRun.LOGGER.error("Cache mode should remain the same.") TestRun.LOGGER.error("Cache mode should remain the same.")
with TestRun.step("Stop cache."): with TestRun.step("Stop cache"):
cache.stop() cache.stop()
with TestRun.step("Check md5 sum of test file again."): with TestRun.step("Copy test data from the exported object to a file"):
Dd().block_size(test_file_size).input(core.path).output(test_file.full_path).oflag( dd = (
"direct" Dd()
).run() .block_size(test_file_size)
.input(core.path)
.output(test_file.full_path)
.oflag("direct")
)
dd.run()
with TestRun.step("Compare md5 sum of test files"):
target_file_md5 = test_file.md5sum() target_file_md5 = test_file.md5sum()
compare_files(test_file_md5_before, target_file_md5) compare_files(test_file_md5_before, target_file_md5)
@ -491,7 +506,7 @@ def create_test_file():
bs = Size(512, Unit.KibiByte) bs = Size(512, Unit.KibiByte)
cnt = int(cache_size.value / bs.value) cnt = int(cache_size.value / bs.value)
test_file = File.create_file(test_file_path) test_file = File.create_file(test_file_path)
dd = Dd().output(test_file_path).input("/dev/zero").block_size(bs).count(cnt) dd = Dd().output(test_file_path).input("/dev/zero").block_size(bs).count(cnt).oflag("direct")
dd.run() dd.run()
test_file.refresh_item() test_file.refresh_item()
return test_file return test_file

View File

@ -12,7 +12,7 @@ from core.test_run import TestRun
from test_utils.size import Size, Unit from test_utils.size import Size, Unit
from api.cas.cache_config import CacheLineSize, CacheMode, CacheStatus from api.cas.cache_config import CacheLineSize, CacheMode, CacheStatus
from api.cas.casadm_params import StatsFilter from api.cas.casadm_params import StatsFilter
from api.cas.casadm_parser import get_core_info_by_path from api.cas.casadm_parser import get_core_info_for_cache_by_path
from api.cas.core import CoreStatus, Core from api.cas.core import CoreStatus, Core
from test_tools.dd import Dd from test_tools.dd import Dd
from api.cas.cli import standby_activate_cmd from api.cas.cli import standby_activate_cmd
@ -173,7 +173,11 @@ def test_activate_incomplete_cache():
TestRun.fail(f"Expected one inactive core. Got {inactive_core_count}") TestRun.fail(f"Expected one inactive core. Got {inactive_core_count}")
with TestRun.step("Check if core is in an appropriate state"): with TestRun.step("Check if core is in an appropriate state"):
core_status = CoreStatus[get_core_info_by_path(core_dev_path)["status"].lower()] core_status = CoreStatus[
get_core_info_for_cache_by_path(
core_disk_path=core_dev_path, target_cache_id=cache.cache_id
)["status"].lower()
]
if core_status != CoreStatus.inactive: if core_status != CoreStatus.inactive:
TestRun.fail( TestRun.fail(
"The core is in an invalid state. " "The core is in an invalid state. "

View File

@ -1,11 +1,13 @@
# #
# Copyright(c) 2020-2021 Intel Corporation # Copyright(c) 2020-2021 Intel Corporation
# Copyright(c) 2024 Huawei Technologies Co., Ltd.
# SPDX-License-Identifier: BSD-3-Clause # SPDX-License-Identifier: BSD-3-Clause
# #
import math import math
import posixpath
import pytest import pytest
import os
from api.cas import casadm, cli_messages from api.cas import casadm, cli_messages
from api.cas.cache_config import CacheLineSize from api.cas.cache_config import CacheLineSize
from core.test_run import TestRun from core.test_run import TestRun
@ -32,7 +34,7 @@ def test_device_capabilities():
""" """
core_device = TestRun.disks['core'] core_device = TestRun.disks['core']
max_io_size_path = os.path.join(disk_utils.get_sysfs_path(core_device.get_device_id()), max_io_size_path = posixpath.join(disk_utils.get_sysfs_path(core_device.get_device_id()),
'queue/max_sectors_kb') 'queue/max_sectors_kb')
default_max_io_size = fs_utils.read_file(max_io_size_path) default_max_io_size = fs_utils.read_file(max_io_size_path)
@ -148,7 +150,7 @@ def measure_capabilities(dev):
dev_id = dev.parent_device.get_device_id() if isinstance(dev, Partition) \ dev_id = dev.parent_device.get_device_id() if isinstance(dev, Partition) \
else dev.get_device_id() else dev.get_device_id()
for c in capabilities: for c in capabilities:
path = os.path.join(disk_utils.get_sysfs_path(dev_id), 'queue', c) path = posixpath.join(disk_utils.get_sysfs_path(dev_id), 'queue', c)
command = f"cat {path}" command = f"cat {path}"
output = TestRun.executor.run(command) output = TestRun.executor.run(command)
if output.exit_code == 0: if output.exit_code == 0: