Merge pull request #230 from Ostrokrzew/init2

Cache configuration load - every mode - 1 cache 1 or 4 cores
This commit is contained in:
Daniel Madej 2020-01-02 13:03:46 +01:00 committed by GitHub
commit 0d663a37f1
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
3 changed files with 390 additions and 2 deletions

View File

@ -95,6 +95,12 @@ class FlushParametersAlru:
self.staleness_time = staleness_time self.staleness_time = staleness_time
self.wake_up_time = wake_up_time self.wake_up_time = wake_up_time
def __eq__(self, other):
return self.activity_threshold == other.activity_threshold and \
self.flush_max_buffers == other.flush_max_buffers and \
self.staleness_time == other.staleness_time and \
self.wake_up_time == other.wake_up_time
@staticmethod @staticmethod
def default_alru_params(): def default_alru_params():
alru_params = FlushParametersAlru() alru_params = FlushParametersAlru()
@ -112,6 +118,10 @@ class FlushParametersAcp:
self.flush_max_buffers = flush_max_buffers self.flush_max_buffers = flush_max_buffers
self.wake_up_time = wake_up_time self.wake_up_time = wake_up_time
def __eq__(self, other):
return self.flush_max_buffers == other.flush_max_buffers and \
self.wake_up_time == other.wake_up_time
@staticmethod @staticmethod
def default_acp_params(): def default_acp_params():
acp_params = FlushParametersAcp() acp_params = FlushParametersAcp()

View File

@ -214,7 +214,7 @@ def get_cas_devices_dict():
def get_flush_parameters_alru(cache_id: int): def get_flush_parameters_alru(cache_id: int):
casadm_output = casadm.get_param_cleaning_alru(cache_id, casadm_output = casadm.get_param_cleaning_alru(cache_id,
casadm.OutputFormat.csv).stdout.spltlines() casadm.OutputFormat.csv).stdout.splitlines()
flush_parameters = FlushParametersAlru() flush_parameters = FlushParametersAlru()
for line in casadm_output: for line in casadm_output:
if 'max buffers' in line: if 'max buffers' in line:
@ -222,7 +222,7 @@ def get_flush_parameters_alru(cache_id: int):
if 'Activity threshold' in line: if 'Activity threshold' in line:
flush_parameters.activity_threshold = Time(milliseconds=int(line.split(',')[1])) flush_parameters.activity_threshold = Time(milliseconds=int(line.split(',')[1]))
if 'Stale buffer time' in line: if 'Stale buffer time' in line:
flush_parameters.staneless_time = Time(seconds=int(line.split(',')[1])) flush_parameters.staleness_time = Time(seconds=int(line.split(',')[1]))
if 'Wake up time' in line: if 'Wake up time' in line:
flush_parameters.wake_up_time = Time(seconds=int(line.split(',')[1])) flush_parameters.wake_up_time = Time(seconds=int(line.split(',')[1]))
return flush_parameters return flush_parameters

View File

@ -0,0 +1,378 @@
#
# Copyright(c) 2019 Intel Corporation
# SPDX-License-Identifier: BSD-3-Clause-Clear
#
import pytest
from api.cas import casadm, casadm_parser
from api.cas.cache_config import *
from storage_devices.disk import DiskType, DiskTypeSet, DiskTypeLowerThan
from test_tools.fio.fio import Fio
from test_tools.fio.fio_param import *
from test_utils.size import Size, Unit
@pytest.mark.parametrize("core_number", [4, 1])
@pytest.mark.parametrize("cleaning_policy", CleaningPolicy)
@pytest.mark.parametrize("cache_mode", CacheMode)
@pytest.mark.parametrize("cache_line_size", CacheLineSize)
@pytest.mark.require_disk("cache", DiskTypeSet([DiskType.optane, DiskType.nand]))
@pytest.mark.require_disk("core", DiskTypeLowerThan("cache"))
def test_load_x_to_one_without_params(cache_mode, cleaning_policy, cache_line_size, core_number):
"""
title: Test for loading CAS with 1 cache and 1 or 4 cores without extra params.
description: |
Verify that loading cache configurations works properly in every mode
with 1 cache and 1 or 4 cores. Use minimal number of possible parameters
with load command.
pass_criteria:
- Cache loads successfully.
- No errors in cache are found.
"""
with TestRun.step(f"Prepare 1 cache and {core_number} core devices"):
cache_dev = TestRun.disks['cache']
cache_dev.create_partitions([Size(512, Unit.MebiByte)])
cache_dev = cache_dev.partitions[0]
core_dev = TestRun.disks['core']
core_size = []
for i in range(core_number):
core_size.append(Size(1, Unit.GibiByte))
core_dev.create_partitions(core_size)
with TestRun.step(f"Start cache with {core_number} cores."):
cache = casadm.start_cache(cache_dev, cache_mode, cache_line_size, force=True)
cores = []
for i in range(core_number):
cores.append(cache.add_core(core_dev.partitions[i]))
caches_count = len(casadm_parser.get_caches())
if caches_count != 1:
TestRun.fail(f"Expected caches count: 1; Actual caches count: {caches_count}.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != core_number:
TestRun.fail(f"Expected cores count: {core_number}; Actual cores count: {cores_count}.")
with TestRun.step("Configure cleaning policy."):
cache.set_cleaning_policy(cleaning_policy)
if cleaning_policy == CleaningPolicy.alru:
alru = FlushParametersAlru()
alru.activity_threshold = Time(milliseconds=1000)
alru.flush_max_buffers = 10
alru.staleness_time = Time(seconds=60)
alru.wake_up_time = Time(seconds=5)
cache.set_params_alru(alru)
if cleaning_policy == CleaningPolicy.acp:
acp = FlushParametersAcp()
acp.flush_max_buffers = 100
acp.wake_up_time = Time(seconds=5)
cache.set_params_acp(acp)
with TestRun.step("Run FIO on exported object"):
for core in cores:
Fio().create_command() \
.io_engine(IoEngine.libaio) \
.io_depth(64) \
.size(Size(1, Unit.GibiByte)) \
.read_write(ReadWrite.randrw) \
.block_size(cache_line_size) \
.target(f"{core.system_path}") \
.run()
with TestRun.step("Stop cache."):
cache.stop()
caches_count = len(casadm_parser.get_caches())
if caches_count != 0:
TestRun.fail(f"There are still {caches_count} caches running after stopping service.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != 0:
TestRun.fail(f"There are still {cores_count} cores running after stopping service.")
with TestRun.step("Load cache."):
cache = casadm.load_cache(cache_dev)
caches_count = len(casadm_parser.get_caches())
if caches_count != 1:
TestRun.fail(f"Expected caches count: 1; Actual caches count: {caches_count}.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != core_number:
TestRun.fail(f"Expected cores count: {core_number}; Actual cores count: {cores_count}.")
with TestRun.step("Compare cache configuration before and after load."):
if cache_mode != cache.get_cache_mode():
TestRun.fail("Cache modes are different.")
if cache_line_size != cache.get_cache_line_size():
TestRun.fail("Cache line sizes are different.")
if cleaning_policy != cache.get_cleaning_policy():
TestRun.fail("Cleaning policies are different.")
if cleaning_policy == CleaningPolicy.alru:
if alru != cache.get_flush_parameters_alru():
TestRun.fail("Cleaning policy parameters are different.")
if cleaning_policy == CleaningPolicy.acp:
if acp != cache.get_flush_parameters_acp():
TestRun.fail("Cleaning policy parameters are different.")
with TestRun.step("Run FIO again on exported object"):
for core in cores:
Fio().create_command() \
.io_engine(IoEngine.libaio) \
.io_depth(64) \
.size(Size(1, Unit.GibiByte)) \
.read_write(ReadWrite.randrw) \
.block_size(cache_line_size) \
.target(f"{core.system_path}") \
.run()
with TestRun.step("Check if there are no error statistics."):
if cache.get_statistics().error_stats.total_errors != 0:
TestRun.fail("There are errors in the cache.")
with TestRun.step("Stop caches."):
casadm.stop_all_caches()
@pytest.mark.parametrize("core_number", [1, 4])
@pytest.mark.parametrize("cleaning_policy", CleaningPolicy)
@pytest.mark.parametrize("cache_mode", CacheMode)
@pytest.mark.parametrize("cache_line_size", CacheLineSize)
@pytest.mark.require_disk("cache", DiskTypeSet([DiskType.optane, DiskType.nand]))
@pytest.mark.require_disk("core", DiskTypeLowerThan("cache"))
def test_load_x_to_one_with_params(cache_mode, cleaning_policy, cache_line_size, core_number):
"""
title: Test for loading CAS with 1 cache and 1 or 4 cores with maximum extra params.
description: |
Verify that loading cache configurations works properly in every mode
with 1 cache and 1 or 4 cores. Use maximal number of possible parameters
with load command.
pass_criteria:
- Cache loads successfully.
- No errors in cache are found.
"""
with TestRun.step(f"Prepare 1 cache and {core_number} core devices"):
cache_dev = TestRun.disks['cache']
cache_dev.create_partitions([Size(512, Unit.MebiByte)])
cache_dev = cache_dev.partitions[0]
core_dev = TestRun.disks['core']
core_size = []
for i in range(core_number):
core_size.append(Size(1, Unit.GibiByte))
core_dev.create_partitions(core_size)
with TestRun.step(f"Start cache with {core_number} cores."):
cache = casadm.start_cache(cache_dev, cache_mode, cache_line_size, force=True)
id_cache = cache.cache_id
cores = []
for i in range(core_number):
cores.append(cache.add_core(core_dev.partitions[i]))
caches_count = len(casadm_parser.get_caches())
if caches_count != 1:
TestRun.fail(f"Expected caches count: 1; Actual caches count: {caches_count}.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != core_number:
TestRun.fail(f"Expected cores count: {core_number}; Actual cores count: {cores_count}.")
with TestRun.step("Configure cleaning policy."):
cache.set_cleaning_policy(cleaning_policy)
if cleaning_policy == CleaningPolicy.alru:
alru = FlushParametersAlru()
alru.activity_threshold = Time(milliseconds=1000)
alru.flush_max_buffers = 10
alru.staleness_time = Time(seconds=60)
alru.wake_up_time = Time(seconds=5)
cache.set_params_alru(alru)
if cleaning_policy == CleaningPolicy.acp:
acp = FlushParametersAcp()
acp.flush_max_buffers = 100
acp.wake_up_time = Time(seconds=5)
cache.set_params_acp(acp)
with TestRun.step("Run FIO on exported object"):
for core in cores:
Fio().create_command() \
.io_engine(IoEngine.libaio) \
.io_depth(64) \
.size(Size(1, Unit.GibiByte)) \
.read_write(ReadWrite.randrw) \
.block_size(cache_line_size) \
.target(f"{core.system_path}") \
.run()
with TestRun.step("Stop cache."):
cache.stop()
caches_count = len(casadm_parser.get_caches())
if caches_count != 0:
TestRun.fail(f"There are still {caches_count} caches running after stopping service.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != 0:
TestRun.fail(f"There are still {cores_count} cores running after stopping service.")
with TestRun.step("Load cache."):
cache = casadm.start_cache(cache_dev, cache_mode, cache_line_size, id_cache, load=True)
caches_count = len(casadm_parser.get_caches())
if caches_count != 1:
TestRun.fail(f"Expected caches count: 1; Actual caches count: {caches_count}.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != core_number:
TestRun.fail(f"Expected cores count: {core_number}; Actual cores count: {cores_count}.")
with TestRun.step("Compare cache configuration before and after load."):
if cache_mode != cache.get_cache_mode():
TestRun.fail("Cache modes are different.")
if cache_line_size != cache.get_cache_line_size():
TestRun.fail("Cache line sizes are different.")
if cleaning_policy != cache.get_cleaning_policy():
TestRun.fail("Cleaning policies are different.")
if cleaning_policy == CleaningPolicy.alru:
if alru != cache.get_flush_parameters_alru():
TestRun.fail("Cleaning policy parameters are different.")
if cleaning_policy == CleaningPolicy.acp:
if acp != cache.get_flush_parameters_acp():
TestRun.fail("Cleaning policy parameters are different.")
with TestRun.step("Run FIO again on exported object"):
for core in cores:
Fio().create_command() \
.io_engine(IoEngine.libaio) \
.io_depth(64) \
.size(Size(1, Unit.GibiByte)) \
.read_write(ReadWrite.randrw) \
.block_size(cache_line_size) \
.target(f"{core.system_path}") \
.run()
with TestRun.step("Check if there are no error statistics."):
if cache.get_statistics().error_stats.total_errors != 0:
TestRun.fail("There are errors in the cache.")
with TestRun.step("Stop caches."):
casadm.stop_all_caches()
@pytest.mark.parametrize("core_number", [1, 4])
@pytest.mark.parametrize("cleaning_policy", CleaningPolicy)
@pytest.mark.parametrize("cache_mode", [[CacheMode.WT, CacheMode.WB],
[CacheMode.WB, CacheMode.WA],
[CacheMode.WA, CacheMode.PT],
[CacheMode.PT, CacheMode.WO],
[CacheMode.WO, CacheMode.WT]])
@pytest.mark.parametrize("cache_line_size", [[CacheLineSize.LINE_4KiB, CacheLineSize.LINE_8KiB],
[CacheLineSize.LINE_8KiB, CacheLineSize.LINE_16KiB],
[CacheLineSize.LINE_16KiB, CacheLineSize.LINE_32KiB],
[CacheLineSize.LINE_32KiB, CacheLineSize.LINE_64KiB],
[CacheLineSize.LINE_64KiB, CacheLineSize.LINE_4KiB]])
@pytest.mark.require_disk("cache", DiskTypeSet([DiskType.optane, DiskType.nand]))
@pytest.mark.require_disk("core", DiskTypeLowerThan("cache"))
def test_load_x_to_one_diff_params(cache_mode, cleaning_policy, cache_line_size, core_number):
f"""
title: Test for loading CAS with 1 cache and 1 or 4 cores with different params.
description: |
Verify that loading cache configurations works properly in every mode
with 1 cache and 1 or 4 cores. Use different parameters with load command.
pass_criteria:
- OpenCAS should load successfully but with saved configuration.
- No errors in cache are found.
"""
with TestRun.step(f"Prepare 1 cache and {core_number} core devices"):
cache_dev = TestRun.disks['cache']
cache_dev.create_partitions([Size(512, Unit.MebiByte)])
cache_dev = cache_dev.partitions[0]
core_dev = TestRun.disks['core']
core_size = []
for i in range(core_number):
core_size.append(Size(1, Unit.GibiByte))
core_dev.create_partitions(core_size)
with TestRun.step(f"Start cache with {core_number} cores."):
cache = casadm.start_cache(cache_dev, cache_mode[0], cache_line_size[0], force=True)
id_cache = cache.cache_id
cores = []
for i in range(core_number):
cores.append(cache.add_core(core_dev.partitions[i]))
caches_count = len(casadm_parser.get_caches())
if caches_count != 1:
TestRun.fail(f"Expected caches count: 1; Actual caches count: {caches_count}.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != core_number:
TestRun.fail(f"Expected cores count: {core_number}; Actual cores count: {cores_count}.")
with TestRun.step("Configure cleaning policy."):
cache.set_cleaning_policy(cleaning_policy)
if cleaning_policy == CleaningPolicy.alru:
alru = FlushParametersAlru()
alru.activity_threshold = Time(milliseconds=1000)
alru.flush_max_buffers = 10
alru.staleness_time = Time(seconds=60)
alru.wake_up_time = Time(seconds=5)
cache.set_params_alru(alru)
if cleaning_policy == CleaningPolicy.acp:
acp = FlushParametersAcp()
acp.flush_max_buffers = 100
acp.wake_up_time = Time(seconds=5)
cache.set_params_acp(acp)
with TestRun.step("Run FIO on exported object"):
for core in cores:
Fio().create_command() \
.io_engine(IoEngine.libaio) \
.io_depth(64) \
.size(Size(1, Unit.GibiByte)) \
.read_write(ReadWrite.randrw) \
.block_size(cache_line_size[0]) \
.target(f"{core.system_path}") \
.run()
with TestRun.step("Stop cache."):
cache.stop()
caches_count = len(casadm_parser.get_caches())
if caches_count != 0:
TestRun.fail(f"There are still {caches_count} caches running after stopping service.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != 0:
TestRun.fail(f"There are still {cores_count} cores running after stopping service.")
with TestRun.step("Load cache."):
try:
cache = casadm.start_cache(cache_dev, cache_mode[1], cache_line_size[1],
id_cache + 1, load=True)
except Exception:
TestRun.LOGGER.error("Cannot pass other cache ID to cache load.")
TestRun.LOGGER.info("Load cache without passing cache ID to command.")
cache = casadm.start_cache(cache_dev, cache_mode[1], cache_line_size[1], load=True)
caches_count = len(casadm_parser.get_caches())
if caches_count != 1:
TestRun.fail(f"Expected caches count: 1; Actual caches count: {caches_count}.")
cores_count = len(casadm_parser.get_cores(cache.cache_id))
if cores_count != core_number:
TestRun.fail(f"Expected cores count: {core_number}; Actual cores count: {cores_count}.")
with TestRun.step("Compare cache configuration before and after load."):
if cache_mode[0] != cache.get_cache_mode():
TestRun.fail("Cache modes are different. Should be the same.")
if cache_line_size[0] != cache.get_cache_line_size():
TestRun.fail("Cache line sizes are different. Should be the same.")
if id_cache != cache.cache_id:
TestRun.fail("Cache IDs are different. Should be the same.")
if cleaning_policy != cache.get_cleaning_policy():
TestRun.fail("Cleaning policies are different.")
if cleaning_policy == CleaningPolicy.alru:
if alru != cache.get_flush_parameters_alru():
TestRun.fail("Cleaning policy parameters are different.")
if cleaning_policy == CleaningPolicy.acp:
if acp != cache.get_flush_parameters_acp():
TestRun.fail("Cleaning policy parameters are different.")
with TestRun.step("Run FIO again on exported object"):
for core in cores:
Fio().create_command() \
.io_engine(IoEngine.libaio) \
.io_depth(64) \
.size(Size(1, Unit.GibiByte)) \
.read_write(ReadWrite.randrw) \
.block_size(cache_line_size[1]) \
.target(f"{core.system_path}") \
.run()
with TestRun.step("Check if there are no error statistics."):
if cache.get_statistics().error_stats.total_errors != 0:
TestRun.fail("There are errors in the cache.")
with TestRun.step("Stop caches."):
casadm.stop_all_caches()