/* * Copyright(c) 2019-2021 Intel Corporation * SPDX-License-Identifier: BSD-3-Clause-Clear */ #include "../ocf_cache_priv.h" #include "../eviction/eviction.h" #ifndef __OCF_METADATA_CONCURRENCY_H__ #define __OCF_METADATA_CONCURRENCY_H__ #define OCF_METADATA_RD 0 #define OCF_METADATA_WR 1 int ocf_metadata_concurrency_init(struct ocf_metadata_lock *metadata_lock); void ocf_metadata_concurrency_deinit(struct ocf_metadata_lock *metadata_lock); int ocf_metadata_concurrency_attached_init( struct ocf_metadata_lock *metadata_lock, ocf_cache_t cache, uint32_t hash_table_entries, uint32_t colision_table_pages); void ocf_metadata_concurrency_attached_deinit( struct ocf_metadata_lock *metadata_lock); static inline void ocf_metadata_eviction_lock( struct ocf_metadata_lock *metadata_lock, unsigned ev_list) { env_spinlock_lock(&metadata_lock->eviction[ev_list]); } static inline void ocf_metadata_eviction_unlock( struct ocf_metadata_lock *metadata_lock, unsigned ev_list) { env_spinlock_unlock(&metadata_lock->eviction[ev_list]); } static inline void ocf_metadata_eviction_lock_all( struct ocf_metadata_lock *metadata_lock) { uint32_t i; for (i = 0; i < OCF_NUM_EVICTION_LISTS; i++) ocf_metadata_eviction_lock(metadata_lock, i); } static inline void ocf_metadata_eviction_unlock_all( struct ocf_metadata_lock *metadata_lock) { uint32_t i; for (i = 0; i < OCF_NUM_EVICTION_LISTS; i++) ocf_metadata_eviction_unlock(metadata_lock, i); } #define OCF_METADATA_EVICTION_LOCK(cline) \ ocf_metadata_eviction_lock(&cache->metadata.lock, \ cline % OCF_NUM_EVICTION_LISTS) #define OCF_METADATA_EVICTION_UNLOCK(cline) \ ocf_metadata_eviction_unlock(&cache->metadata.lock, \ cline % OCF_NUM_EVICTION_LISTS) #define OCF_METADATA_EVICTION_LOCK_ALL() \ ocf_metadata_eviction_lock_all(&cache->metadata.lock) #define OCF_METADATA_EVICTION_UNLOCK_ALL() \ ocf_metadata_eviction_unlock_all(&cache->metadata.lock) static inline void ocf_metadata_partition_lock( struct ocf_metadata_lock *metadata_lock, ocf_part_id_t part_id) { env_spinlock_lock(&metadata_lock->partition[part_id]); } static inline void ocf_metadata_partition_unlock( struct ocf_metadata_lock *metadata_lock, ocf_part_id_t part_id) { env_spinlock_unlock(&metadata_lock->partition[part_id]); } void ocf_metadata_start_exclusive_access( struct ocf_metadata_lock *metadata_lock); int ocf_metadata_try_start_exclusive_access( struct ocf_metadata_lock *metadata_lock); void ocf_metadata_end_exclusive_access( struct ocf_metadata_lock *metadata_lock); int ocf_metadata_try_start_shared_access( struct ocf_metadata_lock *metadata_lock); void ocf_metadata_start_shared_access( struct ocf_metadata_lock *metadata_lock); void ocf_metadata_end_shared_access( struct ocf_metadata_lock *metadata_lock); static inline void ocf_metadata_status_bits_lock( struct ocf_metadata_lock *metadata_lock, int rw) { if (rw == OCF_METADATA_WR) env_rwlock_write_lock(&metadata_lock->status); else if (rw == OCF_METADATA_RD) env_rwlock_read_lock(&metadata_lock->status); else ENV_BUG(); } static inline void ocf_metadata_status_bits_unlock( struct ocf_metadata_lock *metadata_lock, int rw) { if (rw == OCF_METADATA_WR) env_rwlock_write_unlock(&metadata_lock->status); else if (rw == OCF_METADATA_RD) env_rwlock_read_unlock(&metadata_lock->status); else ENV_BUG(); } #define OCF_METADATA_BITS_LOCK_RD() \ ocf_metadata_status_bits_lock(&cache->metadata.lock, \ OCF_METADATA_RD) #define OCF_METADATA_BITS_UNLOCK_RD() \ ocf_metadata_status_bits_unlock(&cache->metadata.lock, \ OCF_METADATA_RD) #define OCF_METADATA_BITS_LOCK_WR() \ ocf_metadata_status_bits_lock(&cache->metadata.lock, \ OCF_METADATA_WR) #define OCF_METADATA_BITS_UNLOCK_WR() \ ocf_metadata_status_bits_unlock(&cache->metadata.lock, \ OCF_METADATA_WR) void ocf_hb_cline_prot_lock_rd(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); void ocf_hb_cline_prot_unlock_rd(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); void ocf_hb_id_prot_lock_wr(struct ocf_metadata_lock *metadata_lock, ocf_cache_line_t hash); void ocf_hb_id_prot_unlock_wr(struct ocf_metadata_lock *metadata_lock, ocf_cache_line_t hash); /* caller must hold global metadata read lock */ bool ocf_hb_cline_naked_trylock_rd(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); void ocf_hb_cline_naked_unlock_rd(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); void ocf_hb_cline_prot_lock_wr(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); void ocf_hb_cline_prot_unlock_wr(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); /* caller must hold global metadata read lock */ bool ocf_hb_cline_naked_trylock_wr(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); void ocf_hb_cline_naked_unlock_wr(struct ocf_metadata_lock *metadata_lock, uint32_t core_id, uint64_t core_line); /* lock entire request in deadlock-free manner */ void ocf_hb_req_prot_lock_rd(struct ocf_request *req); void ocf_hb_req_prot_unlock_rd(struct ocf_request *req); void ocf_hb_req_prot_lock_wr(struct ocf_request *req); void ocf_hb_req_prot_unlock_wr(struct ocf_request *req); void ocf_hb_req_prot_lock_upgrade(struct ocf_request *req); /* collision table page lock interface */ void ocf_collision_start_shared_access(struct ocf_metadata_lock *metadata_lock, uint32_t page); void ocf_collision_end_shared_access(struct ocf_metadata_lock *metadata_lock, uint32_t page); void ocf_collision_start_exclusive_access(struct ocf_metadata_lock *metadata_lock, uint32_t page); void ocf_collision_end_exclusive_access(struct ocf_metadata_lock *metadata_lock, uint32_t page); #endif