mirror of https://github.com/pvnis/srsRAN_4G.git
You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
117 lines
3.1 KiB
C++
117 lines
3.1 KiB
C++
/**
|
|
*
|
|
* \section COPYRIGHT
|
|
*
|
|
* Copyright 2013-2021 Software Radio Systems Limited
|
|
*
|
|
* By using this file, you agree to the terms and conditions set
|
|
* forth in the LICENSE file which can be found at the top level of
|
|
* the distribution.
|
|
*
|
|
*/
|
|
|
|
#ifndef SRSRAN_CACHED_ALLOC_H
|
|
#define SRSRAN_CACHED_ALLOC_H
|
|
|
|
#include "../intrusive_list.h"
|
|
#include "memblock_cache.h"
|
|
#include <deque>
|
|
#include <queue>
|
|
|
|
namespace srsran {
|
|
|
|
/**
|
|
* Custom Allocator that caches deallocated memory blocks in a stack to be reused in future allocations.
|
|
* This minimizes the number of new/delete calls, when the rate of insertions/removals match (e.g. a queue)
|
|
* This allocator is not thread-safe. It assumes the container is being used in a single-threaded environment,
|
|
* or being mutexed when altered, which is a reasonable assumption
|
|
* @tparam T object type
|
|
*/
|
|
template <typename T>
|
|
class cached_alloc : public std::allocator<T>
|
|
{
|
|
struct memblock_t : public intrusive_double_linked_list_element<> {
|
|
explicit memblock_t(size_t sz) : block_size(sz) {}
|
|
size_t block_size;
|
|
};
|
|
const size_t min_n = (sizeof(memblock_t) + sizeof(T) - 1) / sizeof(T);
|
|
|
|
public:
|
|
using value_type = T;
|
|
|
|
~cached_alloc()
|
|
{
|
|
while (not free_list.empty()) {
|
|
memblock_t& b = free_list.front();
|
|
free_list.pop_front();
|
|
size_t n = b.block_size;
|
|
b.~memblock_t();
|
|
std::allocator<T>::deallocate(reinterpret_cast<T*>(&b), n);
|
|
}
|
|
}
|
|
cached_alloc() = default;
|
|
cached_alloc(cached_alloc<T>&& other) noexcept = default;
|
|
cached_alloc(const cached_alloc<T>& other) noexcept : cached_alloc() {}
|
|
template <typename U>
|
|
explicit cached_alloc(const cached_alloc<U>& other) noexcept : cached_alloc()
|
|
{
|
|
// start empty, as cached blocks cannot be copied
|
|
}
|
|
cached_alloc& operator=(const cached_alloc<T>& other) noexcept { return *this; }
|
|
cached_alloc& operator=(cached_alloc&& other) noexcept = default;
|
|
|
|
T* allocate(size_t n, const void* ptr = nullptr)
|
|
{
|
|
size_t req_n = std::max(n, min_n);
|
|
for (memblock_t& b : free_list) {
|
|
if (b.block_size == req_n) {
|
|
free_list.pop(&b);
|
|
b.~memblock_t();
|
|
return reinterpret_cast<T*>(&b);
|
|
}
|
|
}
|
|
return std::allocator<T>::allocate(req_n, ptr);
|
|
}
|
|
void deallocate(T* p, size_t n) noexcept
|
|
{
|
|
size_t req_n = std::max(n, min_n);
|
|
auto* block = reinterpret_cast<memblock_t*>(p);
|
|
new (block) memblock_t(req_n);
|
|
free_list.push_front(block);
|
|
}
|
|
|
|
template <typename U>
|
|
struct rebind {
|
|
using other = cached_alloc<U>;
|
|
};
|
|
|
|
private:
|
|
intrusive_double_linked_list<memblock_t> free_list;
|
|
};
|
|
|
|
} // namespace srsran
|
|
|
|
template <typename T1, typename T2>
|
|
bool operator==(const srsran::cached_alloc<T1>& lhs, const srsran::cached_alloc<T2>& rhs) noexcept
|
|
{
|
|
return &lhs == &rhs;
|
|
}
|
|
|
|
template <typename T1, typename T2>
|
|
bool operator!=(const srsran::cached_alloc<T1>& lhs, const srsran::cached_alloc<T2>& rhs) noexcept
|
|
{
|
|
return not(lhs == rhs);
|
|
}
|
|
|
|
namespace srsran {
|
|
|
|
template <typename T>
|
|
using deque = std::deque<T, cached_alloc<T> >;
|
|
|
|
template <typename T>
|
|
using queue = std::queue<T, srsran::deque<T> >;
|
|
|
|
} // namespace srsran
|
|
|
|
#endif // SRSRAN_CACHED_ALLOC_H
|