Change background worker pool to become a global variable

Allocate rrc::ue objects in batches in the background worker pool
master
Francisco 4 years ago committed by Francisco Paisana
parent 11cffdcbec
commit f610a7b83d

@ -23,26 +23,17 @@ namespace srslte {
class task_scheduler class task_scheduler
{ {
public: public:
explicit task_scheduler(uint32_t default_extern_tasks_size = 512, explicit task_scheduler(uint32_t default_extern_tasks_size = 512, uint32_t nof_timers_prealloc = 100) :
uint32_t nof_background_threads = 0, external_tasks{default_extern_tasks_size}, timers{nof_timers_prealloc}
uint32_t nof_timers_prealloc = 100) :
external_tasks{default_extern_tasks_size},
timers{nof_timers_prealloc},
background_tasks{nof_background_threads}
{ {
background_queue_id = external_tasks.add_queue(); background_queue_id = external_tasks.add_queue();
// Start background thread
if (background_tasks.nof_workers() > 0) {
background_tasks.start();
}
} }
task_scheduler(const task_scheduler&) = delete;
task_scheduler(task_scheduler&&) = delete;
task_scheduler& operator=(const task_scheduler&) = delete;
task_scheduler& operator=(task_scheduler&&) = delete;
void stop() void stop() { external_tasks.reset(); }
{
background_tasks.stop();
external_tasks.reset();
}
srslte::unique_timer get_unique_timer() { return timers.get_unique_timer(); } srslte::unique_timer get_unique_timer() { return timers.get_unique_timer(); }
@ -56,17 +47,6 @@ public:
//! Enqueues internal task to be run in next tic //! Enqueues internal task to be run in next tic
void defer_task(srslte::move_task_t func) { internal_tasks.push_back(std::move(func)); } void defer_task(srslte::move_task_t func) { internal_tasks.push_back(std::move(func)); }
//! Delegates a task to a thread pool that runs in the background
void enqueue_background_task(std::function<void(uint32_t)> f)
{
if (background_tasks.nof_workers() > 0) {
background_tasks.push_task(std::move(f));
} else {
external_tasks.push(background_queue_id,
std::bind([](const std::function<void(uint32_t)>& task) { task(0); }, std::move(f)));
}
}
//! Defer the handling of the result of a background task to next tic //! Defer the handling of the result of a background task to next tic
void notify_background_task_result(srslte::move_task_t task) void notify_background_task_result(srslte::move_task_t task)
{ {
@ -117,7 +97,6 @@ private:
} }
} }
srslte::task_thread_pool background_tasks; ///< Thread pool used for long, low-priority tasks
int background_queue_id = -1; ///< Queue for handling the outcomes of tasks run in the background int background_queue_id = -1; ///< Queue for handling the outcomes of tasks run in the background
srslte::task_multiqueue external_tasks; srslte::task_multiqueue external_tasks;
srslte::timer_handler timers; srslte::timer_handler timers;
@ -131,7 +110,6 @@ public:
task_sched_handle(task_scheduler* sched_) : sched(sched_) {} task_sched_handle(task_scheduler* sched_) : sched(sched_) {}
srslte::unique_timer get_unique_timer() { return sched->get_unique_timer(); } srslte::unique_timer get_unique_timer() { return sched->get_unique_timer(); }
void enqueue_background_task(std::function<void(uint32_t)> f) { sched->enqueue_background_task(std::move(f)); }
void notify_background_task_result(srslte::move_task_t task) void notify_background_task_result(srslte::move_task_t task)
{ {
sched->notify_background_task_result(std::move(task)); sched->notify_background_task_result(std::move(task));
@ -153,7 +131,6 @@ public:
ext_task_sched_handle(task_scheduler* sched_) : sched(sched_) {} ext_task_sched_handle(task_scheduler* sched_) : sched(sched_) {}
srslte::unique_timer get_unique_timer() { return sched->get_unique_timer(); } srslte::unique_timer get_unique_timer() { return sched->get_unique_timer(); }
void enqueue_background_task(std::function<void(uint32_t)> f) { sched->enqueue_background_task(std::move(f)); }
void notify_background_task_result(srslte::move_task_t task) void notify_background_task_result(srslte::move_task_t task)
{ {
sched->notify_background_task_result(std::move(task)); sched->notify_background_task_result(std::move(task));

@ -20,6 +20,8 @@
#ifndef SRSLTE_THREAD_POOL_H #ifndef SRSLTE_THREAD_POOL_H
#define SRSLTE_THREAD_POOL_H #define SRSLTE_THREAD_POOL_H
#include "srslte/adt/move_callback.h"
#include "srslte/srslog/srslog.h"
#include <condition_variable> #include <condition_variable>
#include <functional> #include <functional>
#include <memory> #include <memory>
@ -86,17 +88,22 @@ private:
class task_thread_pool class task_thread_pool
{ {
using task_t = std::function<void(uint32_t worker_id)>; using task_t = srslte::move_callback<void()>;
public: public:
explicit task_thread_pool(uint32_t nof_workers); task_thread_pool(uint32_t nof_workers = 1, bool start_deferred = false, int32_t prio_ = -1, uint32_t mask_ = 255);
task_thread_pool(const task_thread_pool&) = delete;
task_thread_pool(task_thread_pool&&) = delete;
task_thread_pool& operator=(const task_thread_pool&) = delete;
task_thread_pool& operator=(task_thread_pool&&) = delete;
~task_thread_pool(); ~task_thread_pool();
void start(int32_t prio = -1, uint32_t mask = 255);
void stop(); void stop();
void start(int32_t prio_ = -1, uint32_t mask_ = 255);
void set_nof_workers(uint32_t nof_workers);
void push_task(const task_t& task);
void push_task(task_t&& task); void push_task(task_t&& task);
uint32_t nof_pending_tasks(); uint32_t nof_pending_tasks() const;
size_t nof_workers() const { return workers.size(); } size_t nof_workers() const { return workers.size(); }
private: private:
@ -105,7 +112,6 @@ private:
public: public:
explicit worker_t(task_thread_pool* parent_, uint32_t id); explicit worker_t(task_thread_pool* parent_, uint32_t id);
void stop(); void stop();
void setup(int32_t prio, uint32_t mask);
bool is_running() const { return running; } bool is_running() const { return running; }
uint32_t id() const { return id_; } uint32_t id() const { return id_; }
@ -119,13 +125,19 @@ private:
bool running = false; bool running = false;
}; };
int32_t prio = -1;
uint32_t mask = 255;
srslog::basic_logger& logger;
std::queue<task_t> pending_tasks; std::queue<task_t> pending_tasks;
std::vector<worker_t> workers; std::vector<std::unique_ptr<worker_t> > workers;
std::mutex queue_mutex; mutable std::mutex queue_mutex;
std::condition_variable cv_empty; std::condition_variable cv_empty;
bool running; bool running = false;
}; };
srslte::task_thread_pool& get_background_workers();
} // namespace srslte } // namespace srslte
#endif // SRSLTE_THREAD_POOL_H #endif // SRSLTE_THREAD_POOL_H

@ -40,7 +40,7 @@ public:
// run pending tasks without updating timers // run pending tasks without updating timers
void run_pending_tasks() { task_sched.run_pending_tasks(); } void run_pending_tasks() { task_sched.run_pending_tasks(); }
srslte::task_scheduler task_sched{512, 0, 100}; srslte::task_scheduler task_sched{512, 100};
}; };
class rlc_dummy_interface : public rlc_interface_mac class rlc_dummy_interface : public rlc_interface_mac

@ -11,6 +11,7 @@
*/ */
#include "srslte/common/thread_pool.h" #include "srslte/common/thread_pool.h"
#include "srslte/srslog/srslog.h"
#include <assert.h> #include <assert.h>
#include <chrono> #include <chrono>
#include <stdio.h> #include <stdio.h>
@ -244,11 +245,11 @@ uint32_t thread_pool::get_nof_workers()
* once a worker is available * once a worker is available
*************************************************************************/ *************************************************************************/
task_thread_pool::task_thread_pool(uint32_t nof_workers) : running(false) task_thread_pool::task_thread_pool(uint32_t nof_workers, bool start_deferred, int32_t prio_, uint32_t mask_) :
logger(srslog::fetch_basic_logger("POOL")), workers(std::max(1u, nof_workers))
{ {
workers.reserve(nof_workers); if (not start_deferred) {
for (uint32_t i = 0; i < nof_workers; ++i) { start(prio_, mask_);
workers.emplace_back(this, i);
} }
} }
@ -257,12 +258,34 @@ task_thread_pool::~task_thread_pool()
stop(); stop();
} }
void task_thread_pool::start(int32_t prio, uint32_t mask) void task_thread_pool::set_nof_workers(uint32_t nof_workers)
{ {
std::lock_guard<std::mutex> lock(queue_mutex); std::lock_guard<std::mutex> lock(queue_mutex);
if (workers.size() > nof_workers) {
logger.error("Reducing the number of workers dynamically not supported");
return;
}
uint32_t old_size = workers.size();
workers.resize(nof_workers);
if (running) {
for (uint32_t i = old_size; i < nof_workers; ++i) {
workers[i].reset(new worker_t(this, i));
}
}
}
void task_thread_pool::start(int32_t prio_, uint32_t mask_)
{
std::lock_guard<std::mutex> lock(queue_mutex);
if (running) {
logger.error("Starting thread pool that has already started");
return;
}
prio = prio_;
mask = mask_;
running = true; running = true;
for (worker_t& w : workers) { for (uint32_t i = 0; i < workers.size(); ++i) {
w.setup(prio, mask); workers[i].reset(new worker_t(this, i));
} }
} }
@ -272,8 +295,8 @@ void task_thread_pool::stop()
if (running) { if (running) {
running = false; running = false;
bool workers_running = false; bool workers_running = false;
for (worker_t& w : workers) { for (std::unique_ptr<worker_t>& w : workers) {
if (w.is_running()) { if (w->is_running()) {
workers_running = true; workers_running = true;
break; break;
} }
@ -282,21 +305,12 @@ void task_thread_pool::stop()
if (workers_running) { if (workers_running) {
cv_empty.notify_all(); cv_empty.notify_all();
} }
for (worker_t& w : workers) { for (std::unique_ptr<worker_t>& w : workers) {
w.stop(); w->stop();
} }
} }
} }
void task_thread_pool::push_task(const task_t& task)
{
{
std::lock_guard<std::mutex> lock(queue_mutex);
pending_tasks.push(task);
}
cv_empty.notify_one();
}
void task_thread_pool::push_task(task_t&& task) void task_thread_pool::push_task(task_t&& task)
{ {
{ {
@ -306,17 +320,20 @@ void task_thread_pool::push_task(task_t&& task)
cv_empty.notify_one(); cv_empty.notify_one();
} }
uint32_t task_thread_pool::nof_pending_tasks() uint32_t task_thread_pool::nof_pending_tasks() const
{ {
std::lock_guard<std::mutex> lock(queue_mutex); std::lock_guard<std::mutex> lock(queue_mutex);
return pending_tasks.size(); return pending_tasks.size();
} }
task_thread_pool::worker_t::worker_t(srslte::task_thread_pool* parent_, uint32_t my_id) : task_thread_pool::worker_t::worker_t(srslte::task_thread_pool* parent_, uint32_t my_id) :
parent(parent_), parent(parent_), thread(std::string("TASKWORKER") + std::to_string(my_id)), id_(my_id), running(true)
thread(std::string("TASKWORKER") + std::to_string(my_id)),
id_(my_id)
{ {
if (parent->mask == 255) {
start(parent->prio);
} else {
start_cpu_mask(parent->prio, parent->mask);
}
} }
void task_thread_pool::worker_t::stop() void task_thread_pool::worker_t::stop()
@ -324,16 +341,6 @@ void task_thread_pool::worker_t::stop()
wait_thread_finish(); wait_thread_finish();
} }
void task_thread_pool::worker_t::setup(int32_t prio, uint32_t mask)
{
running = true;
if (mask == 255) {
start(prio);
} else {
start_cpu_mask(prio, mask);
}
}
bool task_thread_pool::worker_t::wait_task(task_t* task) bool task_thread_pool::worker_t::wait_task(task_t* task)
{ {
std::unique_lock<std::mutex> lock(parent->queue_mutex); std::unique_lock<std::mutex> lock(parent->queue_mutex);
@ -355,7 +362,7 @@ void task_thread_pool::worker_t::run_thread()
// main loop // main loop
task_t task; task_t task;
while (wait_task(&task)) { while (wait_task(&task)) {
task(id()); task();
} }
// on exit, notify pool class // on exit, notify pool class
@ -363,4 +370,11 @@ void task_thread_pool::worker_t::run_thread()
running = false; running = false;
} }
// Global thread pool for long, low-priority tasks
task_thread_pool& get_background_workers()
{
static task_thread_pool background_workers;
return background_workers;
}
} // namespace srslte } // namespace srslte

@ -14,6 +14,7 @@
#include "srslte/common/multiqueue.h" #include "srslte/common/multiqueue.h"
#include "srslte/common/thread_pool.h" #include "srslte/common/thread_pool.h"
#include <iostream> #include <iostream>
#include <map>
#include <thread> #include <thread>
#include <unistd.h> #include <unistd.h>
@ -235,16 +236,14 @@ int test_task_thread_pool()
// Description: check whether the tasks are successfully distributed between workers // Description: check whether the tasks are successfully distributed between workers
uint32_t nof_workers = 4, nof_runs = 10000; uint32_t nof_workers = 4, nof_runs = 10000;
std::vector<int> count_worker(nof_workers, 0); std::mutex count_mutex;
std::vector<std::mutex> count_mutex(nof_workers); std::map<std::thread::id, int> count_worker;
task_thread_pool thread_pool(nof_workers); task_thread_pool thread_pool(nof_workers);
thread_pool.start();
auto task = [&count_worker, &count_mutex](uint32_t worker_id) { auto task = [&count_worker, &count_mutex]() {
std::lock_guard<std::mutex> lock(count_mutex[worker_id]); std::lock_guard<std::mutex> lock(count_mutex);
// std::cout << "hello world from worker " << worker_id << std::endl; count_worker[std::this_thread::get_id()]++;
count_worker[worker_id]++;
}; };
for (uint32_t i = 0; i < nof_runs; ++i) { for (uint32_t i = 0; i < nof_runs; ++i) {
@ -259,12 +258,12 @@ int test_task_thread_pool()
thread_pool.stop(); thread_pool.stop();
uint32_t total_count = 0; uint32_t total_count = 0;
for (uint32_t i = 0; i < nof_workers; ++i) { for (auto& w : count_worker) {
if (count_worker[i] < 10) { if (w.second < 10) {
printf("WARNING: the number of tasks %d assigned to worker %d is too low\n", count_worker[i], i); std::cout << "WARNING: the number of tasks " << w.second << " assigned to worker " << w.first << " is too low";
} }
total_count += count_worker[i]; total_count += w.second;
printf("worker %d: %d runs\n", i, count_worker[i]); std::cout << "worker " << w.first << ": " << w.second << " runs\n";
} }
if (total_count != nof_runs) { if (total_count != nof_runs) {
printf("Number of task runs=%d does not match total=%d\n", total_count, nof_runs); printf("Number of task runs=%d does not match total=%d\n", total_count, nof_runs);
@ -289,14 +288,14 @@ int test_task_thread_pool2()
task_thread_pool thread_pool(nof_workers); task_thread_pool thread_pool(nof_workers);
thread_pool.start(); thread_pool.start();
auto task = [&workers_started, &workers_finished, &mut](uint32_t worker_id) { auto task = [&workers_started, &workers_finished, &mut]() {
{ {
std::lock_guard<std::mutex> lock(mut); std::lock_guard<std::mutex> lock(mut);
workers_started++; workers_started++;
} }
sleep(1); sleep(1);
std::lock_guard<std::mutex> lock(mut); std::lock_guard<std::mutex> lock(mut);
std::cout << "worker " << worker_id << " has finished\n"; std::cout << "worker has finished\n";
workers_finished++; workers_finished++;
}; };

@ -39,12 +39,10 @@ int test_task_scheduler_no_pool()
// TEST: background task is run, despite there are no pool workers // TEST: background task is run, despite there are no pool workers
state = task_result::null; state = task_result::null;
task_sched.enqueue_background_task([&task_sched, &state](uint32_t worker_id) { srslte::get_background_workers().push_task([&task_sched, &state]() {
task_sched.notify_background_task_result([&state]() { state = task_result::external; }); task_sched.notify_background_task_result([&state]() { state = task_result::external; });
}); });
TESTASSERT(state == task_result::null); TESTASSERT(state == task_result::null);
task_sched.run_next_task(); // runs background task
TESTASSERT(state == task_result::null);
task_sched.run_next_task(); // runs notification task_sched.run_next_task(); // runs notification
TESTASSERT(state == task_result::external); TESTASSERT(state == task_result::external);
@ -56,7 +54,7 @@ int test_task_scheduler_with_pool()
srslte::task_scheduler task_sched{5, 2}; srslte::task_scheduler task_sched{5, 2};
task_result state = task_result::null; task_result state = task_result::null;
task_sched.enqueue_background_task([&task_sched, &state](uint32_t worker_id) { srslte::get_background_workers().push_task([&task_sched, &state]() {
task_sched.notify_background_task_result([&state]() { state = task_result::external; }); task_sched.notify_background_task_result([&state]() { state = task_result::external; });
}); });
TESTASSERT(state == task_result::null); TESTASSERT(state == task_result::null);

@ -29,7 +29,7 @@ enb_stack_lte::enb_stack_lte(srslte::logger* logger_, srslog::sink& log_sink) :
s1ap_logger(srslog::fetch_basic_logger("S1AP", log_sink, false)), s1ap_logger(srslog::fetch_basic_logger("S1AP", log_sink, false)),
gtpu_logger(srslog::fetch_basic_logger("GTPU", log_sink, false)), gtpu_logger(srslog::fetch_basic_logger("GTPU", log_sink, false)),
stack_logger(srslog::fetch_basic_logger("STCK", log_sink, false)), stack_logger(srslog::fetch_basic_logger("STCK", log_sink, false)),
task_sched(512, 0, 128), task_sched(512, 128),
pdcp(&task_sched, pdcp_logger), pdcp(&task_sched, pdcp_logger),
mac(&task_sched, mac_logger), mac(&task_sched, mac_logger),
rlc(rlc_logger), rlc(rlc_logger),
@ -39,6 +39,7 @@ enb_stack_lte::enb_stack_lte(srslte::logger* logger_, srslog::sink& log_sink) :
logger(logger_), logger(logger_),
mac_pcap(srslte_rat_t::lte) mac_pcap(srslte_rat_t::lte)
{ {
get_background_workers().set_nof_workers(2);
enb_task_queue = task_sched.make_task_queue(); enb_task_queue = task_sched.make_task_queue();
mme_task_queue = task_sched.make_task_queue(); mme_task_queue = task_sched.make_task_queue();
gtpu_task_queue = task_sched.make_task_queue(); gtpu_task_queue = task_sched.make_task_queue();
@ -187,6 +188,7 @@ void enb_stack_lte::stop_impl()
} }
task_sched.stop(); task_sched.stop();
get_background_workers().stop();
started = false; started = false;
} }

@ -16,7 +16,7 @@
namespace srsenb { namespace srsenb {
gnb_stack_nr::gnb_stack_nr(srslte::logger* logger_) : logger(logger_), task_sched{512, 1, 128}, thread("gNB") gnb_stack_nr::gnb_stack_nr(srslte::logger* logger_) : logger(logger_), task_sched{512, 128}, thread("gNB")
{ {
m_mac.reset(new mac_nr()); m_mac.reset(new mac_nr());
m_rlc.reset(new rlc_nr("RLC")); m_rlc.reset(new rlc_nr("RLC"));
@ -115,6 +115,7 @@ void gnb_stack_nr::stop()
m_pdcp->stop(); m_pdcp->stop();
m_mac->stop(); m_mac->stop();
srslte::get_background_workers().stop();
running = false; running = false;
} }
} }

@ -476,6 +476,8 @@ uint16_t mac::allocate_ue()
ue_db[rnti] = std::move(ue_ptr); ue_db[rnti] = std::move(ue_ptr);
} }
// Allocate one new UE object in advance
srslte::get_background_workers().push_task([this]() { prealloc_ue(1); });
return rnti; return rnti;
} }
@ -486,11 +488,6 @@ uint16_t mac::reserve_new_crnti(const sched_interface::ue_cfg_t& ue_cfg)
return rnti; return rnti;
} }
task_sched.enqueue_background_task([this](uint32_t wid) {
// Allocate one new UE object in advance
prealloc_ue(1);
});
// Add new user to the scheduler so that it can RX/TX SRB0 // Add new user to the scheduler so that it can RX/TX SRB0
if (scheduler.ue_cfg(rnti, ue_cfg) != SRSLTE_SUCCESS) { if (scheduler.ue_cfg(rnti, ue_cfg) != SRSLTE_SUCCESS) {
logger.error("Registering new user rnti=0x%x to SCHED", rnti); logger.error("Registering new user rnti=0x%x to SCHED", rnti);
@ -556,9 +553,6 @@ void mac::rach_detected(uint32_t tti, uint32_t enb_cc_idx, uint32_t preamble_idx
time_adv, time_adv,
rnti); rnti);
}); });
// Allocate one new UE object in advance
prealloc_ue(1);
} }
void mac::prealloc_ue(uint32_t nof_ue) void mac::prealloc_ue(uint32_t nof_ue)

@ -160,9 +160,6 @@ int rrc::add_user(uint16_t rnti, const sched_interface::ue_cfg_t& sched_ue_cfg)
logger.error("Adding user rnti=0x%x - Failed to allocate user resources", rnti); logger.error("Adding user rnti=0x%x - Failed to allocate user resources", rnti);
return SRSLTE_ERROR; return SRSLTE_ERROR;
} }
if (ue_pool.capacity() <= 4) {
task_sched.defer_task([]() { rrc::ue_pool.reserve(16); });
}
users.insert(std::make_pair(rnti, std::move(u))); users.insert(std::make_pair(rnti, std::move(u)));
} }
rlc->add_user(rnti); rlc->add_user(rnti);

@ -65,7 +65,11 @@ int rrc::ue::init()
void* rrc::ue::operator new(size_t sz) void* rrc::ue::operator new(size_t sz)
{ {
assert(sz == sizeof(ue)); assert(sz == sizeof(ue));
return rrc::ue_pool.allocate_node(sz); void* memchunk = rrc::ue_pool.allocate_node(sz);
if (ue_pool.capacity() <= 4) {
srslte::get_background_workers().push_task([]() { rrc::ue_pool.reserve(4); });
}
return memchunk;
} }
void rrc::ue::operator delete(void* ptr)noexcept void rrc::ue::operator delete(void* ptr)noexcept
{ {

@ -43,10 +43,11 @@ ue_stack_lte::ue_stack_lte(srslog::sink& log_sink) :
pdcp(&task_sched, "PDCP"), pdcp(&task_sched, "PDCP"),
nas(&task_sched), nas(&task_sched),
thread("STACK"), thread("STACK"),
task_sched(512, 2, 64), task_sched(512, 64),
tti_tprof("tti_tprof", "STCK", TTI_STAT_PERIOD), tti_tprof("tti_tprof", "STCK", TTI_STAT_PERIOD),
mac_pcap(srslte_rat_t::lte) mac_pcap(srslte_rat_t::lte)
{ {
get_background_workers().set_nof_workers(2);
ue_task_queue = task_sched.make_task_queue(); ue_task_queue = task_sched.make_task_queue();
gw_queue_id = task_sched.make_task_queue(); gw_queue_id = task_sched.make_task_queue();
cfg_task_queue = task_sched.make_task_queue(); cfg_task_queue = task_sched.make_task_queue();
@ -156,16 +157,7 @@ int ue_stack_lte::init(const stack_args_t& args_, srslte::logger* logger_)
mac_nr_args_t mac_nr_args = {}; mac_nr_args_t mac_nr_args = {};
mac_nr.init(mac_nr_args, phy_nr, &rlc); mac_nr.init(mac_nr_args, phy_nr, &rlc);
rrc_nr.init(phy_nr, rrc_nr.init(phy_nr, &mac_nr, &rlc, &pdcp, gw, &rrc, usim.get(), task_sched.get_timer_handler(), nullptr, args.rrc_nr);
&mac_nr,
&rlc,
&pdcp,
gw,
&rrc,
usim.get(),
task_sched.get_timer_handler(),
nullptr,
args.rrc_nr);
rrc.init(phy, &mac, &rlc, &pdcp, &nas, usim.get(), gw, &rrc_nr, args.rrc); rrc.init(phy, &mac, &rlc, &pdcp, &nas, usim.get(), gw, &rrc_nr, args.rrc);
running = true; running = true;
@ -200,6 +192,9 @@ void ue_stack_lte::stop_impl()
if (args.pcap.nas_enable) { if (args.pcap.nas_enable) {
nas_pcap.close(); nas_pcap.close();
} }
task_sched.stop();
get_background_workers().stop();
} }
bool ue_stack_lte::switch_on() bool ue_stack_lte::switch_on()

@ -18,8 +18,9 @@ using namespace srslte;
namespace srsue { namespace srsue {
ue_stack_nr::ue_stack_nr(srslte::logger* logger_) : ue_stack_nr::ue_stack_nr(srslte::logger* logger_) :
logger(logger_), thread("STACK"), task_sched(64, 2, 64), rlc_log("RLC"), pdcp_log("PDCP") logger(logger_), thread("STACK"), task_sched(64, 64), rlc_log("RLC"), pdcp_log("PDCP")
{ {
get_background_workers().set_nof_workers(2);
mac.reset(new mac_nr(&task_sched)); mac.reset(new mac_nr(&task_sched));
pdcp.reset(new srslte::pdcp(&task_sched, "PDCP")); pdcp.reset(new srslte::pdcp(&task_sched, "PDCP"));
rlc.reset(new srslte::rlc("RLC")); rlc.reset(new srslte::rlc("RLC"));
@ -102,6 +103,8 @@ void ue_stack_nr::stop_impl()
rlc->stop(); rlc->stop();
pdcp->stop(); pdcp->stop();
mac->stop(); mac->stop();
get_background_workers().stop();
} }
bool ue_stack_nr::switch_on() bool ue_stack_nr::switch_on()

@ -17,11 +17,10 @@ using namespace srsue;
int rrc_nr_cap_request_test() int rrc_nr_cap_request_test()
{ {
srslte::log_ref rrc_log("RRC"); srslte::log_ref rrc_log("RRC");
rrc_log->set_level(srslte::LOG_LEVEL_DEBUG); rrc_log->set_level(srslte::LOG_LEVEL_DEBUG);
rrc_log->set_hex_limit(-1); rrc_log->set_hex_limit(-1);
srslte::task_scheduler task_sched{512, 0, 100}; srslte::task_scheduler task_sched{512, 100};
srslte::task_sched_handle task_sched_handle(&task_sched); srslte::task_sched_handle task_sched_handle(&task_sched);
rrc_nr rrc_nr(task_sched_handle); rrc_nr rrc_nr(task_sched_handle);
srslte::byte_buffer_t caps; srslte::byte_buffer_t caps;

Loading…
Cancel
Save