[improvement](executor) Support task group schedule in pipeline engine (#17615)
This commit is contained in:
@ -17,9 +17,13 @@
|
||||
|
||||
#include "task_queue.h"
|
||||
|
||||
#include "runtime/task_group/task_group.h"
|
||||
|
||||
namespace doris {
|
||||
namespace pipeline {
|
||||
|
||||
TaskQueue::~TaskQueue() = default;
|
||||
|
||||
PipelineTask* SubWorkTaskQueue::try_take(bool is_steal) {
|
||||
if (_queue.empty()) {
|
||||
return nullptr;
|
||||
@ -35,7 +39,7 @@ PipelineTask* SubWorkTaskQueue::try_take(bool is_steal) {
|
||||
|
||||
//////////////////// WorkTaskQueue ////////////////////
|
||||
|
||||
WorkTaskQueue::WorkTaskQueue() : _closed(false) {
|
||||
NormalWorkTaskQueue::NormalWorkTaskQueue() : _closed(false) {
|
||||
double factor = 1;
|
||||
for (int i = 0; i < SUB_QUEUE_LEVEL; ++i) {
|
||||
_sub_queues[i].set_factor_for_normal(factor);
|
||||
@ -49,13 +53,13 @@ WorkTaskQueue::WorkTaskQueue() : _closed(false) {
|
||||
}
|
||||
}
|
||||
|
||||
void WorkTaskQueue::close() {
|
||||
void NormalWorkTaskQueue::close() {
|
||||
std::unique_lock<std::mutex> lock(_work_size_mutex);
|
||||
_closed = true;
|
||||
_wait_task.notify_all();
|
||||
}
|
||||
|
||||
PipelineTask* WorkTaskQueue::try_take_unprotected(bool is_steal) {
|
||||
PipelineTask* NormalWorkTaskQueue::try_take_unprotected(bool is_steal) {
|
||||
if (_total_task_size == 0 || _closed) {
|
||||
return nullptr;
|
||||
}
|
||||
@ -86,7 +90,7 @@ PipelineTask* WorkTaskQueue::try_take_unprotected(bool is_steal) {
|
||||
return task;
|
||||
}
|
||||
|
||||
int WorkTaskQueue::_compute_level(PipelineTask* task) {
|
||||
int NormalWorkTaskQueue::_compute_level(PipelineTask* task) {
|
||||
uint32_t schedule_time = task->total_schedule_time();
|
||||
for (int i = 0; i < SUB_QUEUE_LEVEL - 1; ++i) {
|
||||
if (schedule_time <= _task_schedule_limit[i]) {
|
||||
@ -96,13 +100,13 @@ int WorkTaskQueue::_compute_level(PipelineTask* task) {
|
||||
return SUB_QUEUE_LEVEL - 1;
|
||||
}
|
||||
|
||||
PipelineTask* WorkTaskQueue::try_take(bool is_steal) {
|
||||
PipelineTask* NormalWorkTaskQueue::try_take(bool is_steal) {
|
||||
// TODO other efficient lock? e.g. if get lock fail, return null_ptr
|
||||
std::unique_lock<std::mutex> lock(_work_size_mutex);
|
||||
return try_take_unprotected(is_steal);
|
||||
}
|
||||
|
||||
PipelineTask* WorkTaskQueue::take(uint32_t timeout_ms) {
|
||||
PipelineTask* NormalWorkTaskQueue::take(uint32_t timeout_ms) {
|
||||
std::unique_lock<std::mutex> lock(_work_size_mutex);
|
||||
auto task = try_take_unprotected(false);
|
||||
if (task) {
|
||||
@ -117,7 +121,7 @@ PipelineTask* WorkTaskQueue::take(uint32_t timeout_ms) {
|
||||
}
|
||||
}
|
||||
|
||||
Status WorkTaskQueue::push(PipelineTask* task) {
|
||||
Status NormalWorkTaskQueue::push(PipelineTask* task) {
|
||||
if (_closed) {
|
||||
return Status::InternalError("WorkTaskQueue closed");
|
||||
}
|
||||
@ -129,23 +133,27 @@ Status WorkTaskQueue::push(PipelineTask* task) {
|
||||
return Status::OK();
|
||||
}
|
||||
|
||||
////////////////// TaskQueue ////////////
|
||||
NormalTaskQueue::~NormalTaskQueue() = default;
|
||||
|
||||
void TaskQueue::close() {
|
||||
NormalTaskQueue::NormalTaskQueue(size_t core_size) : TaskQueue(core_size), _closed(false) {
|
||||
_async_queue.reset(new NormalWorkTaskQueue[core_size]);
|
||||
}
|
||||
|
||||
void NormalTaskQueue::close() {
|
||||
_closed = true;
|
||||
for (int i = 0; i < _core_size; ++i) {
|
||||
_async_queue[i].close();
|
||||
}
|
||||
}
|
||||
|
||||
PipelineTask* TaskQueue::try_take(size_t core_id) {
|
||||
PipelineTask* NormalTaskQueue::take(size_t core_id) {
|
||||
PipelineTask* task = nullptr;
|
||||
while (!_closed) {
|
||||
task = _async_queue[core_id].try_take(false);
|
||||
if (task) {
|
||||
break;
|
||||
}
|
||||
task = steal_take(core_id);
|
||||
task = _steal_take(core_id);
|
||||
if (task) {
|
||||
break;
|
||||
}
|
||||
@ -160,7 +168,7 @@ PipelineTask* TaskQueue::try_take(size_t core_id) {
|
||||
return task;
|
||||
}
|
||||
|
||||
PipelineTask* TaskQueue::steal_take(size_t core_id) {
|
||||
PipelineTask* NormalTaskQueue::_steal_take(size_t core_id) {
|
||||
DCHECK(core_id < _core_size);
|
||||
size_t next_id = core_id;
|
||||
for (size_t i = 1; i < _core_size; ++i) {
|
||||
@ -177,7 +185,7 @@ PipelineTask* TaskQueue::steal_take(size_t core_id) {
|
||||
return nullptr;
|
||||
}
|
||||
|
||||
Status TaskQueue::push_back(PipelineTask* task) {
|
||||
Status NormalTaskQueue::push_back(PipelineTask* task) {
|
||||
int core_id = task->get_previous_core_id();
|
||||
if (core_id < 0) {
|
||||
core_id = _next_core.fetch_add(1) % _core_size;
|
||||
@ -185,11 +193,158 @@ Status TaskQueue::push_back(PipelineTask* task) {
|
||||
return push_back(task, core_id);
|
||||
}
|
||||
|
||||
Status TaskQueue::push_back(PipelineTask* task, size_t core_id) {
|
||||
Status NormalTaskQueue::push_back(PipelineTask* task, size_t core_id) {
|
||||
DCHECK(core_id < _core_size);
|
||||
task->put_in_runnable_queue();
|
||||
return _async_queue[core_id].push(task);
|
||||
}
|
||||
|
||||
bool TaskGroupTaskQueue::TaskGroupSchedEntityComparator::operator()(
|
||||
const taskgroup::TGEntityPtr& lhs_ptr, const taskgroup::TGEntityPtr& rhs_ptr) const {
|
||||
int64_t lhs_val = lhs_ptr->vruntime_ns();
|
||||
int64_t rhs_val = rhs_ptr->vruntime_ns();
|
||||
if (lhs_val != rhs_val) {
|
||||
return lhs_val < rhs_val;
|
||||
} else {
|
||||
auto l_share = lhs_ptr->cpu_share();
|
||||
auto r_share = rhs_ptr->cpu_share();
|
||||
if (l_share != r_share) {
|
||||
return l_share < rhs_val;
|
||||
} else {
|
||||
return lhs_ptr < rhs_ptr;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
TaskGroupTaskQueue::TaskGroupTaskQueue(size_t core_size) : TaskQueue(core_size) {}
|
||||
|
||||
TaskGroupTaskQueue::~TaskGroupTaskQueue() = default;
|
||||
|
||||
void TaskGroupTaskQueue::close() {
|
||||
std::unique_lock<std::mutex> lock(_rs_mutex);
|
||||
_closed = true;
|
||||
_wait_task.notify_all();
|
||||
}
|
||||
|
||||
Status TaskGroupTaskQueue::push_back(PipelineTask* task) {
|
||||
return _push_back<false>(task);
|
||||
}
|
||||
|
||||
Status TaskGroupTaskQueue::push_back(PipelineTask* task, size_t core_id) {
|
||||
return _push_back<true>(task);
|
||||
}
|
||||
|
||||
template <bool from_executor>
|
||||
Status TaskGroupTaskQueue::_push_back(PipelineTask* task) {
|
||||
auto* entity = task->get_task_group()->task_entity();
|
||||
std::unique_lock<std::mutex> lock(_rs_mutex);
|
||||
entity->push_back(task);
|
||||
if (_group_entities.find(entity) == _group_entities.end()) {
|
||||
_enqueue_task_group<from_executor>(entity);
|
||||
}
|
||||
_wait_task.notify_one();
|
||||
return Status::OK();
|
||||
}
|
||||
|
||||
// TODO pipeline support steal
|
||||
PipelineTask* TaskGroupTaskQueue::take(size_t core_id) {
|
||||
std::unique_lock<std::mutex> lock(_rs_mutex);
|
||||
taskgroup::TGEntityPtr entity = nullptr;
|
||||
while (entity == nullptr) {
|
||||
if (_closed) {
|
||||
return nullptr;
|
||||
}
|
||||
if (_group_entities.empty()) {
|
||||
_wait_task.wait(lock);
|
||||
} else {
|
||||
entity = _next_tg_entity();
|
||||
if (!entity) {
|
||||
_wait_task.wait_for(lock, std::chrono::milliseconds(WAIT_CORE_TASK_TIMEOUT_MS));
|
||||
}
|
||||
}
|
||||
}
|
||||
DCHECK(entity->task_size() > 0);
|
||||
if (entity->task_size() == 1) {
|
||||
_dequeue_task_group(entity);
|
||||
}
|
||||
return entity->take();
|
||||
}
|
||||
|
||||
template <bool from_worker>
|
||||
void TaskGroupTaskQueue::_enqueue_task_group(taskgroup::TGEntityPtr tg_entity) {
|
||||
_total_cpu_share += tg_entity->cpu_share();
|
||||
if constexpr (!from_worker) {
|
||||
/**
|
||||
* If a task group entity leaves task queue for a long time, its v runtime will be very
|
||||
* small. This can cause it to preempt too many execution time. So, in order to avoid this
|
||||
* situation, it is necessary to adjust the task group's v runtime.
|
||||
* */
|
||||
auto old_v_ns = tg_entity->vruntime_ns();
|
||||
auto* min_entity = _min_tg_entity.load();
|
||||
if (min_entity) {
|
||||
int64_t new_vruntime_ns = min_entity->vruntime_ns() - _ideal_runtime_ns(tg_entity) / 2;
|
||||
if (new_vruntime_ns > old_v_ns) {
|
||||
tg_entity->adjust_vruntime_ns(new_vruntime_ns);
|
||||
}
|
||||
} else if (old_v_ns < _min_tg_v_runtime_ns) {
|
||||
tg_entity->adjust_vruntime_ns(_min_tg_v_runtime_ns);
|
||||
}
|
||||
}
|
||||
_group_entities.emplace(tg_entity);
|
||||
VLOG_DEBUG << "enqueue tg " << tg_entity->debug_string()
|
||||
<< ", group entity size: " << _group_entities.size();
|
||||
_update_min_tg();
|
||||
}
|
||||
|
||||
void TaskGroupTaskQueue::_dequeue_task_group(taskgroup::TGEntityPtr tg_entity) {
|
||||
_total_cpu_share -= tg_entity->cpu_share();
|
||||
_group_entities.erase(tg_entity);
|
||||
VLOG_DEBUG << "dequeue tg " << tg_entity->debug_string()
|
||||
<< ", group entity size: " << _group_entities.size();
|
||||
_update_min_tg();
|
||||
}
|
||||
|
||||
void TaskGroupTaskQueue::_update_min_tg() {
|
||||
auto* min_entity = _next_tg_entity();
|
||||
_min_tg_entity = min_entity;
|
||||
if (min_entity) {
|
||||
auto min_v_runtime = min_entity->vruntime_ns();
|
||||
if (min_v_runtime > _min_tg_v_runtime_ns) {
|
||||
_min_tg_v_runtime_ns = min_v_runtime;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// like sched_fair.c calc_delta_fair, THREAD_TIME_SLICE maybe a dynamic value.
|
||||
int64_t TaskGroupTaskQueue::_ideal_runtime_ns(taskgroup::TGEntityPtr tg_entity) const {
|
||||
return PipelineTask::THREAD_TIME_SLICE * _core_size * tg_entity->cpu_share() / _total_cpu_share;
|
||||
}
|
||||
|
||||
taskgroup::TGEntityPtr TaskGroupTaskQueue::_next_tg_entity() {
|
||||
taskgroup::TGEntityPtr res = nullptr;
|
||||
for (auto* entity : _group_entities) {
|
||||
res = entity;
|
||||
break;
|
||||
}
|
||||
return res;
|
||||
}
|
||||
|
||||
void TaskGroupTaskQueue::update_statistics(PipelineTask* task, int64_t time_spent) {
|
||||
std::unique_lock<std::mutex> lock(_rs_mutex);
|
||||
auto* group = task->get_task_group();
|
||||
auto* entity = group->task_entity();
|
||||
auto find_entity = _group_entities.find(entity);
|
||||
bool is_in_queue = find_entity != _group_entities.end();
|
||||
VLOG_DEBUG << "update_statistics " << entity->debug_string() << ", in queue:" << is_in_queue;
|
||||
if (is_in_queue) {
|
||||
_group_entities.erase(entity);
|
||||
}
|
||||
entity->incr_runtime_ns(time_spent);
|
||||
if (is_in_queue) {
|
||||
_group_entities.emplace(entity);
|
||||
_update_min_tg();
|
||||
}
|
||||
}
|
||||
|
||||
} // namespace pipeline
|
||||
} // namespace doris
|
||||
Reference in New Issue
Block a user