mem tracker can be logically divided into 4 layers: 1)process 2)type 3)query/load/compation task etc. 4)exec node etc.
type includes
enum Type {
GLOBAL = 0, // Life cycle is the same as the process, e.g. Cache and default Orphan
QUERY = 1, // Count the memory consumption of all Query tasks.
LOAD = 2, // Count the memory consumption of all Load tasks.
COMPACTION = 3, // Count the memory consumption of all Base and Cumulative tasks.
SCHEMA_CHANGE = 4, // Count the memory consumption of all SchemaChange tasks.
CLONE = 5, // Count the memory consumption of all EngineCloneTask. Note: Memory that does not contain make/release snapshots.
BATCHLOAD = 6, // Count the memory consumption of all EngineBatchLoadTask.
CONSISTENCY = 7 // Count the memory consumption of all EngineChecksumTask.
}
Object pointers are no longer saved between each layer, and the values of process and each type are periodically aggregated.
other fix:
In [fix](memtracker) Fix transmit_tracker null pointer because phamp is not thread safe #13528, I tried to separate the memory that was manually abandoned in the query from the orphan mem tracker. But in the actual test, the accuracy of this part of the memory cannot be guaranteed, so put it back to the orphan mem tracker again.
288 lines
11 KiB
C++
288 lines
11 KiB
C++
// Licensed to the Apache Software Foundation (ASF) under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing,
|
|
// software distributed under the License is distributed on an
|
|
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
// KIND, either express or implied. See the License for the
|
|
// specific language governing permissions and limitations
|
|
// under the License.
|
|
|
|
#include "scanner_scheduler.h"
|
|
|
|
#include "common/config.h"
|
|
#include "util/priority_thread_pool.hpp"
|
|
#include "util/priority_work_stealing_thread_pool.hpp"
|
|
#include "util/telemetry/telemetry.h"
|
|
#include "util/thread.h"
|
|
#include "util/threadpool.h"
|
|
#include "vec/core/block.h"
|
|
#include "vec/exec/scan/vscanner.h"
|
|
#include "vec/exprs/vexpr.h"
|
|
|
|
namespace doris::vectorized {
|
|
|
|
ScannerScheduler::ScannerScheduler() {}
|
|
|
|
ScannerScheduler::~ScannerScheduler() {
|
|
if (!_is_init) {
|
|
return;
|
|
}
|
|
|
|
for (int i = 0; i < QUEUE_NUM; i++) {
|
|
_pending_queues[i]->shutdown();
|
|
}
|
|
|
|
_is_closed = true;
|
|
|
|
_scheduler_pool->shutdown();
|
|
_local_scan_thread_pool->shutdown();
|
|
_remote_scan_thread_pool->shutdown();
|
|
|
|
_scheduler_pool->wait();
|
|
_local_scan_thread_pool->join();
|
|
_remote_scan_thread_pool->join();
|
|
|
|
for (int i = 0; i < QUEUE_NUM; i++) {
|
|
delete _pending_queues[i];
|
|
}
|
|
delete[] _pending_queues;
|
|
}
|
|
|
|
Status ScannerScheduler::init(ExecEnv* env) {
|
|
// 1. scheduling thread pool and scheduling queues
|
|
ThreadPoolBuilder("SchedulingThreadPool")
|
|
.set_min_threads(QUEUE_NUM)
|
|
.set_max_threads(QUEUE_NUM)
|
|
.build(&_scheduler_pool);
|
|
|
|
_pending_queues = new BlockingQueue<ScannerContext*>*[QUEUE_NUM];
|
|
for (int i = 0; i < QUEUE_NUM; i++) {
|
|
_pending_queues[i] = new BlockingQueue<ScannerContext*>(INT32_MAX);
|
|
_scheduler_pool->submit_func([this, i] { this->_schedule_thread(i); });
|
|
}
|
|
|
|
// 2. local scan thread pool
|
|
_local_scan_thread_pool.reset(new PriorityWorkStealingThreadPool(
|
|
config::doris_scanner_thread_pool_thread_num, env->store_paths().size(),
|
|
config::doris_scanner_thread_pool_queue_size, "local_scan"));
|
|
|
|
// 3. remote scan thread pool
|
|
_remote_scan_thread_pool.reset(
|
|
new PriorityThreadPool(config::doris_scanner_thread_pool_thread_num,
|
|
config::doris_scanner_thread_pool_queue_size, "remote_scan"));
|
|
|
|
_is_init = true;
|
|
return Status::OK();
|
|
}
|
|
|
|
Status ScannerScheduler::submit(ScannerContext* ctx) {
|
|
if (ctx->queue_idx == -1) {
|
|
ctx->queue_idx = (_queue_idx++ % QUEUE_NUM);
|
|
}
|
|
if (!_pending_queues[ctx->queue_idx]->blocking_put(ctx)) {
|
|
return Status::InternalError("failed to submit scanner context to scheduler");
|
|
}
|
|
return Status::OK();
|
|
}
|
|
|
|
void ScannerScheduler::_schedule_thread(int queue_id) {
|
|
BlockingQueue<ScannerContext*>* queue = _pending_queues[queue_id];
|
|
while (!_is_closed) {
|
|
ScannerContext* ctx;
|
|
bool ok = queue->blocking_get(&ctx);
|
|
if (!ok) {
|
|
// maybe closed
|
|
continue;
|
|
}
|
|
|
|
_schedule_scanners(ctx);
|
|
// If ctx is done, no need to schedule it again.
|
|
// But should notice that there may still scanners running in scanner pool.
|
|
}
|
|
return;
|
|
}
|
|
|
|
void ScannerScheduler::_schedule_scanners(ScannerContext* ctx) {
|
|
ctx->incr_num_ctx_scheduling(1);
|
|
if (ctx->done()) {
|
|
ctx->update_num_running(0, -1);
|
|
return;
|
|
}
|
|
|
|
std::list<VScanner*> this_run;
|
|
ctx->get_next_batch_of_scanners(&this_run);
|
|
if (this_run.empty()) {
|
|
// There will be 2 cases when this_run is empty:
|
|
// 1. The blocks queue reaches limit.
|
|
// The consumer will continue scheduling the ctx.
|
|
// 2. All scanners are running.
|
|
// There running scanner will schedule the ctx after they are finished.
|
|
// So here we just return to stop scheduling ctx.
|
|
ctx->update_num_running(0, -1);
|
|
return;
|
|
}
|
|
|
|
ctx->update_num_running(this_run.size(), -1);
|
|
// Submit scanners to thread pool
|
|
// TODO(cmy): How to handle this "nice"?
|
|
int nice = 1;
|
|
auto cur_span = opentelemetry::trace::Tracer::GetCurrentSpan();
|
|
auto iter = this_run.begin();
|
|
ctx->incr_num_scanner_scheduling(this_run.size());
|
|
if (ctx->thread_token != nullptr) {
|
|
while (iter != this_run.end()) {
|
|
(*iter)->start_wait_worker_timer();
|
|
auto s = ctx->thread_token->submit_func(
|
|
[this, scanner = *iter, parent_span = cur_span, ctx] {
|
|
opentelemetry::trace::Scope scope {parent_span};
|
|
this->_scanner_scan(this, ctx, scanner);
|
|
});
|
|
if (s.ok()) {
|
|
this_run.erase(iter++);
|
|
} else {
|
|
ctx->set_status_on_error(s);
|
|
break;
|
|
}
|
|
}
|
|
} else {
|
|
while (iter != this_run.end()) {
|
|
PriorityThreadPool::Task task;
|
|
task.work_function = [this, scanner = *iter, parent_span = cur_span, ctx] {
|
|
opentelemetry::trace::Scope scope {parent_span};
|
|
this->_scanner_scan(this, ctx, scanner);
|
|
};
|
|
task.priority = nice;
|
|
task.queue_id = (*iter)->queue_id();
|
|
(*iter)->start_wait_worker_timer();
|
|
|
|
TabletStorageType type = (*iter)->get_storage_type();
|
|
bool ret = false;
|
|
if (type == TabletStorageType::STORAGE_TYPE_LOCAL) {
|
|
ret = _local_scan_thread_pool->offer(task);
|
|
} else {
|
|
ret = _remote_scan_thread_pool->offer(task);
|
|
}
|
|
if (ret) {
|
|
this_run.erase(iter++);
|
|
} else {
|
|
ctx->set_status_on_error(
|
|
Status::InternalError("failed to submit scanner to scanner pool"));
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void ScannerScheduler::_scanner_scan(ScannerScheduler* scheduler, ScannerContext* ctx,
|
|
VScanner* scanner) {
|
|
INIT_AND_SCOPE_REENTRANT_SPAN_IF(ctx->state()->enable_profile(), ctx->state()->get_tracer(),
|
|
ctx->scan_span(), "VScanner::scan");
|
|
SCOPED_ATTACH_TASK(scanner->runtime_state());
|
|
SCOPED_CONSUME_MEM_TRACKER(scanner->runtime_state()->scanner_mem_tracker().get());
|
|
Thread::set_self_name("_scanner_scan");
|
|
scanner->update_wait_worker_timer();
|
|
// Do not use ScopedTimer. There is no guarantee that, the counter
|
|
// (_scan_cpu_timer, the class member) is not destroyed after `_running_thread==0`.
|
|
ThreadCpuStopWatch cpu_watch;
|
|
cpu_watch.start();
|
|
Status status = Status::OK();
|
|
bool eos = false;
|
|
RuntimeState* state = ctx->state();
|
|
DCHECK(nullptr != state);
|
|
if (!scanner->is_open()) {
|
|
status = scanner->open(state);
|
|
if (!status.ok()) {
|
|
ctx->set_status_on_error(status);
|
|
eos = true;
|
|
}
|
|
scanner->set_opened();
|
|
}
|
|
|
|
scanner->try_append_late_arrival_runtime_filter();
|
|
|
|
// Because we use thread pool to scan data from storage. One scanner can't
|
|
// use this thread too long, this can starve other query's scanner. So, we
|
|
// need yield this thread when we do enough work. However, OlapStorage read
|
|
// data in pre-aggregate mode, then we can't use storage returned data to
|
|
// judge if we need to yield. So we record all raw data read in this round
|
|
// scan, if this exceeds row number or bytes threshold, we yield this thread.
|
|
std::vector<vectorized::Block*> blocks;
|
|
int64_t raw_rows_read = scanner->raw_rows_read();
|
|
int64_t raw_rows_threshold = raw_rows_read + config::doris_scanner_row_num;
|
|
int64_t raw_bytes_read = 0;
|
|
int64_t raw_bytes_threshold = config::doris_scanner_row_bytes;
|
|
bool get_free_block = true;
|
|
int num_rows_in_block = 0;
|
|
|
|
// Only set to true when ctx->done() return true.
|
|
// Use this flag because we need distinguish eos from `should_stop`.
|
|
// If eos is true, we still need to return blocks,
|
|
// but is should_stop is true, no need to return blocks
|
|
bool should_stop = false;
|
|
// Has to wait at least one full block, or it will cause a lot of schedule task in priority
|
|
// queue, it will affect query latency and query concurrency for example ssb 3.3.
|
|
while (!eos && raw_bytes_read < raw_bytes_threshold &&
|
|
((raw_rows_read < raw_rows_threshold && get_free_block) ||
|
|
num_rows_in_block < state->batch_size())) {
|
|
if (UNLIKELY(ctx->done())) {
|
|
// No need to set status on error here.
|
|
// Because done() maybe caused by "should_stop"
|
|
should_stop = true;
|
|
break;
|
|
}
|
|
|
|
auto block = ctx->get_free_block(&get_free_block);
|
|
status = scanner->get_block(state, block, &eos);
|
|
VLOG_ROW << "VOlapScanNode input rows: " << block->rows() << ", eos: " << eos;
|
|
if (!status.ok()) {
|
|
LOG(WARNING) << "Scan thread read VOlapScanner failed: " << status.to_string();
|
|
// Add block ptr in blocks, prevent mem leak in read failed
|
|
blocks.push_back(block);
|
|
break;
|
|
}
|
|
|
|
raw_bytes_read += block->bytes();
|
|
num_rows_in_block += block->rows();
|
|
if (UNLIKELY(block->rows() == 0)) {
|
|
ctx->return_free_block(block);
|
|
} else {
|
|
if (!blocks.empty() && blocks.back()->rows() + block->rows() <= state->batch_size()) {
|
|
vectorized::MutableBlock(blocks.back()).merge(*block);
|
|
ctx->return_free_block(block);
|
|
} else {
|
|
blocks.push_back(block);
|
|
}
|
|
}
|
|
raw_rows_read = scanner->raw_rows_read();
|
|
} // end for while
|
|
|
|
// if we failed, check status.
|
|
if (UNLIKELY(!status.ok())) {
|
|
// _transfer_done = true;
|
|
ctx->set_status_on_error(status);
|
|
eos = true;
|
|
std::for_each(blocks.begin(), blocks.end(), std::default_delete<vectorized::Block>());
|
|
} else if (should_stop) {
|
|
// No need to return blocks because of should_stop, just delete them
|
|
std::for_each(blocks.begin(), blocks.end(), std::default_delete<vectorized::Block>());
|
|
} else if (!blocks.empty()) {
|
|
ctx->append_blocks_to_queue(blocks);
|
|
}
|
|
|
|
if (eos || should_stop) {
|
|
scanner->mark_to_need_to_close();
|
|
}
|
|
|
|
ctx->push_back_scanner_and_reschedule(scanner);
|
|
}
|
|
|
|
} // namespace doris::vectorized
|