Files
doris/be/src/exec/cross_join_node.cpp
HuangWei 10f822eb43 [MemTracker] make all MemTrackers shared (#4135)
We make all MemTrackers shared, in order to show MemTracker real-time consumptions on the web.
As follows:
1. nearly all MemTracker raw ptr -> shared_ptr
2. Use CreateTracker() to create new MemTracker(in order to add itself to its parent)
3. RowBatch & MemPool still use raw ptrs of MemTracker, it's easy to ensure RowBatch & MemPool destructor exec 
     before MemTracker's destructor. So we don't change these code.
4. MemTracker can use RuntimeProfile's counter to calc consumption. So RuntimeProfile's counter need to be shared 
    too. We add a shared counter pool to store the shared counter, don't change other counters of RuntimeProfile.
Note that, this PR doesn't change the MemTracker tree structure. So there still have some orphan trackers, e.g. RowBlockV2's MemTracker. If you find some shared MemTrackers are little memory consumption & too time-consuming, you could make them be the orphan, then it's fine to use the raw ptr.
2020-07-31 21:57:21 +08:00

204 lines
6.6 KiB
C++

// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
#include "exec/cross_join_node.h"
#include <sstream>
#include "exprs/expr.h"
#include "gen_cpp/PlanNodes_types.h"
#include "runtime/row_batch.h"
#include "runtime/runtime_state.h"
#include "util/debug_util.h"
#include "util/runtime_profile.h"
namespace doris {
CrossJoinNode::CrossJoinNode(
ObjectPool* pool, const TPlanNode& tnode, const DescriptorTbl& descs)
: BlockingJoinNode("CrossJoinNode", TJoinOp::CROSS_JOIN, pool, tnode, descs) {
}
Status CrossJoinNode::prepare(RuntimeState* state) {
DCHECK(_join_op == TJoinOp::CROSS_JOIN);
RETURN_IF_ERROR(BlockingJoinNode::prepare(state));
_build_batch_pool.reset(new ObjectPool());
return Status::OK();
}
Status CrossJoinNode::close(RuntimeState* state) {
// avoid double close
if (is_closed()) {
return Status::OK();
}
_build_batches.reset();
_build_batch_pool.reset();
BlockingJoinNode::close(state);
return Status::OK();
}
Status CrossJoinNode::construct_build_side(RuntimeState* state) {
// Do a full scan of child(1) and store all build row batches.
RETURN_IF_ERROR(child(1)->open(state));
while (true) {
RowBatch* batch = _build_batch_pool->add(
new RowBatch(child(1)->row_desc(), state->batch_size(), mem_tracker().get()));
RETURN_IF_CANCELLED(state);
// TODO(zhaochun):
// RETURN_IF_ERROR(state->CheckQueryState());
bool eos = true;
RETURN_IF_ERROR(child(1)->get_next(state, batch, &eos));
// to prevent use too many memory
RETURN_IF_LIMIT_EXCEEDED(state, "Cross join, while getting next from the child 1.");
SCOPED_TIMER(_build_timer);
_build_batches.add_row_batch(batch);
VLOG_ROW << build_list_debug_string();
COUNTER_SET(_build_row_counter,
static_cast<int64_t>(_build_batches.total_num_rows()));
if (eos) {
break;
}
}
return Status::OK();
}
void CrossJoinNode::init_get_next(TupleRow* first_left_row) {
_current_build_row = _build_batches.iterator();
}
Status CrossJoinNode::get_next(RuntimeState* state, RowBatch* output_batch, bool* eos) {
// RETURN_IF_ERROR(exec_debug_action(TExecNodePhase::GETNEXT, state));
RETURN_IF_CANCELLED(state);
*eos = false;
// TOOD(zhaochun)
// RETURN_IF_ERROR(state->check_query_state());
SCOPED_TIMER(_runtime_profile->total_time_counter());
if (reached_limit() || _eos) {
*eos = true;
return Status::OK();
}
ScopedTimer<MonotonicStopWatch> timer(_left_child_timer);
while (!_eos) {
// Compute max rows that should be added to output_batch
int64_t max_added_rows = output_batch->capacity() - output_batch->num_rows();
if (limit() != -1) {
max_added_rows = std::min(max_added_rows, limit() - rows_returned());
}
// Continue processing this row batch
_num_rows_returned +=
process_left_child_batch(output_batch, _left_batch.get(), max_added_rows);
COUNTER_SET(_rows_returned_counter, _num_rows_returned);
if (reached_limit() || output_batch->is_full()) {
*eos = reached_limit();
break;
}
// Check to see if we're done processing the current left child batch
if (_current_build_row.at_end() && _left_batch_pos == _left_batch->num_rows()) {
_left_batch->transfer_resource_ownership(output_batch);
_left_batch_pos = 0;
if (output_batch->is_full()) {
break;
}
if (_left_side_eos) {
*eos = _eos = true;
break;
} else {
timer.stop();
RETURN_IF_ERROR(child(0)->get_next(state, _left_batch.get(), &_left_side_eos));
timer.start();
COUNTER_UPDATE(_left_child_row_counter, _left_batch->num_rows());
}
}
}
return Status::OK();
}
std::string CrossJoinNode::build_list_debug_string() {
std::stringstream out;
out << "BuildList(";
out << _build_batches.debug_string(child(1)->row_desc());
out << ")";
return out.str();
}
// TODO: this can be replaced with a codegen'd function
int CrossJoinNode::process_left_child_batch(RowBatch* output_batch, RowBatch* batch,
int max_added_rows) {
int row_idx = output_batch->add_rows(max_added_rows);
DCHECK(row_idx != RowBatch::INVALID_ROW_INDEX);
uint8_t* output_row_mem = reinterpret_cast<uint8_t*>(output_batch->get_row(row_idx));
TupleRow* output_row = reinterpret_cast<TupleRow*>(output_row_mem);
int rows_returned = 0;
ExprContext* const* ctxs = &_conjunct_ctxs[0];
int ctx_size = _conjunct_ctxs.size();
while (true) {
while (!_current_build_row.at_end()) {
create_output_row(output_row, _current_left_child_row, _current_build_row.get_row());
_current_build_row.next();
if (!eval_conjuncts(ctxs, ctx_size, output_row)) {
continue;
}
++rows_returned;
// Filled up out batch or hit limit
if (UNLIKELY(rows_returned == max_added_rows)) {
output_batch->commit_rows(rows_returned);
return rows_returned;
}
// Advance to next out row
output_row_mem += output_batch->row_byte_size();
output_row = reinterpret_cast<TupleRow*>(output_row_mem);
}
DCHECK(_current_build_row.at_end());
// Advance to the next row in the left child batch
if (UNLIKELY(_left_batch_pos == batch->num_rows())) {
output_batch->commit_rows(rows_returned);
return rows_returned;
}
_current_left_child_row = batch->get_row(_left_batch_pos++);
_current_build_row = _build_batches.iterator();
}
output_batch->commit_rows(rows_returned);
return rows_returned;
}
}