imay commented on a change in pull request #1292: Support Grouping Sets, Rollup and Cube to extend group by statement URL: https://github.com/apache/incubator-doris/pull/1292#discussion_r294033200
########## File path: be/src/exec/repeat_node.cpp ########## @@ -0,0 +1,214 @@ +// Licensed to the Apache Software Foundation (ASF) under one +// or more contributor license agreements. See the NOTICE file +// distributed with this work for additional information +// regarding copyright ownership. The ASF licenses this file +// to you under the Apache License, Version 2.0 (the +// "License"); you may not use this file except in compliance +// with the License. You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, +// software distributed under the License is distributed on an +// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +// KIND, either express or implied. See the License for the +// specific language governing permissions and limitations +// under the License. + +#include "exec/repeat_node.h" + +#include "exprs/expr.h" +#include "runtime/raw_value.h" +#include "runtime/row_batch.h" +#include "runtime/runtime_state.h" +#include "util/runtime_profile.h" + +namespace doris { + +RepeatNode::RepeatNode(ObjectPool* pool, const TPlanNode& tnode, + const DescriptorTbl& descs) + : ExecNode(pool, tnode, descs), + _slot_id_set_list(tnode.repeat_node.slot_id_set_list), + _repeat_id_list(tnode.repeat_node.repeat_id_list), + _output_tuple_id(tnode.repeat_node.output_tuple_id), + _tuple_desc(nullptr), + _child_row_batch(nullptr), + _child_eos(false), + _repeat_id_idx(0), + _runtime_state(nullptr) { +} + +RepeatNode::~RepeatNode() { +} + +Status RepeatNode::prepare(RuntimeState* state) { + SCOPED_TIMER(_runtime_profile->total_time_counter()); + RETURN_IF_ERROR(ExecNode::prepare(state)); + + _runtime_state = state; + _tuple_desc = state->desc_tbl().get_tuple_descriptor(_output_tuple_id); + if (_tuple_desc == NULL) { + return Status("Failed to get tuple descriptor."); + } + + return Status::OK; +} + +Status RepeatNode::open(RuntimeState* state) { + SCOPED_TIMER(_runtime_profile->total_time_counter()); + RETURN_IF_ERROR(ExecNode::open(state)); + RETURN_IF_CANCELLED(state); + RETURN_IF_ERROR(child(0)->open(state)); + return Status::OK; +} + +Status RepeatNode::get_repeated_batch( + RowBatch* child_row_batch, int repeat_id_idx, RowBatch* row_batch) { + DCHECK(repeat_id_idx >= 0); + DCHECK(repeat_id_idx <= (int)_repeat_id_list.size()); + DCHECK(child_row_batch != nullptr); + DCHECK_EQ(row_batch->num_rows(), 0); + + // Fill all slots according to child + MemPool* tuple_pool = row_batch->tuple_data_pool(); + const vector<TupleDescriptor*>& src_tuple_descs = child_row_batch->row_desc().tuple_descriptors(); + const vector<TupleDescriptor*>& dst_tuple_descs = row_batch->row_desc().tuple_descriptors(); + vector<Tuple*> dst_tuples(src_tuple_descs.size(), nullptr); + for (int i = 0; i < child_row_batch->num_rows(); ++i) { + int row_idx = row_batch->add_row(); + TupleRow* dst_row = row_batch->get_row(row_idx); + TupleRow* src_row = child_row_batch->get_row(i); + + vector<TupleDescriptor*>::const_iterator src_it = src_tuple_descs.begin(); + vector<TupleDescriptor*>::const_iterator dst_it = dst_tuple_descs.begin(); + for (int j = 0; src_it != src_tuple_descs.end() && dst_it != dst_tuple_descs.end(); + ++src_it, ++dst_it, ++j) { + Tuple* src_tuple = src_row->get_tuple(j); + if (src_tuple == NULL) { + continue; + } + + if (dst_tuples[j] == nullptr) { + int size = row_batch->capacity() * (*dst_it)->byte_size(); + void* tuple_buffer = tuple_pool->allocate(size); + if (tuple_buffer == nullptr) { + return Status("Allocate memory for row batch failed."); + } + dst_tuples[j] = reinterpret_cast<Tuple*>(tuple_buffer); + } else { + char* new_tuple = reinterpret_cast<char*>(dst_tuples[j]); + new_tuple += (*dst_it)->byte_size(); + dst_tuples[j] = reinterpret_cast<Tuple*>(new_tuple); + } + dst_row->set_tuple(j, dst_tuples[j]); + memset(dst_tuples[j], 0, (*dst_it)->num_null_bytes()); + + for (int k = 0; k < (*src_it)->slots().size(); k++) { + SlotDescriptor* src_slot_desc = (*src_it)->slots()[k]; + SlotDescriptor* dst_slot_desc = (*dst_it)->slots()[k]; + DCHECK_EQ(src_slot_desc->type().type, dst_slot_desc->type().type); + DCHECK_EQ(src_slot_desc->col_name(), dst_slot_desc->col_name()); + + if (_slot_id_set_list[0].find(src_slot_desc->id()) != _slot_id_set_list[0].end()) { + std::set<SlotId> repeat_ids = _slot_id_set_list[repeat_id_idx]; + if (repeat_ids.find(src_slot_desc->id()) == repeat_ids.end()) { + dst_tuples[j]->set_null(dst_slot_desc->null_indicator_offset()); + continue; + } + } + + bool src_slot_null = src_tuple->is_null(src_slot_desc->null_indicator_offset()); + void* src_slot = NULL; + if (!src_slot_null) src_slot = src_tuple->get_slot(src_slot_desc->tuple_offset()); + RawValue::write(src_slot, dst_tuples[j], dst_slot_desc, tuple_pool); + } + } + row_batch->commit_last_row(); + } + + Tuple* tuple = nullptr; + int64_t groupingId = _repeat_id_list[repeat_id_idx]; + // Fill grouping ID to tuple + for (int i = 0; i < child_row_batch->num_rows(); ++i) { + int row_idx = i; + TupleRow* row = row_batch->get_row(row_idx); + + if (tuple == nullptr) { + int size = row_batch->capacity() * _tuple_desc->byte_size(); + void* tuple_buffer = tuple_pool->allocate(size); + if (tuple_buffer == nullptr) { + return Status("Allocate memory for row batch failed."); + } + tuple = reinterpret_cast<Tuple*>(tuple_buffer); + } else { + char* new_tuple = reinterpret_cast<char*>(tuple); + new_tuple += _tuple_desc->byte_size(); + tuple = reinterpret_cast<Tuple*>(new_tuple); + } + + row->set_tuple(src_tuple_descs.size(), tuple); + memset(tuple, 0, _tuple_desc->num_null_bytes()); + + // GROUPING__ID located in index 0 + const SlotDescriptor* slot_desc = _tuple_desc->slots()[0]; + tuple->set_not_null(slot_desc->null_indicator_offset()); + RawValue::write(&groupingId, tuple, slot_desc, tuple_pool); + } + + return Status::OK; +} + +Status RepeatNode::get_next(RuntimeState* state, RowBatch* row_batch, bool* eos) { + SCOPED_TIMER(_runtime_profile->total_time_counter()); + RETURN_IF_ERROR(exec_debug_action(TExecNodePhase::GETNEXT)); + RETURN_IF_CANCELLED(state); + + // current child has finished its repeat, get child's next batch + if (_child_row_batch.get() == nullptr) { + if (_child_eos) { + *eos = true; + return Status::OK; + } + + _child_row_batch.reset( + new RowBatch(child(0)->row_desc(), state->batch_size(), mem_tracker())); + RETURN_IF_ERROR(child(0)->get_next(state, _child_row_batch.get(), &_child_eos)); + + if (_child_row_batch->num_rows() <= 0) { + _child_row_batch.reset(nullptr); + *eos = true; + return Status::OK; + } + } + + DCHECK_EQ(row_batch->num_rows(), 0); + RETURN_IF_ERROR(get_repeated_batch(_child_row_batch.get(), _repeat_id_idx, row_batch)); + _repeat_id_idx++; + + int size = _repeat_id_list.size(); + if (_repeat_id_idx >= size) { + _child_row_batch.reset(nullptr); + _repeat_id_idx = 0; + } + + return Status::OK; +} + +Status RepeatNode::close(RuntimeState* state) { + if (is_closed()) { + return Status::OK; + } + RETURN_IF_ERROR(child(0)->close(state)); + return ExecNode::close(state); Review comment: you should reset _child_row_batch here to make memory leak check pass ---------------------------------------------------------------- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: [email protected] With regards, Apache Git Services --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
