Files
doris/be/src/vec/columns/column_jsonb.cpp

425 lines
14 KiB
C++

// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
#include "vec/columns/column_jsonb.h"
#include "util/jsonb_parser.h"
#include "vec/columns/column_string.h"
#include "vec/columns/columns_common.h"
#include "vec/common/arena.h"
#include "vec/common/assert_cast.h"
#include "vec/common/memcmp_small.h"
#include "vec/common/string_buffer.hpp"
#include "vec/common/unaligned.h"
namespace doris::vectorized {
MutableColumnPtr ColumnJsonb::clone_resized(size_t to_size) const {
auto res = ColumnJsonb::create();
if (to_size == 0) return res;
size_t from_size = size();
if (to_size <= from_size) {
/// Just cut column.
res->offsets.assign(offsets.begin(), offsets.begin() + to_size);
res->chars.assign(chars.begin(), chars.begin() + offsets[to_size - 1]);
} else {
Offset offset = 0;
if (from_size > 0) {
res->offsets.assign(offsets.begin(), offsets.end());
res->chars.assign(chars.begin(), chars.end());
offset = offsets.back();
}
/// Empty strings are just zero terminating bytes.
res->chars.resize_fill(res->chars.size() + to_size - from_size);
res->offsets.resize(to_size);
for (size_t i = from_size; i < to_size; ++i) {
++offset;
res->offsets[i] = offset;
}
res->offsets.resize_fill(to_size, chars.size());
}
return res;
}
void ColumnJsonb::insert_range_from(const IColumn& src, size_t start, size_t length) {
if (length == 0) return;
const ColumnJsonb& src_concrete = reinterpret_cast<const ColumnJsonb&>(src);
if (start + length > src_concrete.offsets.size()) {
LOG(FATAL) << "Parameter out of bound in IColumnJsonb::insert_range_from method.";
}
size_t nested_offset = src_concrete.offset_at(start);
size_t nested_length = src_concrete.offsets[start + length - 1] - nested_offset;
size_t old_chars_size = chars.size();
chars.resize(old_chars_size + nested_length);
memcpy(&chars[old_chars_size], &src_concrete.chars[nested_offset], nested_length);
if (start == 0 && offsets.empty()) {
offsets.assign(src_concrete.offsets.begin(), src_concrete.offsets.begin() + length);
} else {
size_t old_size = offsets.size();
size_t prev_max_offset = offsets.back(); /// -1th index is Ok, see PaddedPODArray
offsets.resize(old_size + length);
for (size_t i = 0; i < length; ++i)
offsets[old_size + i] =
src_concrete.offsets[start + i] - nested_offset + prev_max_offset;
}
}
void ColumnJsonb::insert_indices_from(const IColumn& src, const int* indices_begin,
const int* indices_end) {
for (auto x = indices_begin; x != indices_end; ++x) {
if (*x == -1) {
ColumnJsonb::insert_default();
} else {
ColumnJsonb::insert_from(src, *x);
}
}
}
ColumnPtr ColumnJsonb::filter(const Filter& filt, ssize_t result_size_hint) const {
if (offsets.size() == 0) return ColumnJsonb::create();
auto res = ColumnJsonb::create();
Chars& res_chars = res->chars;
Offsets& res_offsets = res->offsets;
filter_arrays_impl<UInt8>(chars, offsets, res_chars, res_offsets, filt, result_size_hint);
return res;
}
ColumnPtr ColumnJsonb::permute(const Permutation& perm, size_t limit) const {
size_t size = offsets.size();
if (limit == 0)
limit = size;
else
limit = std::min(size, limit);
if (perm.size() < limit) {
LOG(FATAL) << "Size of permutation is less than required.";
}
if (limit == 0) return ColumnJsonb::create();
auto res = ColumnJsonb::create();
Chars& res_chars = res->chars;
Offsets& res_offsets = res->offsets;
if (limit == size)
res_chars.resize(chars.size());
else {
size_t new_chars_size = 0;
for (size_t i = 0; i < limit; ++i) new_chars_size += size_at(perm[i]);
res_chars.resize(new_chars_size);
}
res_offsets.resize(limit);
Offset current_new_offset = 0;
for (size_t i = 0; i < limit; ++i) {
size_t j = perm[i];
size_t json_offset = offsets[j - 1];
size_t json_size = offsets[j] - json_offset;
memcpy_small_allow_read_write_overflow15(&res_chars[current_new_offset],
&chars[json_offset], json_size);
current_new_offset += json_size;
res_offsets[i] = current_new_offset;
}
return res;
}
StringRef ColumnJsonb::serialize_value_into_arena(size_t n, Arena& arena,
char const*& begin) const {
IColumn::Offset json_size = size_at(n);
size_t offset = offset_at(n);
StringRef res;
res.size = sizeof(json_size) + json_size;
char* pos = arena.alloc_continue(res.size, begin);
memcpy(pos, &json_size, sizeof(json_size));
memcpy(pos + sizeof(json_size), &chars[offset], json_size);
res.data = pos;
return res;
}
const char* ColumnJsonb::deserialize_and_insert_from_arena(const char* pos) {
const IColumn::Offset json_size = unaligned_load<size_t>(pos);
pos += sizeof(json_size);
const size_t old_size = chars.size();
const size_t new_size = old_size + json_size;
chars.resize(new_size);
memcpy(chars.data() + old_size, pos, json_size);
offsets.push_back(new_size);
return pos + json_size;
}
size_t ColumnJsonb::get_max_row_byte_size() const {
size_t max_size = 0;
size_t num_rows = offsets.size();
for (size_t i = 0; i < num_rows; ++i) {
max_size = std::max(max_size, size_at(i));
}
return max_size + sizeof(uint32_t);
}
void ColumnJsonb::serialize_vec(std::vector<StringRef>& keys, size_t num_rows,
size_t max_row_byte_size) const {
for (size_t i = 0; i < num_rows; ++i) {
uint32_t offset(offset_at(i));
uint32_t string_size(size_at(i));
auto* ptr = const_cast<char*>(keys[i].data + keys[i].size);
memcpy(ptr, &string_size, sizeof(string_size));
memcpy(ptr + sizeof(string_size), &chars[offset], string_size);
keys[i].size += sizeof(string_size) + string_size;
}
}
void ColumnJsonb::serialize_vec_with_null_map(std::vector<StringRef>& keys, size_t num_rows,
const uint8_t* null_map,
size_t max_row_byte_size) const {
for (size_t i = 0; i < num_rows; ++i) {
if (null_map[i] == 0) {
uint32_t offset(offset_at(i));
uint32_t string_size(size_at(i));
auto* ptr = const_cast<char*>(keys[i].data + keys[i].size);
memcpy(ptr, &string_size, sizeof(string_size));
memcpy(ptr + sizeof(string_size), &chars[offset], string_size);
keys[i].size += sizeof(string_size) + string_size;
}
}
}
template <typename Type>
ColumnPtr ColumnJsonb::index_impl(const PaddedPODArray<Type>& indexes, size_t limit) const {
if (limit == 0) return ColumnJsonb::create();
auto res = ColumnJsonb::create();
Chars& res_chars = res->chars;
Offsets& res_offsets = res->offsets;
size_t new_chars_size = 0;
for (size_t i = 0; i < limit; ++i) new_chars_size += size_at(indexes[i]);
res_chars.resize(new_chars_size);
res_offsets.resize(limit);
Offset current_new_offset = 0;
for (size_t i = 0; i < limit; ++i) {
size_t j = indexes[i];
size_t json_offset = offsets[j - 1];
size_t json_size = offsets[j] - json_offset;
memcpy_small_allow_read_write_overflow15(&res_chars[current_new_offset],
&chars[json_offset], json_size);
current_new_offset += json_size;
res_offsets[i] = current_new_offset;
}
return res;
}
template <bool positive>
struct ColumnJsonb::less {
const ColumnJsonb& parent;
explicit less(const ColumnJsonb& parent_) : parent(parent_) {}
bool operator()(size_t lhs, size_t rhs) const {
int res = memcmp_small_allow_overflow15(
parent.chars.data() + parent.offset_at(lhs), parent.size_at(lhs) - 1,
parent.chars.data() + parent.offset_at(rhs), parent.size_at(rhs) - 1);
return positive ? (res < 0) : (res > 0);
}
};
void ColumnJsonb::get_permutation(bool reverse, size_t limit, int /*nan_direction_hint*/,
Permutation& res) const {
size_t s = offsets.size();
res.resize(s);
for (size_t i = 0; i < s; ++i) res[i] = i;
if (limit >= s) limit = 0;
if (limit) {
if (reverse)
std::partial_sort(res.begin(), res.begin() + limit, res.end(), less<false>(*this));
else
std::partial_sort(res.begin(), res.begin() + limit, res.end(), less<true>(*this));
} else {
if (reverse)
std::sort(res.begin(), res.end(), less<false>(*this));
else
std::sort(res.begin(), res.end(), less<true>(*this));
}
}
ColumnPtr ColumnJsonb::replicate(const Offsets& replicate_offsets) const {
size_t col_size = size();
if (col_size != replicate_offsets.size()) {
LOG(FATAL) << "Size of offsets doesn't match size of column.";
}
auto res = ColumnJsonb::create();
if (0 == col_size) return res;
Chars& res_chars = res->chars;
Offsets& res_offsets = res->offsets;
res_chars.reserve(chars.size() / col_size * replicate_offsets.back());
res_offsets.reserve(replicate_offsets.back());
Offset prev_replicate_offset = 0;
Offset prev_json_offset = 0;
Offset current_new_offset = 0;
for (size_t i = 0; i < col_size; ++i) {
size_t size_to_replicate = replicate_offsets[i] - prev_replicate_offset;
size_t json_size = offsets[i] - prev_json_offset;
for (size_t j = 0; j < size_to_replicate; ++j) {
current_new_offset += json_size;
res_offsets.push_back(current_new_offset);
res_chars.resize(res_chars.size() + json_size);
memcpy_small_allow_read_write_overflow15(&res_chars[res_chars.size() - json_size],
&chars[prev_json_offset], json_size);
}
prev_replicate_offset = replicate_offsets[i];
prev_json_offset = offsets[i];
}
return res;
}
void ColumnJsonb::replicate(const uint32_t* counts, size_t target_size, IColumn& column,
size_t begin, int count_sz) const {
size_t col_size = count_sz < 0 ? size() : count_sz;
if (0 == col_size) return;
auto& res = reinterpret_cast<ColumnJsonb&>(column);
Chars& res_chars = res.chars;
Offsets& res_offsets = res.offsets;
res_chars.reserve(chars.size() / col_size * target_size);
res_offsets.reserve(target_size);
size_t base = begin > 0 ? offset_at(begin - 1) : 0;
Offset prev_json_offset = 0 + base;
Offset current_new_offset = 0;
size_t end = begin + col_size;
for (size_t i = begin; i < end; ++i) {
size_t size_to_replicate = counts[i];
size_t json_size = offsets[i] - prev_json_offset;
for (size_t j = 0; j < size_to_replicate; ++j) {
current_new_offset += json_size;
res_offsets.push_back(current_new_offset);
res_chars.resize(res_chars.size() + json_size);
memcpy_small_allow_read_write_overflow15(&res_chars[res_chars.size() - json_size],
&chars[prev_json_offset], json_size);
}
prev_json_offset = offsets[i];
}
}
void ColumnJsonb::reserve(size_t n) {
offsets.reserve(n);
chars.reserve(n);
}
MutableColumnPtr ColumnJsonb::get_shrinked_column() {
auto shrinked_column = ColumnJsonb::create();
shrinked_column->get_offsets().reserve(offsets.size());
shrinked_column->get_chars().reserve(chars.size());
for (int i = 0; i < size(); i++) {
StringRef str = get_data_at(i);
reinterpret_cast<ColumnJsonb*>(shrinked_column.get())
->insert_data(str.data, strnlen(str.data, str.size));
}
return shrinked_column;
}
void ColumnJsonb::resize(size_t n) {
auto origin_size = size();
if (origin_size > n) {
offsets.resize(n);
} else if (origin_size < n) {
insert_many_defaults(n - origin_size);
}
}
void ColumnJsonb::get_extremes(Field& min, Field& max) const {
min = String();
max = String();
size_t col_size = size();
if (col_size == 0) return;
size_t min_idx = 0;
size_t max_idx = 0;
less<true> less_op(*this);
for (size_t i = 1; i < col_size; ++i) {
if (less_op(i, min_idx))
min_idx = i;
else if (less_op(max_idx, i))
max_idx = i;
}
get(min_idx, min);
get(max_idx, max);
}
void ColumnJsonb::protect() {
get_chars().protect();
get_offsets().protect();
}
} // namespace doris::vectorized