1842 lines
80 KiB
C++
1842 lines
80 KiB
C++
// Licensed to the Apache Software Foundation (ASF) under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing,
|
|
// software distributed under the License is distributed on an
|
|
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
|
// KIND, either express or implied. See the License for the
|
|
// specific language governing permissions and limitations
|
|
// under the License.
|
|
|
|
#include <fmt/core.h>
|
|
#include <stdint.h>
|
|
|
|
#include <algorithm>
|
|
#include <cctype>
|
|
// IWYU pragma: no_include <bthread/errno.h>
|
|
#include <lz4/lz4hc.h>
|
|
|
|
#include <cerrno> // IWYU pragma: keep
|
|
#include <cstdlib>
|
|
#include <cstring>
|
|
#include <fstream> // IWYU pragma: keep
|
|
#include <functional>
|
|
#include <iostream>
|
|
#include <map>
|
|
#include <memory>
|
|
#include <mutex>
|
|
#include <random>
|
|
#include <string>
|
|
#include <utility>
|
|
#include <vector>
|
|
|
|
#include "common/config.h"
|
|
#include "common/logging.h"
|
|
#include "common/status.h"
|
|
#include "config.h"
|
|
#include "io/fs/file_writer.h"
|
|
#include "io/fs/local_file_system.h"
|
|
#include "util/cpu_info.h"
|
|
|
|
namespace doris::config {
|
|
|
|
// Dir of custom config file
|
|
DEFINE_String(custom_config_dir, "${DORIS_HOME}/conf");
|
|
|
|
// Dir of jdbc drivers
|
|
DEFINE_String(jdbc_drivers_dir, "${DORIS_HOME}/jdbc_drivers");
|
|
|
|
// cluster id
|
|
DEFINE_Int32(cluster_id, "-1");
|
|
// port on which BackendService is exported
|
|
DEFINE_Int32(be_port, "9060");
|
|
|
|
// port for brpc
|
|
DEFINE_Int32(brpc_port, "8060");
|
|
|
|
DEFINE_Int32(arrow_flight_sql_port, "-1");
|
|
|
|
// If the external client cannot directly access priority_networks, set public_host to be accessible
|
|
// to external client.
|
|
// There are usually two usage scenarios:
|
|
// 1. in production environment, it is often inconvenient to expose Doris BE nodes to the external network.
|
|
// However, a reverse proxy (such as Nginx) can be added to all Doris BE nodes, and the external client will be
|
|
// randomly routed to a Doris BE node when connecting to Nginx. set public_host to the host of Nginx.
|
|
// 2. if priority_networks is an internal network IP, and BE node has its own independent external IP,
|
|
// but Doris currently does not support modifying priority_networks, setting public_host to the real external IP.
|
|
DEFINE_mString(public_host, "");
|
|
|
|
// If the BE node is connected to the external network through a reverse proxy like Nginx
|
|
// and need to use Arrow Flight SQL, should add a server in Nginx to reverse proxy
|
|
// `Nginx:arrow_flight_sql_proxy_port` to `BE_priority_networks:arrow_flight_sql_port`. For example:
|
|
// upstream arrowflight {
|
|
// server 10.16.10.8:8069;
|
|
// server 10.16.10.8:8068;
|
|
//}
|
|
// server {
|
|
// listen 8167 http2;
|
|
// listen [::]:8167 http2;
|
|
// server_name doris.arrowflight.com;
|
|
// }
|
|
DEFINE_Int32(arrow_flight_sql_proxy_port, "-1");
|
|
|
|
// the number of bthreads for brpc, the default value is set to -1,
|
|
// which means the number of bthreads is #cpu-cores
|
|
DEFINE_Int32(brpc_num_threads, "256");
|
|
// the time of brpc server keep idle connection, setting this value too small may cause rpc between backends to fail,
|
|
// the default value is set to -1, which means never close idle connection.
|
|
DEFINE_Int32(brpc_idle_timeout_sec, "-1");
|
|
|
|
// Declare a selection strategy for those servers have many ips.
|
|
// Note that there should at most one ip match this list.
|
|
// this is a list in semicolon-delimited format, in CIDR notation, e.g. 10.10.10.0/24
|
|
// If no ip match this rule, will choose one randomly.
|
|
DEFINE_String(priority_networks, "");
|
|
|
|
// memory mode
|
|
// performance or compact
|
|
DEFINE_String(memory_mode, "moderate");
|
|
|
|
DEFINE_mBool(enable_use_cgroup_memory_info, "true");
|
|
|
|
// process memory limit specified as number of bytes
|
|
// ('<int>[bB]?'), megabytes ('<float>[mM]'), gigabytes ('<float>[gG]'),
|
|
// or percentage of the physical memory ('<int>%').
|
|
// defaults to bytes if no unit is given"
|
|
// must larger than 0. and if larger than physical memory size,
|
|
// it will be set to physical memory size.
|
|
DEFINE_String(mem_limit, "90%");
|
|
|
|
// Soft memory limit as a fraction of hard memory limit.
|
|
DEFINE_Double(soft_mem_limit_frac, "0.9");
|
|
|
|
// Schema change memory limit as a fraction of soft memory limit.
|
|
DEFINE_Double(schema_change_mem_limit_frac, "0.6");
|
|
|
|
// Many modern allocators (for example, tcmalloc) do not do a mremap for
|
|
// realloc, even in case of large enough chunks of memory. Although this allows
|
|
// you to increase performance and reduce memory consumption during realloc.
|
|
// To fix this, we do mremap manually if the chunk of memory is large enough.
|
|
//
|
|
// The threshold (128 MB, 128 * (1ULL << 20)) is chosen quite large, since changing the address
|
|
// space is very slow, especially in the case of a large number of threads. We
|
|
// expect that the set of operations mmap/something to do/mremap can only be
|
|
// performed about 1000 times per second.
|
|
//
|
|
// P.S. This is also required, because tcmalloc can not allocate a chunk of
|
|
// memory greater than 16 GB.
|
|
DEFINE_mInt64(mmap_threshold, "134217728"); // bytes
|
|
|
|
// When hash table capacity is greater than 2^double_grow_degree(default 2G), grow when 75% of the capacity is satisfied.
|
|
// Increase can reduce the number of hash table resize, but may waste more memory.
|
|
DEFINE_mInt32(hash_table_double_grow_degree, "31");
|
|
|
|
DEFINE_mInt32(max_fill_rate, "2");
|
|
|
|
DEFINE_mInt32(double_resize_threshold, "23");
|
|
|
|
// The maximum low water mark of the system `/proc/meminfo/MemAvailable`, Unit byte, default -1.
|
|
// if it is -1, then low water mark = min(MemTotal - MemLimit, MemTotal * 5%), which is 3.2G on a 64G machine.
|
|
// Turn up max. more memory buffers will be reserved for Memory GC.
|
|
// Turn down max. will use as much memory as possible.
|
|
// note that: `max_` prefix should be removed, but keep it for compatibility.
|
|
DEFINE_Int64(max_sys_mem_available_low_water_mark_bytes, "-1");
|
|
|
|
DEFINE_Int64(memtable_limiter_reserved_memory_bytes, "838860800");
|
|
|
|
// The size of the memory that gc wants to release each time, as a percentage of the mem limit.
|
|
DEFINE_mString(process_minor_gc_size, "5%");
|
|
DEFINE_mString(process_full_gc_size, "10%");
|
|
|
|
// If true, when the process does not exceed the soft mem limit, the query memory will not be limited;
|
|
// when the process memory exceeds the soft mem limit, the query with the largest ratio between the currently
|
|
// used memory and the exec_mem_limit will be canceled.
|
|
// If false, cancel query when the memory used exceeds exec_mem_limit, same as before.
|
|
DEFINE_mBool(enable_query_memory_overcommit, "true");
|
|
|
|
DEFINE_mBool(disable_memory_gc, "false");
|
|
|
|
DEFINE_mInt64(stacktrace_in_alloc_large_memory_bytes, "2147483648");
|
|
|
|
DEFINE_mInt64(crash_in_alloc_large_memory_bytes, "-1");
|
|
|
|
// If memory tracker value is inaccurate, BE will crash. usually used in test environments, default value is false.
|
|
DEFINE_mBool(crash_in_memory_tracker_inaccurate, "false");
|
|
|
|
// default is true. if any memory tracking in Orphan mem tracker will report error.
|
|
// !! not modify the default value of this conf!! otherwise memory errors cannot be detected in time.
|
|
// allocator free memory not need to check, because when the thread memory tracker label is Orphan,
|
|
// use the tracker saved in Allocator.
|
|
DEFINE_mBool(enable_memory_orphan_check, "true");
|
|
|
|
// The maximum time a thread waits for full GC. Currently only query will wait for full gc.
|
|
DEFINE_mInt32(thread_wait_gc_max_milliseconds, "1000");
|
|
|
|
DEFINE_mInt64(pre_serialize_keys_limit_bytes, "16777216");
|
|
|
|
// the port heartbeat service used
|
|
DEFINE_Int32(heartbeat_service_port, "9050");
|
|
// the count of heart beat service
|
|
DEFINE_Int32(heartbeat_service_thread_count, "1");
|
|
// the count of thread to create table
|
|
DEFINE_Int32(create_tablet_worker_count, "3");
|
|
// the count of thread to drop table
|
|
DEFINE_Int32(drop_tablet_worker_count, "3");
|
|
// the count of thread to batch load
|
|
DEFINE_Int32(push_worker_count_normal_priority, "3");
|
|
// the count of thread to high priority batch load
|
|
DEFINE_Int32(push_worker_count_high_priority, "3");
|
|
// the count of thread to publish version
|
|
DEFINE_Int32(publish_version_worker_count, "8");
|
|
// the count of tablet thread to publish version
|
|
DEFINE_Int32(tablet_publish_txn_max_thread, "32");
|
|
// the timeout of EnginPublishVersionTask
|
|
DEFINE_Int32(publish_version_task_timeout_s, "8");
|
|
// the count of thread to calc delete bitmap
|
|
DEFINE_Int32(calc_delete_bitmap_max_thread, "32");
|
|
// the count of thread to clear transaction task
|
|
DEFINE_Int32(clear_transaction_task_worker_count, "1");
|
|
// the count of thread to delete
|
|
DEFINE_Int32(delete_worker_count, "3");
|
|
// the count of thread to alter table
|
|
DEFINE_Int32(alter_tablet_worker_count, "3");
|
|
// the count of thread to alter index
|
|
DEFINE_Int32(alter_index_worker_count, "3");
|
|
// the count of thread to clone
|
|
DEFINE_Int32(clone_worker_count, "3");
|
|
// the count of thread to clone
|
|
DEFINE_Int32(storage_medium_migrate_count, "1");
|
|
// the count of thread to check consistency
|
|
DEFINE_Int32(check_consistency_worker_count, "1");
|
|
// the count of thread to upload
|
|
DEFINE_Int32(upload_worker_count, "1");
|
|
// the count of thread to download
|
|
DEFINE_Int32(download_worker_count, "1");
|
|
// the count of thread to make snapshot
|
|
DEFINE_Int32(make_snapshot_worker_count, "5");
|
|
// the count of thread to release snapshot
|
|
DEFINE_Int32(release_snapshot_worker_count, "5");
|
|
// report random wait a little time to avoid FE receiving multiple be reports at the same time.
|
|
// do not set it to false for production environment
|
|
DEFINE_mBool(report_random_wait, "true");
|
|
// the interval time(seconds) for agent report tasks signature to FE
|
|
DEFINE_mInt32(report_task_interval_seconds, "10");
|
|
// the interval time(seconds) for refresh storage policy from FE
|
|
DEFINE_mInt32(storage_refresh_storage_policy_task_interval_seconds, "5");
|
|
// the interval time(seconds) for agent report disk state to FE
|
|
DEFINE_mInt32(report_disk_state_interval_seconds, "60");
|
|
// the interval time(seconds) for agent report olap table to FE
|
|
DEFINE_mInt32(report_tablet_interval_seconds, "60");
|
|
// the max download speed(KB/s)
|
|
DEFINE_mInt32(max_download_speed_kbps, "50000");
|
|
// download low speed limit(KB/s)
|
|
DEFINE_mInt32(download_low_speed_limit_kbps, "50");
|
|
// download low speed time(seconds)
|
|
DEFINE_mInt32(download_low_speed_time, "300");
|
|
|
|
DEFINE_String(sys_log_dir, "");
|
|
DEFINE_String(user_function_dir, "${DORIS_HOME}/lib/udf");
|
|
// INFO, WARNING, ERROR, FATAL
|
|
DEFINE_mString(sys_log_level, "INFO");
|
|
// TIME-DAY, TIME-HOUR, SIZE-MB-nnn
|
|
DEFINE_String(sys_log_roll_mode, "SIZE-MB-1024");
|
|
// log roll num
|
|
DEFINE_Int32(sys_log_roll_num, "10");
|
|
// verbose log
|
|
DEFINE_Strings(sys_log_verbose_modules, "");
|
|
// verbose log level
|
|
DEFINE_Int32(sys_log_verbose_level, "10");
|
|
// verbose log FLAGS_v
|
|
DEFINE_Int32(sys_log_verbose_flags_v, "-1");
|
|
// log buffer level
|
|
DEFINE_String(log_buffer_level, "");
|
|
// log enable custom date time format
|
|
DEFINE_Bool(sys_log_enable_custom_date_time_format, "false");
|
|
// log custom date time format (https://en.cppreference.com/w/cpp/io/manip/put_time)
|
|
DEFINE_String(sys_log_custom_date_time_format, "%Y-%m-%d %H:%M:%S");
|
|
// log custom date time milliseconds format (fmt::format)
|
|
DEFINE_String(sys_log_custom_date_time_ms_format, ",{:03d}");
|
|
|
|
// number of threads available to serve backend execution requests
|
|
DEFINE_Int32(be_service_threads, "64");
|
|
|
|
// interval between profile reports; in seconds
|
|
DEFINE_mInt32(status_report_interval, "5");
|
|
// The pipeline task has a high concurrency, therefore reducing its report frequency
|
|
DEFINE_mInt32(pipeline_status_report_interval, "10");
|
|
// if true, each disk will have a separate thread pool for scanner
|
|
DEFINE_Bool(doris_enable_scanner_thread_pool_per_disk, "true");
|
|
// the timeout of a work thread to wait the blocking priority queue to get a task
|
|
DEFINE_mInt64(doris_blocking_priority_queue_wait_timeout_ms, "500");
|
|
// number of scanner thread pool size for olap table
|
|
// and the min thread num of remote scanner thread pool
|
|
DEFINE_Int32(doris_scanner_thread_pool_thread_num, "-1");
|
|
DEFINE_Validator(doris_scanner_thread_pool_thread_num, [](const int config) -> bool {
|
|
if (config == -1) {
|
|
CpuInfo::init();
|
|
doris_scanner_thread_pool_thread_num = std::max(48, CpuInfo::num_cores() * 2);
|
|
}
|
|
return true;
|
|
});
|
|
DEFINE_Int32(doris_scanner_min_thread_pool_thread_num, "8");
|
|
DEFINE_Int32(remote_split_source_batch_size, "10240");
|
|
DEFINE_Int32(doris_max_remote_scanner_thread_pool_thread_num, "-1");
|
|
// number of olap scanner thread pool queue size
|
|
DEFINE_Int32(doris_scanner_thread_pool_queue_size, "102400");
|
|
// default thrift client connect timeout(in seconds)
|
|
DEFINE_mInt32(thrift_connect_timeout_seconds, "3");
|
|
DEFINE_mInt32(fetch_rpc_timeout_seconds, "30");
|
|
|
|
// default thrift client retry interval (in milliseconds)
|
|
DEFINE_mInt64(thrift_client_retry_interval_ms, "1000");
|
|
// max message size of thrift request
|
|
// default: 100 * 1024 * 1024
|
|
DEFINE_mInt64(thrift_max_message_size, "104857600");
|
|
// max row count number for single scan range, used in segmentv1
|
|
DEFINE_mInt32(doris_scan_range_row_count, "524288");
|
|
// max bytes number for single scan range, used in segmentv2
|
|
DEFINE_mInt32(doris_scan_range_max_mb, "1024");
|
|
// max bytes number for single scan block, used in segmentv2
|
|
DEFINE_mInt32(doris_scan_block_max_mb, "67108864");
|
|
// size of scanner queue between scanner thread and compute thread
|
|
DEFINE_mInt32(doris_scanner_queue_size, "1024");
|
|
// single read execute fragment row number
|
|
DEFINE_mInt32(doris_scanner_row_num, "16384");
|
|
// single read execute fragment row bytes
|
|
DEFINE_mInt32(doris_scanner_row_bytes, "10485760");
|
|
// single read execute fragment max run time millseconds
|
|
DEFINE_mInt32(doris_scanner_max_run_time_ms, "1000");
|
|
DEFINE_mInt32(min_bytes_in_scanner_queue, "67108864");
|
|
// number of max scan keys
|
|
DEFINE_mInt32(doris_max_scan_key_num, "48");
|
|
// the max number of push down values of a single column.
|
|
// if exceed, no conditions will be pushed down for that column.
|
|
DEFINE_mInt32(max_pushdown_conditions_per_column, "1024");
|
|
// (Advanced) Maximum size of per-query receive-side buffer
|
|
DEFINE_mInt32(exchg_node_buffer_size_bytes, "20485760");
|
|
DEFINE_mInt32(exchg_buffer_queue_capacity_factor, "64");
|
|
|
|
DEFINE_mInt64(column_dictionary_key_ratio_threshold, "0");
|
|
DEFINE_mInt64(column_dictionary_key_size_threshold, "0");
|
|
// memory_limitation_per_thread_for_schema_change_bytes unit bytes
|
|
DEFINE_mInt64(memory_limitation_per_thread_for_schema_change_bytes, "2147483648");
|
|
DEFINE_mInt64(memory_limitation_per_thread_for_storage_migration_bytes, "100000000");
|
|
|
|
DEFINE_mInt32(cache_prune_interval_sec, "10");
|
|
DEFINE_mInt32(cache_periodic_prune_stale_sweep_sec, "300");
|
|
// the clean interval of tablet lookup cache
|
|
DEFINE_mInt32(tablet_lookup_cache_stale_sweep_time_sec, "30");
|
|
DEFINE_mInt32(point_query_row_cache_stale_sweep_time_sec, "300");
|
|
DEFINE_mInt32(disk_stat_monitor_interval, "5");
|
|
DEFINE_mInt32(unused_rowset_monitor_interval, "30");
|
|
DEFINE_String(storage_root_path, "${DORIS_HOME}/storage");
|
|
DEFINE_mString(broken_storage_path, "");
|
|
|
|
// Config is used to check incompatible old format hdr_ format
|
|
// whether doris uses strict way. When config is true, process will log fatal
|
|
// and exit. When config is false, process will only log warning.
|
|
DEFINE_Bool(storage_strict_check_incompatible_old_format, "true");
|
|
|
|
// BE process will exit if the percentage of error disk reach this value.
|
|
DEFINE_mInt32(max_percentage_of_error_disk, "100");
|
|
DEFINE_mInt32(default_num_rows_per_column_file_block, "1024");
|
|
// pending data policy
|
|
DEFINE_mInt32(pending_data_expire_time_sec, "1800");
|
|
// inc_rowset snapshot rs sweep time interval
|
|
DEFINE_mInt32(tablet_rowset_stale_sweep_time_sec, "300");
|
|
// tablet stale rowset sweep by threshold size
|
|
DEFINE_Bool(tablet_rowset_stale_sweep_by_size, "false");
|
|
DEFINE_mInt32(tablet_rowset_stale_sweep_threshold_size, "100");
|
|
// garbage sweep policy
|
|
DEFINE_Int32(max_garbage_sweep_interval, "3600");
|
|
DEFINE_Int32(min_garbage_sweep_interval, "180");
|
|
DEFINE_mInt32(garbage_sweep_batch_size, "100");
|
|
DEFINE_mInt32(snapshot_expire_time_sec, "172800");
|
|
// It is only a recommended value. When the disk space is insufficient,
|
|
// the file storage period under trash dose not have to comply with this parameter.
|
|
DEFINE_mInt32(trash_file_expire_time_sec, "86400");
|
|
// minimum file descriptor number
|
|
// modify them upon necessity
|
|
DEFINE_Int32(min_file_descriptor_number, "60000");
|
|
DEFINE_mBool(disable_segment_cache, "false");
|
|
DEFINE_String(row_cache_mem_limit, "20%");
|
|
|
|
// Cache for storage page size
|
|
DEFINE_String(storage_page_cache_limit, "20%");
|
|
// Shard size for page cache, the value must be power of two.
|
|
// It's recommended to set it to a value close to the number of BE cores in order to reduce lock contentions.
|
|
DEFINE_Int32(storage_page_cache_shard_size, "256");
|
|
// Percentage for index page cache
|
|
// all storage page cache will be divided into data_page_cache and index_page_cache
|
|
DEFINE_Int32(index_page_cache_percentage, "10");
|
|
// whether to disable page cache feature in storage
|
|
DEFINE_mBool(disable_storage_page_cache, "false");
|
|
// whether to disable row cache feature in storage
|
|
DEFINE_mBool(disable_storage_row_cache, "true");
|
|
// whether to disable pk page cache feature in storage
|
|
DEFINE_Bool(disable_pk_storage_page_cache, "false");
|
|
|
|
// Cache for mow primary key storage page size
|
|
DEFINE_String(pk_storage_page_cache_limit, "10%");
|
|
// data page size for primary key index
|
|
DEFINE_Int32(primary_key_data_page_size, "32768");
|
|
|
|
DEFINE_mInt32(data_page_cache_stale_sweep_time_sec, "300");
|
|
DEFINE_mInt32(index_page_cache_stale_sweep_time_sec, "600");
|
|
DEFINE_mInt32(pk_index_page_cache_stale_sweep_time_sec, "600");
|
|
|
|
DEFINE_Bool(enable_low_cardinality_optimize, "true");
|
|
DEFINE_Bool(enable_low_cardinality_cache_code, "true");
|
|
|
|
// be policy
|
|
// whether check compaction checksum
|
|
DEFINE_mBool(enable_compaction_checksum, "false");
|
|
// whether disable automatic compaction task
|
|
DEFINE_mBool(disable_auto_compaction, "false");
|
|
// whether enable vertical compaction
|
|
DEFINE_mBool(enable_vertical_compaction, "true");
|
|
// whether enable ordered data compaction
|
|
DEFINE_mBool(enable_ordered_data_compaction, "true");
|
|
// In vertical compaction, column number for every group
|
|
DEFINE_mInt32(vertical_compaction_num_columns_per_group, "5");
|
|
// In vertical compaction, max memory usage for row_source_buffer
|
|
DEFINE_Int32(vertical_compaction_max_row_source_memory_mb, "200");
|
|
// In vertical compaction, max dest segment file size
|
|
DEFINE_mInt64(vertical_compaction_max_segment_size, "1073741824");
|
|
|
|
// If enabled, segments will be flushed column by column
|
|
DEFINE_mBool(enable_vertical_segment_writer, "true");
|
|
|
|
// In ordered data compaction, min segment size for input rowset
|
|
DEFINE_mInt32(ordered_data_compaction_min_segment_size, "10485760");
|
|
|
|
// This config can be set to limit thread number in compaction thread pool.
|
|
DEFINE_mInt32(max_base_compaction_threads, "4");
|
|
DEFINE_mInt32(max_cumu_compaction_threads, "-1");
|
|
DEFINE_mInt32(max_single_replica_compaction_threads, "-1");
|
|
|
|
DEFINE_Bool(enable_base_compaction_idle_sched, "true");
|
|
DEFINE_mInt64(base_compaction_min_rowset_num, "5");
|
|
DEFINE_mInt64(base_compaction_max_compaction_score, "20");
|
|
DEFINE_mDouble(base_compaction_min_data_ratio, "0.3");
|
|
DEFINE_mInt64(base_compaction_dup_key_max_file_size_mbytes, "1024");
|
|
|
|
DEFINE_Bool(enable_skip_tablet_compaction, "true");
|
|
DEFINE_mInt32(skip_tablet_compaction_second, "10");
|
|
|
|
// output rowset of cumulative compaction total disk size exceed this config size,
|
|
// this rowset will be given to base compaction, unit is m byte.
|
|
DEFINE_mInt64(compaction_promotion_size_mbytes, "1024");
|
|
|
|
// output rowset of cumulative compaction total disk size exceed this config ratio of
|
|
// base rowset's total disk size, this rowset will be given to base compaction. The value must be between
|
|
// 0 and 1.
|
|
DEFINE_mDouble(compaction_promotion_ratio, "0.05");
|
|
|
|
// the smallest size of rowset promotion. When the rowset is less than this config, this
|
|
// rowset will be not given to base compaction. The unit is m byte.
|
|
DEFINE_mInt64(compaction_promotion_min_size_mbytes, "128");
|
|
|
|
// When output rowset of cumulative compaction total version count (end_version - start_version)
|
|
// exceed this config count, the rowset will be moved to base compaction
|
|
// NOTE: this config will work for unique key merge-on-write table only, to reduce version count
|
|
// related cost on delete bitmap more effectively.
|
|
DEFINE_mInt64(compaction_promotion_version_count, "1000");
|
|
|
|
// The lower bound size to do cumulative compaction. When total disk size of candidate rowsets is less than
|
|
// this size, size_based policy may not do to cumulative compaction. The unit is m byte.
|
|
DEFINE_mInt64(compaction_min_size_mbytes, "64");
|
|
|
|
// cumulative compaction policy: min and max delta file's number
|
|
DEFINE_mInt64(cumulative_compaction_min_deltas, "5");
|
|
DEFINE_mInt64(cumulative_compaction_max_deltas, "1000");
|
|
DEFINE_mInt32(cumulative_compaction_max_deltas_factor, "10");
|
|
|
|
// This config can be set to limit thread number in multiget thread pool.
|
|
DEFINE_mInt32(multi_get_max_threads, "10");
|
|
|
|
// The upper limit of "permits" held by all compaction tasks. This config can be set to limit memory consumption for compaction.
|
|
DEFINE_mInt64(total_permits_for_compaction_score, "1000000");
|
|
|
|
// sleep interval in ms after generated compaction tasks
|
|
DEFINE_mInt32(generate_compaction_tasks_interval_ms, "100");
|
|
|
|
// sleep interval in second after update replica infos
|
|
DEFINE_mInt32(update_replica_infos_interval_seconds, "60");
|
|
|
|
// Compaction task number per disk.
|
|
// Must be greater than 2, because Base compaction and Cumulative compaction have at least one thread each.
|
|
DEFINE_mInt32(compaction_task_num_per_disk, "4");
|
|
// compaction thread num for fast disk(typically .SSD), must be greater than 2.
|
|
DEFINE_mInt32(compaction_task_num_per_fast_disk, "8");
|
|
DEFINE_Validator(compaction_task_num_per_disk,
|
|
[](const int config) -> bool { return config >= 2; });
|
|
DEFINE_Validator(compaction_task_num_per_fast_disk,
|
|
[](const int config) -> bool { return config >= 2; });
|
|
DEFINE_Validator(low_priority_compaction_task_num_per_disk,
|
|
[](const int config) -> bool { return config >= 2; });
|
|
|
|
// How many rounds of cumulative compaction for each round of base compaction when compaction tasks generation.
|
|
DEFINE_mInt32(cumulative_compaction_rounds_for_each_base_compaction_round, "9");
|
|
|
|
// Not compact the invisible versions, but with some limitations:
|
|
// if not timeout, keep no more than compaction_keep_invisible_version_max_count versions;
|
|
// if timeout, keep no more than compaction_keep_invisible_version_min_count versions.
|
|
DEFINE_mInt32(compaction_keep_invisible_version_timeout_sec, "1800");
|
|
DEFINE_mInt32(compaction_keep_invisible_version_min_count, "50");
|
|
DEFINE_mInt32(compaction_keep_invisible_version_max_count, "500");
|
|
|
|
// Threshold to logging compaction trace, in seconds.
|
|
DEFINE_mInt32(base_compaction_trace_threshold, "60");
|
|
DEFINE_mInt32(cumulative_compaction_trace_threshold, "10");
|
|
DEFINE_mBool(disable_compaction_trace_log, "true");
|
|
|
|
// Interval to picking rowset to compact, in seconds
|
|
DEFINE_mInt64(pick_rowset_to_compact_interval_sec, "86400");
|
|
|
|
// Compaction priority schedule
|
|
DEFINE_mBool(enable_compaction_priority_scheduling, "true");
|
|
DEFINE_mInt32(low_priority_compaction_task_num_per_disk, "2");
|
|
DEFINE_mInt32(low_priority_compaction_score_threshold, "200");
|
|
|
|
// Thread count to do tablet meta checkpoint, -1 means use the data directories count.
|
|
DEFINE_Int32(max_meta_checkpoint_threads, "-1");
|
|
|
|
// Threshold to logging agent task trace, in seconds.
|
|
DEFINE_mInt32(agent_task_trace_threshold_sec, "2");
|
|
|
|
// This config can be set to limit thread number in tablet migration thread pool.
|
|
DEFINE_Int32(min_tablet_migration_threads, "1");
|
|
DEFINE_Int32(max_tablet_migration_threads, "1");
|
|
|
|
DEFINE_mInt32(finished_migration_tasks_size, "10000");
|
|
// If size less than this, the remaining rowsets will be force to complete
|
|
DEFINE_mInt32(migration_remaining_size_threshold_mb, "10");
|
|
// If the task runs longer than this time, the task will be terminated, in seconds.
|
|
// timeout = std::max(migration_task_timeout_secs, tablet size / 1MB/s)
|
|
DEFINE_mInt32(migration_task_timeout_secs, "300");
|
|
// timeout for try_lock migration lock
|
|
DEFINE_Int64(migration_lock_timeout_ms, "1000");
|
|
|
|
// Port to start debug webserver on
|
|
DEFINE_Int32(webserver_port, "8040");
|
|
// Https enable flag
|
|
DEFINE_Bool(enable_https, "false");
|
|
// Path of certificate
|
|
DEFINE_String(ssl_certificate_path, "");
|
|
// Path of private key
|
|
DEFINE_String(ssl_private_key_path, "");
|
|
// Whether to check authorization
|
|
DEFINE_Bool(enable_all_http_auth, "false");
|
|
// Number of webserver workers
|
|
DEFINE_Int32(webserver_num_workers, "48");
|
|
|
|
DEFINE_Bool(enable_single_replica_load, "true");
|
|
// Number of download workers for single replica load
|
|
DEFINE_Int32(single_replica_load_download_num_workers, "64");
|
|
|
|
// Used for mini Load. mini load data file will be removed after this time.
|
|
DEFINE_Int64(load_data_reserve_hours, "4");
|
|
// log error log will be removed after this time
|
|
DEFINE_mInt64(load_error_log_reserve_hours, "48");
|
|
// error log size limit, default 200MB
|
|
DEFINE_mInt64(load_error_log_limit_bytes, "209715200");
|
|
|
|
DEFINE_Int32(brpc_heavy_work_pool_threads, "-1");
|
|
DEFINE_Int32(brpc_light_work_pool_threads, "-1");
|
|
DEFINE_Int32(brpc_heavy_work_pool_max_queue_size, "-1");
|
|
DEFINE_Int32(brpc_light_work_pool_max_queue_size, "-1");
|
|
DEFINE_Int32(brpc_arrow_flight_work_pool_threads, "-1");
|
|
DEFINE_Int32(brpc_arrow_flight_work_pool_max_queue_size, "-1");
|
|
|
|
//Enable brpc builtin services, see:
|
|
//https://brpc.apache.org/docs/server/basics/#disable-built-in-services-completely
|
|
DEFINE_Bool(enable_brpc_builtin_services, "true");
|
|
|
|
// Enable brpc connection check
|
|
DEFINE_Bool(enable_brpc_connection_check, "false");
|
|
|
|
DEFINE_mInt64(brpc_connection_check_timeout_ms, "10000");
|
|
|
|
// The maximum amount of data that can be processed by a stream load
|
|
DEFINE_mInt64(streaming_load_max_mb, "102400");
|
|
// Some data formats, such as JSON, cannot be streamed.
|
|
// Therefore, it is necessary to limit the maximum number of
|
|
// such data when using stream load to prevent excessive memory consumption.
|
|
DEFINE_mInt64(streaming_load_json_max_mb, "100");
|
|
// the alive time of a TabletsChannel.
|
|
// If the channel does not receive any data till this time,
|
|
// the channel will be removed.
|
|
DEFINE_mInt32(streaming_load_rpc_max_alive_time_sec, "1200");
|
|
// the timeout of a rpc to open the tablet writer in remote BE.
|
|
// short operation time, can set a short timeout
|
|
DEFINE_Int32(tablet_writer_open_rpc_timeout_sec, "60");
|
|
// You can ignore brpc error '[E1011]The server is overcrowded' when writing data.
|
|
DEFINE_mBool(tablet_writer_ignore_eovercrowded, "true");
|
|
DEFINE_mBool(exchange_sink_ignore_eovercrowded, "true");
|
|
DEFINE_mInt32(slave_replica_writer_rpc_timeout_sec, "60");
|
|
// Whether to enable stream load record function, the default is false.
|
|
// False: disable stream load record
|
|
DEFINE_mBool(enable_stream_load_record, "false");
|
|
// batch size of stream load record reported to FE
|
|
DEFINE_mInt32(stream_load_record_batch_size, "50");
|
|
// expire time of stream load record in rocksdb.
|
|
DEFINE_Int32(stream_load_record_expire_time_secs, "28800");
|
|
// time interval to clean expired stream load records
|
|
DEFINE_mInt64(clean_stream_load_record_interval_secs, "1800");
|
|
// The buffer size to store stream table function schema info
|
|
DEFINE_Int64(stream_tvf_buffer_size, "1048576"); // 1MB
|
|
|
|
// OlapTableSink sender's send interval, should be less than the real response time of a tablet writer rpc.
|
|
// You may need to lower the speed when the sink receiver bes are too busy.
|
|
DEFINE_mInt32(olap_table_sink_send_interval_microseconds, "1000");
|
|
DEFINE_mDouble(olap_table_sink_send_interval_auto_partition_factor, "0.001");
|
|
|
|
// Fragment thread pool
|
|
DEFINE_Int32(fragment_pool_thread_num_min, "64");
|
|
DEFINE_Int32(fragment_pool_thread_num_max, "2048");
|
|
DEFINE_Int32(fragment_pool_queue_size, "4096");
|
|
|
|
// Control the number of disks on the machine. If 0, this comes from the system settings.
|
|
DEFINE_Int32(num_disks, "0");
|
|
// The maximum number of the threads per disk is also the max queue depth per disk.
|
|
DEFINE_Int32(num_threads_per_disk, "0");
|
|
// The read size is the size of the reads sent to os.
|
|
// There is a trade off of latency and throughout, trying to keep disks busy but
|
|
// not introduce seeks. The literature seems to agree that with 8 MB reads, random
|
|
// io and sequential io perform similarly.
|
|
DEFINE_Int32(read_size, "8388608"); // 8 * 1024 * 1024, Read Size (in bytes)
|
|
DEFINE_Int32(min_buffer_size, "1024"); // 1024, The minimum read buffer size (in bytes)
|
|
|
|
// for pprof
|
|
DEFINE_String(pprof_profile_dir, "${DORIS_HOME}/log");
|
|
// for jeprofile in jemalloc
|
|
DEFINE_mString(jeprofile_dir, "${DORIS_HOME}/log");
|
|
DEFINE_mBool(enable_je_purge_dirty_pages, "true");
|
|
DEFINE_mString(je_dirty_pages_mem_limit_percent, "5%");
|
|
|
|
// to forward compatibility, will be removed later
|
|
DEFINE_mBool(enable_token_check, "true");
|
|
|
|
// to open/close system metrics
|
|
DEFINE_Bool(enable_system_metrics, "true");
|
|
|
|
// Number of cores Doris will used, this will effect only when it's greater than 0.
|
|
// Otherwise, Doris will use all cores returned from "/proc/cpuinfo".
|
|
DEFINE_Int32(num_cores, "0");
|
|
|
|
// When BE start, If there is a broken disk, BE process will exit by default.
|
|
// Otherwise, we will ignore the broken disk,
|
|
DEFINE_Bool(ignore_broken_disk, "false");
|
|
|
|
// Sleep time in milliseconds between memory maintenance iterations
|
|
DEFINE_mInt32(memory_maintenance_sleep_time_ms, "100");
|
|
|
|
// After full gc, no longer full gc and minor gc during sleep.
|
|
// After minor gc, no minor gc during sleep, but full gc is possible.
|
|
DEFINE_mInt32(memory_gc_sleep_time_ms, "500");
|
|
|
|
// Sleep time in milliseconds between memtbale flush mgr refresh iterations
|
|
DEFINE_mInt64(memtable_mem_tracker_refresh_interval_ms, "5");
|
|
|
|
DEFINE_mInt64(wg_weighted_memory_ratio_refresh_interval_ms, "50");
|
|
|
|
// percent of (active memtables size / all memtables size) when reach hard limit
|
|
DEFINE_mInt32(memtable_hard_limit_active_percent, "50");
|
|
|
|
// percent of (active memtables size / all memtables size) when reach soft limit
|
|
DEFINE_mInt32(memtable_soft_limit_active_percent, "50");
|
|
|
|
// memtable insert memory tracker will multiply input block size with this ratio
|
|
DEFINE_mDouble(memtable_insert_memory_ratio, "1.4");
|
|
// max write buffer size before flush, default 200MB
|
|
DEFINE_mInt64(write_buffer_size, "209715200");
|
|
// max buffer size used in memtable for the aggregated table, default 400MB
|
|
DEFINE_mInt64(write_buffer_size_for_agg, "419430400");
|
|
// max parallel flush task per memtable writer
|
|
DEFINE_mInt32(memtable_flush_running_count_limit, "2");
|
|
|
|
DEFINE_Int32(load_process_max_memory_limit_percent, "50"); // 50%
|
|
|
|
// If the memory consumption of load jobs exceed load_process_max_memory_limit,
|
|
// all load jobs will hang there to wait for memtable flush. We should have a
|
|
// soft limit which can trigger the memtable flush for the load channel who
|
|
// consumes lagest memory size before we reach the hard limit. The soft limit
|
|
// might avoid all load jobs hang at the same time.
|
|
DEFINE_Int32(load_process_soft_mem_limit_percent, "80");
|
|
|
|
// If load memory consumption is within load_process_safe_mem_permit_percent,
|
|
// memtable memory limiter will do nothing.
|
|
DEFINE_Int32(load_process_safe_mem_permit_percent, "5");
|
|
|
|
// result buffer cancelled time (unit: second)
|
|
DEFINE_mInt32(result_buffer_cancelled_interval_time, "300");
|
|
|
|
// arrow flight result sink buffer rows size, default 4096 * 8
|
|
DEFINE_mInt32(arrow_flight_result_sink_buffer_size_rows, "32768");
|
|
// The timeout for ADBC Client to wait for data using arrow flight reader.
|
|
// If the query is very complex and no result is generated after this time, consider increasing this timeout.
|
|
DEFINE_mInt32(arrow_flight_reader_brpc_controller_timeout_ms, "300000");
|
|
|
|
// the increased frequency of priority for remaining tasks in BlockingPriorityQueue
|
|
DEFINE_mInt32(priority_queue_remaining_tasks_increased_frequency, "512");
|
|
|
|
// sync tablet_meta when modifying meta
|
|
DEFINE_mBool(sync_tablet_meta, "false");
|
|
|
|
// default thrift rpc timeout ms
|
|
DEFINE_mInt32(thrift_rpc_timeout_ms, "60000");
|
|
|
|
// txn commit rpc timeout
|
|
DEFINE_mInt32(txn_commit_rpc_timeout_ms, "60000");
|
|
|
|
// If set to true, metric calculator will run
|
|
DEFINE_Bool(enable_metric_calculator, "true");
|
|
|
|
// max consumer num in one data consumer group, for routine load
|
|
DEFINE_mInt32(max_consumer_num_per_group, "3");
|
|
|
|
// the max size of thread pool for routine load task.
|
|
// this should be larger than FE config 'max_routine_load_task_num_per_be' (default 5)
|
|
DEFINE_Int32(max_routine_load_thread_pool_size, "1024");
|
|
|
|
// max external scan cache batch count, means cache max_memory_cache_batch_count * batch_size row
|
|
// default is 20, batch_size's default value is 1024 means 20 * 1024 rows will be cached
|
|
DEFINE_mInt32(max_memory_sink_batch_count, "20");
|
|
|
|
// This configuration is used for the context gc thread schedule period
|
|
// note: unit is minute, default is 5min
|
|
DEFINE_mInt32(scan_context_gc_interval_min, "5");
|
|
|
|
// es scroll keep-alive
|
|
DEFINE_String(es_scroll_keepalive, "5m");
|
|
|
|
// HTTP connection timeout for es
|
|
DEFINE_mInt32(es_http_timeout_ms, "5000");
|
|
|
|
// the max client cache number per each host
|
|
// There are variety of client cache in BE, but currently we use the
|
|
// same cache size configuration.
|
|
// TODO(cmy): use different config to set different client cache if necessary.
|
|
DEFINE_Int32(max_client_cache_size_per_host, "10");
|
|
|
|
// Dir to save files downloaded by SmallFileMgr
|
|
DEFINE_String(small_file_dir, "${DORIS_HOME}/lib/small_file/");
|
|
// path gc
|
|
DEFINE_Bool(path_gc_check, "true");
|
|
DEFINE_mInt32(path_gc_check_interval_second, "86400");
|
|
DEFINE_mInt32(path_gc_check_step, "1000");
|
|
DEFINE_mInt32(path_gc_check_step_interval_ms, "10");
|
|
|
|
// The following 2 configs limit the max usage of disk capacity of a data dir.
|
|
// If both of these 2 threshold reached, no more data can be writen into that data dir.
|
|
// The percent of max used capacity of a data dir
|
|
DEFINE_mInt32(storage_flood_stage_usage_percent, "90"); // 90%
|
|
// The min bytes that should be left of a data dir
|
|
DEFINE_mInt64(storage_flood_stage_left_capacity_bytes, "1073741824"); // 1GB
|
|
// number of thread for flushing memtable per store
|
|
DEFINE_Int32(flush_thread_num_per_store, "6");
|
|
// number of thread for flushing memtable per store, for high priority load task
|
|
DEFINE_Int32(high_priority_flush_thread_num_per_store, "6");
|
|
|
|
DEFINE_Int32(wg_flush_thread_num_per_store, "6");
|
|
|
|
// config for tablet meta checkpoint
|
|
DEFINE_mInt32(tablet_meta_checkpoint_min_new_rowsets_num, "10");
|
|
DEFINE_mInt32(tablet_meta_checkpoint_min_interval_secs, "600");
|
|
DEFINE_Int32(generate_tablet_meta_checkpoint_tasks_interval_secs, "600");
|
|
|
|
// config for default rowset type
|
|
// Valid configs: ALPHA, BETA
|
|
DEFINE_String(default_rowset_type, "BETA");
|
|
|
|
// Maximum size of a single message body in all protocols
|
|
DEFINE_Int64(brpc_max_body_size, "3147483648");
|
|
DEFINE_Int64(brpc_socket_max_unwritten_bytes, "-1");
|
|
// TODO(zxy): expect to be true in v1.3
|
|
// Whether to embed the ProtoBuf Request serialized string together with Tuple/Block data into
|
|
// Controller Attachment and send it through http brpc when the length of the Tuple/Block data
|
|
// is greater than 1.8G. This is to avoid the error of Request length overflow (2G).
|
|
DEFINE_mBool(transfer_large_data_by_brpc, "true");
|
|
|
|
// max number of txns for every txn_partition_map in txn manager
|
|
// this is a self protection to avoid too many txns saving in manager
|
|
DEFINE_mInt64(max_runnings_transactions_per_txn_map, "2000");
|
|
|
|
// tablet_map_lock shard size, the value is 2^n, n=0,1,2,3,4
|
|
// this is a an enhancement for better performance to manage tablet
|
|
DEFINE_Int32(tablet_map_shard_size, "256");
|
|
|
|
// txn_map_lock shard size, the value is 2^n, n=0,1,2,3,4
|
|
// this is a an enhancement for better performance to manage txn
|
|
DEFINE_Int32(txn_map_shard_size, "1024");
|
|
|
|
// txn_lock shard size, the value is 2^n, n=0,1,2,3,4
|
|
// this is a an enhancement for better performance to commit and publish txn
|
|
DEFINE_Int32(txn_shard_size, "1024");
|
|
|
|
// Whether to continue to start be when load tablet from header failed.
|
|
DEFINE_Bool(ignore_load_tablet_failure, "false");
|
|
|
|
// Whether to continue to start be when load tablet from header failed.
|
|
DEFINE_mBool(ignore_rowset_stale_unconsistent_delete, "false");
|
|
|
|
// Set max cache's size of query results, the unit is M byte
|
|
DEFINE_Int32(query_cache_max_size_mb, "256");
|
|
|
|
// Cache memory is pruned when reach query_cache_max_size_mb + query_cache_elasticity_size_mb
|
|
DEFINE_Int32(query_cache_elasticity_size_mb, "128");
|
|
|
|
// Maximum number of cache partitions corresponding to a SQL
|
|
DEFINE_Int32(query_cache_max_partition_count, "1024");
|
|
|
|
// Maximum number of version of a tablet. If the version num of a tablet exceed limit,
|
|
// the load process will reject new incoming load job of this tablet.
|
|
// This is to avoid too many version num.
|
|
DEFINE_mInt32(max_tablet_version_num, "2000");
|
|
|
|
// Frontend mainly use two thrift sever type: THREAD_POOL, THREADED_SELECTOR. if fe use THREADED_SELECTOR model for thrift server,
|
|
// the thrift_server_type_of_fe should be set THREADED_SELECTOR to make be thrift client to fe constructed with TFramedTransport
|
|
DEFINE_String(thrift_server_type_of_fe, "THREAD_POOL");
|
|
|
|
// disable zone map index when page row is too few
|
|
DEFINE_mInt32(zone_map_row_num_threshold, "20");
|
|
|
|
// aws sdk log level
|
|
// Off = 0,
|
|
// Fatal = 1,
|
|
// Error = 2,
|
|
// Warn = 3,
|
|
// Info = 4,
|
|
// Debug = 5,
|
|
// Trace = 6
|
|
// Default to turn off aws sdk log, because aws sdk errors that need to be cared will be output through Doris logs
|
|
DEFINE_Int32(aws_log_level, "0");
|
|
|
|
// the buffer size when read data from remote storage like s3
|
|
DEFINE_mInt32(remote_storage_read_buffer_mb, "16");
|
|
|
|
// The minimum length when TCMalloc Hook consumes/releases MemTracker, consume size
|
|
// smaller than this value will continue to accumulate. specified as number of bytes.
|
|
// Decreasing this value will increase the frequency of consume/release.
|
|
// Increasing this value will cause MemTracker statistics to be inaccurate.
|
|
DEFINE_mInt32(mem_tracker_consume_min_size_bytes, "1048576");
|
|
|
|
// The version information of the tablet will be stored in the memory
|
|
// in an adjacency graph data structure.
|
|
// And as the new version is written and the old version is deleted,
|
|
// the data structure will begin to have empty vertex with no edge associations(orphan vertex).
|
|
// This config is used to control that when the proportion of orphan vertex is greater than the threshold,
|
|
// the adjacency graph will be rebuilt to ensure that the data structure will not expand indefinitely.
|
|
// This config usually only needs to be modified during testing.
|
|
// In most cases, it does not need to be modified.
|
|
DEFINE_mDouble(tablet_version_graph_orphan_vertex_ratio, "0.1");
|
|
|
|
// share delta writers when memtable_on_sink_node = true
|
|
DEFINE_Bool(share_delta_writers, "true");
|
|
// timeout for open load stream rpc in ms
|
|
DEFINE_Int64(open_load_stream_timeout_ms, "60000"); // 60s
|
|
// enable write background when using brpc stream
|
|
DEFINE_mBool(enable_brpc_stream_write_background, "true");
|
|
|
|
// brpc streaming max_buf_size in bytes
|
|
DEFINE_Int64(load_stream_max_buf_size, "20971520"); // 20MB
|
|
// brpc streaming messages_in_batch
|
|
DEFINE_Int32(load_stream_messages_in_batch, "128");
|
|
// brpc streaming StreamWait seconds on EAGAIN
|
|
DEFINE_Int32(load_stream_eagain_wait_seconds, "600");
|
|
// max tasks per flush token in load stream
|
|
DEFINE_Int32(load_stream_flush_token_max_tasks, "15");
|
|
// max wait flush token time in load stream
|
|
DEFINE_Int32(load_stream_max_wait_flush_token_time_ms, "600000");
|
|
|
|
// max send batch parallelism for OlapTableSink
|
|
// The value set by the user for send_batch_parallelism is not allowed to exceed max_send_batch_parallelism_per_job,
|
|
// if exceed, the value of send_batch_parallelism would be max_send_batch_parallelism_per_job
|
|
DEFINE_mInt32(max_send_batch_parallelism_per_job, "5");
|
|
DEFINE_Validator(max_send_batch_parallelism_per_job,
|
|
[](const int config) -> bool { return config >= 1; });
|
|
|
|
// number of send batch thread pool size
|
|
DEFINE_Int32(send_batch_thread_pool_thread_num, "64");
|
|
// number of send batch thread pool queue size
|
|
DEFINE_Int32(send_batch_thread_pool_queue_size, "102400");
|
|
|
|
// Limit the number of segment of a newly created rowset.
|
|
// The newly created rowset may to be compacted after loading,
|
|
// so if there are too many segment in a rowset, the compaction process
|
|
// will run out of memory.
|
|
// When doing compaction, each segment may take at least 1MB buffer.
|
|
DEFINE_mInt32(max_segment_num_per_rowset, "1000");
|
|
DEFINE_mInt32(segment_compression_threshold_kb, "256");
|
|
|
|
// The connection timeout when connecting to external table such as odbc table.
|
|
DEFINE_mInt32(external_table_connect_timeout_sec, "30");
|
|
|
|
// Time to clean up useless JDBC connection pool cache
|
|
DEFINE_mInt32(jdbc_connection_pool_cache_clear_time_sec, "28800");
|
|
|
|
// Global bitmap cache capacity for aggregation cache, size in bytes
|
|
DEFINE_Int64(delete_bitmap_agg_cache_capacity, "104857600");
|
|
// The default delete bitmap cache is set to 100MB,
|
|
// which can be insufficient and cause performance issues when the amount of user data is large.
|
|
// To mitigate the problem of an inadequate cache,
|
|
// we will take the larger of 0.5% of the total memory and 100MB as the delete bitmap cache size.
|
|
DEFINE_String(delete_bitmap_dynamic_agg_cache_limit, "0.5%");
|
|
DEFINE_mInt32(delete_bitmap_agg_cache_stale_sweep_time_sec, "1800");
|
|
|
|
// reference https://github.com/edenhill/librdkafka/blob/master/INTRODUCTION.md#broker-version-compatibility
|
|
// If the dependent kafka broker version older than 0.10.0.0,
|
|
// the value of kafka_api_version_request should be false, and the
|
|
// value set by the fallback version kafka_broker_version_fallback will be used,
|
|
// and the valid values are: 0.9.0.x, 0.8.x.y.
|
|
DEFINE_String(kafka_api_version_request, "true");
|
|
DEFINE_String(kafka_broker_version_fallback, "0.10.0");
|
|
DEFINE_String(kafka_debug, "disable");
|
|
|
|
// The number of pool siz of routine load consumer.
|
|
// If you meet the error describe in https://github.com/edenhill/librdkafka/issues/3608
|
|
// Change this size to 0 to fix it temporarily.
|
|
DEFINE_mInt32(routine_load_consumer_pool_size, "1024");
|
|
|
|
// Used in single-stream-multi-table load. When receive a batch of messages from kafka,
|
|
// if the size of batch is more than this threshold, we will request plans for all related tables.
|
|
DEFINE_Int32(multi_table_batch_plan_threshold, "200");
|
|
|
|
// Used in single-stream-multi-table load. When receiving a batch of messages from Kafka,
|
|
// if the size of the table wait for plan is more than this threshold, we will request plans for all related tables.
|
|
// The param is aimed to avoid requesting and executing too many plans at once.
|
|
// Performing small batch processing on multiple tables during the loaded process can reduce the pressure of a single RPC
|
|
// and improve the real-time processing of data.
|
|
DEFINE_Int32(multi_table_max_wait_tables, "5");
|
|
|
|
// When the timeout of a load task is less than this threshold,
|
|
// Doris treats it as a high priority task.
|
|
// high priority tasks use a separate thread pool for flush and do not block rpc by memory cleanup logic.
|
|
// this threshold is mainly used to identify routine load tasks and should not be modified if not necessary.
|
|
DEFINE_mInt32(load_task_high_priority_threshold_second, "120");
|
|
|
|
// The min timeout of load rpc (add batch, close, etc.)
|
|
// Because a load rpc may be blocked for a while.
|
|
// Increase this config may avoid rpc timeout.
|
|
DEFINE_mInt32(min_load_rpc_timeout_ms, "20000");
|
|
|
|
// use which protocol to access function service, candicate is baidu_std/h2:grpc
|
|
DEFINE_String(function_service_protocol, "h2:grpc");
|
|
|
|
// use which load balancer to select server to connect
|
|
DEFINE_String(rpc_load_balancer, "rr");
|
|
|
|
// a soft limit of string type length, the hard limit is 2GB - 4, but if too long will cause very low performance,
|
|
// so we set a soft limit, default is 1MB
|
|
DEFINE_mInt32(string_type_length_soft_limit_bytes, "1048576");
|
|
|
|
DEFINE_Validator(string_type_length_soft_limit_bytes,
|
|
[](const int config) -> bool { return config > 0 && config <= 2147483643; });
|
|
|
|
DEFINE_mInt32(jsonb_type_length_soft_limit_bytes, "1048576");
|
|
|
|
DEFINE_Validator(jsonb_type_length_soft_limit_bytes,
|
|
[](const int config) -> bool { return config > 0 && config <= 2147483643; });
|
|
|
|
// Threshold of reading a small file into memory
|
|
DEFINE_mInt32(in_memory_file_size, "1048576"); // 1MB
|
|
|
|
// ParquetReaderWrap prefetch buffer size
|
|
DEFINE_Int32(parquet_reader_max_buffer_size, "50");
|
|
// Max size of parquet page header in bytes
|
|
DEFINE_mInt32(parquet_header_max_size_mb, "1");
|
|
// Max buffer size for parquet row group
|
|
DEFINE_mInt32(parquet_rowgroup_max_buffer_mb, "128");
|
|
// Max buffer size for parquet chunk column
|
|
DEFINE_mInt32(parquet_column_max_buffer_mb, "8");
|
|
DEFINE_mDouble(max_amplified_read_ratio, "0.8");
|
|
DEFINE_mInt32(merged_oss_min_io_size, "1048576");
|
|
DEFINE_mInt32(merged_hdfs_min_io_size, "8192");
|
|
|
|
// OrcReader
|
|
DEFINE_mInt32(orc_natural_read_size_mb, "8");
|
|
DEFINE_mInt64(big_column_size_buffer, "65535");
|
|
DEFINE_mInt64(small_column_size_buffer, "100");
|
|
|
|
// When the rows number reached this limit, will check the filter rate the of bloomfilter
|
|
// if it is lower than a specific threshold, the predicate will be disabled.
|
|
DEFINE_mInt32(rf_predicate_check_row_num, "204800");
|
|
|
|
// cooldown task configs
|
|
DEFINE_Int32(cooldown_thread_num, "5");
|
|
DEFINE_mInt64(generate_cooldown_task_interval_sec, "20");
|
|
DEFINE_mInt32(remove_unused_remote_files_interval_sec, "21600"); // 6h
|
|
DEFINE_mInt32(confirm_unused_remote_files_interval_sec, "60");
|
|
DEFINE_Int32(cold_data_compaction_thread_num, "2");
|
|
DEFINE_mInt32(cold_data_compaction_interval_sec, "1800");
|
|
|
|
DEFINE_String(tmp_file_dir, "tmp");
|
|
|
|
DEFINE_Int32(s3_transfer_executor_pool_size, "2");
|
|
|
|
DEFINE_Bool(enable_time_lut, "true");
|
|
DEFINE_mBool(enable_simdjson_reader, "true");
|
|
|
|
DEFINE_mBool(enable_query_like_bloom_filter, "true");
|
|
// number of s3 scanner thread pool size
|
|
DEFINE_Int32(doris_remote_scanner_thread_pool_thread_num, "48");
|
|
// number of s3 scanner thread pool queue size
|
|
DEFINE_Int32(doris_remote_scanner_thread_pool_queue_size, "102400");
|
|
|
|
// limit the queue of pending batches which will be sent by a single nodechannel
|
|
DEFINE_mInt64(nodechannel_pending_queue_max_bytes, "67108864");
|
|
|
|
// The batch size for sending data by brpc streaming client
|
|
DEFINE_mInt64(brpc_streaming_client_batch_bytes, "262144");
|
|
|
|
// Max waiting time to wait the "plan fragment start" rpc.
|
|
// If timeout, the fragment will be cancelled.
|
|
// This parameter is usually only used when the FE loses connection,
|
|
// and the BE can automatically cancel the relevant fragment after the timeout,
|
|
// so as to avoid occupying the execution thread for a long time.
|
|
DEFINE_mInt32(max_fragment_start_wait_time_seconds, "30");
|
|
|
|
DEFINE_mInt32(fragment_mgr_cancel_worker_interval_seconds, "1");
|
|
|
|
// Node role tag for backend. Mix role is the default role, and computation role have no
|
|
// any tablet.
|
|
DEFINE_String(be_node_role, "mix");
|
|
|
|
// Hide webserver page for safety.
|
|
// Hide the be config page for webserver.
|
|
DEFINE_Bool(hide_webserver_config_page, "false");
|
|
|
|
DEFINE_Bool(enable_segcompaction, "true");
|
|
|
|
// Max number of segments allowed in a single segcompaction task.
|
|
DEFINE_Int32(segcompaction_batch_size, "10");
|
|
|
|
// Max row count allowed in a single source segment, bigger segments will be skipped.
|
|
DEFINE_Int32(segcompaction_candidate_max_rows, "1048576");
|
|
|
|
// Max file size allowed in a single source segment, bigger segments will be skipped.
|
|
DEFINE_Int64(segcompaction_candidate_max_bytes, "104857600");
|
|
|
|
// Max total row count allowed in a single segcompaction task.
|
|
DEFINE_Int32(segcompaction_task_max_rows, "1572864");
|
|
|
|
// Max total file size allowed in a single segcompaction task.
|
|
DEFINE_Int64(segcompaction_task_max_bytes, "157286400");
|
|
|
|
// Global segcompaction thread pool size.
|
|
DEFINE_mInt32(segcompaction_num_threads, "5");
|
|
|
|
// enable java udf and jdbc scannode
|
|
DEFINE_Bool(enable_java_support, "true");
|
|
|
|
// Set config randomly to check more issues in github workflow
|
|
DEFINE_Bool(enable_fuzzy_mode, "false");
|
|
|
|
DEFINE_Bool(enable_debug_points, "false");
|
|
|
|
DEFINE_Int32(pipeline_executor_size, "0");
|
|
DEFINE_Bool(enable_workload_group_for_scan, "false");
|
|
DEFINE_mInt64(workload_group_scan_task_wait_timeout_ms, "10000");
|
|
|
|
DEFINE_mBool(variant_enable_flatten_nested, "false");
|
|
DEFINE_mDouble(variant_ratio_of_defaults_as_sparse_column, "1");
|
|
DEFINE_mInt64(variant_threshold_rows_to_estimate_sparse_column, "2048");
|
|
DEFINE_mBool(variant_throw_exeception_on_invalid_json, "false");
|
|
|
|
// block file cache
|
|
DEFINE_Bool(enable_file_cache, "false");
|
|
// format: [{"path":"/path/to/file_cache","total_size":21474836480,"query_limit":10737418240}]
|
|
// format: [{"path":"/path/to/file_cache","total_size":21474836480,"query_limit":10737418240},{"path":"/path/to/file_cache2","total_size":21474836480,"query_limit":10737418240}]
|
|
DEFINE_String(file_cache_path, "");
|
|
// thread will sleep 10ms per scan file num to limit IO
|
|
DEFINE_Int64(async_file_cache_init_file_num_interval, "1000");
|
|
DEFINE_Int64(async_file_cache_init_sleep_interval_ms, "20");
|
|
DEFINE_Int64(file_cache_max_file_segment_size, "4194304"); // 4MB
|
|
// 4KB <= file_cache_max_file_segment_size <= 256MB
|
|
DEFINE_Validator(file_cache_max_file_segment_size, [](const int64_t config) -> bool {
|
|
return config >= 4096 && config <= 268435456;
|
|
});
|
|
DEFINE_Int64(file_cache_min_file_segment_size, "1048576"); // 1MB
|
|
// 4KB <= file_cache_min_file_segment_size <= 256MB
|
|
DEFINE_Validator(file_cache_min_file_segment_size, [](const int64_t config) -> bool {
|
|
return config >= 4096 && config <= 268435456 &&
|
|
config <= config::file_cache_max_file_segment_size;
|
|
});
|
|
DEFINE_Bool(clear_file_cache, "false");
|
|
DEFINE_Bool(enable_file_cache_query_limit, "false");
|
|
DEFINE_mInt32(file_cache_wait_sec_after_fail, "0"); // // zero for no waiting and retrying
|
|
DEFINE_mInt32(file_cache_max_evict_num_per_round, "5000");
|
|
|
|
DEFINE_mInt32(index_cache_entry_stay_time_after_lookup_s, "1800");
|
|
DEFINE_mInt32(inverted_index_cache_stale_sweep_time_sec, "600");
|
|
// inverted index searcher cache size
|
|
DEFINE_String(inverted_index_searcher_cache_limit, "10%");
|
|
// set `true` to enable insert searcher into cache when write inverted index data
|
|
DEFINE_Bool(enable_write_index_searcher_cache, "true");
|
|
DEFINE_Bool(enable_inverted_index_cache_check_timestamp, "true");
|
|
DEFINE_Int32(inverted_index_fd_number_limit_percent, "20"); // 20%
|
|
DEFINE_Int32(inverted_index_query_cache_shards, "256");
|
|
|
|
// inverted index match bitmap cache size
|
|
DEFINE_String(inverted_index_query_cache_limit, "10%");
|
|
|
|
// inverted index
|
|
DEFINE_mDouble(inverted_index_ram_buffer_size, "512");
|
|
// -1 indicates not working.
|
|
// Normally we should not change this, it's useful for testing.
|
|
DEFINE_mInt32(inverted_index_max_buffered_docs, "-1");
|
|
// dict path for chinese analyzer
|
|
DEFINE_String(inverted_index_dict_path, "${DORIS_HOME}/dict");
|
|
DEFINE_Int32(inverted_index_read_buffer_size, "4096");
|
|
// tree depth for bkd index
|
|
DEFINE_Int32(max_depth_in_bkd_tree, "32");
|
|
// index compaction
|
|
DEFINE_mBool(inverted_index_compaction_enable, "false");
|
|
// Only for debug, do not use in production
|
|
DEFINE_mBool(debug_inverted_index_compaction, "false");
|
|
// index by RAM directory
|
|
DEFINE_mBool(inverted_index_ram_dir_enable, "true");
|
|
// use num_broadcast_buffer blocks as buffer to do broadcast
|
|
DEFINE_Int32(num_broadcast_buffer, "32");
|
|
|
|
// max depth of expression tree allowed.
|
|
DEFINE_Int32(max_depth_of_expr_tree, "600");
|
|
|
|
// Report a tablet as bad when io errors occurs more than this value.
|
|
DEFINE_mInt64(max_tablet_io_errors, "-1");
|
|
|
|
// Report a tablet as bad when its path not found
|
|
DEFINE_Int32(tablet_path_check_interval_seconds, "-1");
|
|
DEFINE_mInt32(tablet_path_check_batch_size, "1000");
|
|
|
|
// it must be larger than or equal to 5MB
|
|
DEFINE_mInt32(s3_write_buffer_size, "5242880");
|
|
// The timeout config for S3 buffer allocation
|
|
DEFINE_mInt32(s3_writer_buffer_allocation_timeout, "300");
|
|
DEFINE_mInt64(file_cache_max_file_reader_cache_size, "1000000");
|
|
|
|
//disable shrink memory by default
|
|
DEFINE_mBool(enable_shrink_memory, "false");
|
|
DEFINE_mInt32(schema_cache_capacity, "1024");
|
|
DEFINE_mInt32(schema_cache_sweep_time_sec, "100");
|
|
|
|
// max number of segment cache, default -1 for backward compatibility fd_number*2/5
|
|
DEFINE_Int32(segment_cache_capacity, "-1");
|
|
DEFINE_Int32(segment_cache_fd_percentage, "20");
|
|
DEFINE_mInt32(estimated_mem_per_column_reader, "512");
|
|
DEFINE_Int32(segment_cache_memory_percentage, "5");
|
|
|
|
// enable feature binlog, default false
|
|
DEFINE_Bool(enable_feature_binlog, "false");
|
|
|
|
// enable set in BitmapValue
|
|
DEFINE_Bool(enable_set_in_bitmap_value, "false");
|
|
|
|
DEFINE_Int64(max_hdfs_file_handle_cache_num, "1000");
|
|
DEFINE_Int32(max_hdfs_file_handle_cache_time_sec, "3600");
|
|
DEFINE_Int64(max_external_file_meta_cache_num, "1000");
|
|
DEFINE_mInt32(common_obj_lru_cache_stale_sweep_time_sec, "900");
|
|
// Apply delete pred in cumu compaction
|
|
DEFINE_mBool(enable_delete_when_cumu_compaction, "false");
|
|
|
|
// max_write_buffer_number for rocksdb
|
|
DEFINE_Int32(rocksdb_max_write_buffer_number, "5");
|
|
|
|
DEFINE_mBool(allow_zero_date, "false");
|
|
DEFINE_Bool(allow_invalid_decimalv2_literal, "false");
|
|
DEFINE_mString(kerberos_ccache_path, "");
|
|
DEFINE_mString(kerberos_krb5_conf_path, "/etc/krb5.conf");
|
|
|
|
DEFINE_mString(get_stack_trace_tool, "libunwind");
|
|
DEFINE_mString(dwarf_location_info_mode, "FAST");
|
|
DEFINE_mBool(enable_address_sanitizers_with_stack_trace, "true");
|
|
|
|
// the ratio of _prefetch_size/_batch_size in AutoIncIDBuffer
|
|
DEFINE_mInt64(auto_inc_prefetch_size_ratio, "10");
|
|
|
|
// the ratio of _low_level_water_level_mark/_batch_size in AutoIncIDBuffer
|
|
DEFINE_mInt64(auto_inc_low_water_level_mark_size_ratio, "3");
|
|
|
|
// number of threads that fetch auto-inc ranges from FE
|
|
DEFINE_mInt64(auto_inc_fetch_thread_num, "3");
|
|
// default 4GB
|
|
DEFINE_mInt64(lookup_connection_cache_bytes_limit, "4294967296");
|
|
|
|
// level of compression when using LZ4_HC, whose defalut value is LZ4HC_CLEVEL_DEFAULT
|
|
DEFINE_mInt64(LZ4_HC_compression_level, "9");
|
|
|
|
DEFINE_mBool(enable_merge_on_write_correctness_check, "true");
|
|
// rowid conversion correctness check when compaction for mow table
|
|
DEFINE_mBool(enable_rowid_conversion_correctness_check, "false");
|
|
// missing rows correctness check when compaction for mow table
|
|
DEFINE_mBool(enable_missing_rows_correctness_check, "false");
|
|
// When the number of missing versions is more than this value, do not directly
|
|
// retry the publish and handle it through async publish.
|
|
DEFINE_mInt32(mow_publish_max_discontinuous_version_num, "20");
|
|
// When the version is not continuous for MOW table in publish phase and the gap between
|
|
// current txn's publishing version and the max version of the tablet exceeds this value,
|
|
// don't print warning log
|
|
DEFINE_mInt32(publish_version_gap_logging_threshold, "200");
|
|
|
|
// The secure path with user files, used in the `local` table function.
|
|
DEFINE_mString(user_files_secure_path, "${DORIS_HOME}");
|
|
|
|
DEFINE_Int32(fe_expire_duration_seconds, "60");
|
|
|
|
DEFINE_Int32(grace_shutdown_wait_seconds, "120");
|
|
|
|
DEFINE_Int16(bitmap_serialize_version, "1");
|
|
|
|
// group commit config
|
|
DEFINE_String(group_commit_wal_path, "");
|
|
DEFINE_Int32(group_commit_replay_wal_retry_num, "10");
|
|
DEFINE_Int32(group_commit_replay_wal_retry_interval_seconds, "5");
|
|
DEFINE_Int32(group_commit_replay_wal_retry_interval_max_seconds, "1800");
|
|
DEFINE_Int32(group_commit_relay_wal_threads, "10");
|
|
// This config can be set to limit thread number in group commit request fragment thread pool.
|
|
DEFINE_Int32(group_commit_insert_threads, "10");
|
|
DEFINE_Int32(group_commit_memory_rows_for_max_filter_ratio, "10000");
|
|
DEFINE_Bool(wait_internal_group_commit_finish, "false");
|
|
// Max size(bytes) of group commit queues, used for mem back pressure, defult 64M.
|
|
DEFINE_mInt32(group_commit_queue_mem_limit, "67108864");
|
|
// Max size(bytes) or percentage(%) of wal disk usage, used for disk space back pressure, default 10% of the disk available space.
|
|
// group_commit_wal_max_disk_limit=1024 or group_commit_wal_max_disk_limit=10% can be automatically identified.
|
|
DEFINE_String(group_commit_wal_max_disk_limit, "10%");
|
|
DEFINE_Bool(group_commit_wait_replay_wal_finish, "false");
|
|
|
|
DEFINE_mInt32(scan_thread_nice_value, "0");
|
|
DEFINE_mInt32(tablet_schema_cache_recycle_interval, "3600");
|
|
DEFINE_mInt32(tablet_schema_cache_capacity, "102400");
|
|
|
|
DEFINE_Bool(exit_on_exception, "false");
|
|
// This config controls whether the s3 file writer would flush cache asynchronously
|
|
DEFINE_Bool(enable_flush_file_cache_async, "true");
|
|
|
|
// cgroup
|
|
DEFINE_mString(doris_cgroup_cpu_path, "");
|
|
|
|
DEFINE_mBool(enable_workload_group_memory_gc, "true");
|
|
|
|
DEFINE_Bool(ignore_always_true_predicate_for_segment, "true");
|
|
|
|
// Ingest binlog work pool size, -1 is disable, 0 is hardware concurrency
|
|
DEFINE_Int32(ingest_binlog_work_pool_size, "-1");
|
|
|
|
// Download binlog rate limit, unit is KB/s, 0 means no limit
|
|
DEFINE_Int32(download_binlog_rate_limit_kbs, "0");
|
|
|
|
DEFINE_mInt32(buffered_reader_read_timeout_ms, "600000");
|
|
|
|
DEFINE_Bool(enable_snapshot_action, "false");
|
|
|
|
DEFINE_mInt32(variant_max_merged_tablet_schema_size, "2048");
|
|
|
|
DEFINE_mBool(enable_column_type_check, "true");
|
|
// 128 MB
|
|
DEFINE_mInt64(local_exchange_buffer_mem_limit, "134217728");
|
|
|
|
// Default 300s, if its value <= 0, then log is disabled
|
|
DEFINE_mInt64(enable_debug_log_timeout_secs, "0");
|
|
|
|
// Tolerance for the number of partition id 0 in rowset, default 0
|
|
DEFINE_Int32(ignore_invalid_partition_id_rowset_num, "0");
|
|
|
|
DEFINE_mInt32(report_query_statistics_interval_ms, "3000");
|
|
// 30s
|
|
DEFINE_mInt32(query_statistics_reserve_timeout_ms, "30000");
|
|
|
|
// consider two high usage disk at the same available level if they do not exceed this diff.
|
|
DEFINE_mDouble(high_disk_avail_level_diff_usages, "0.15");
|
|
|
|
// create tablet in partition random robin idx lru size, default 10000
|
|
DEFINE_Int32(partition_disk_index_lru_size, "10000");
|
|
// limit the storage space that query spill files can use
|
|
DEFINE_String(spill_storage_root_path, "");
|
|
DEFINE_String(spill_storage_limit, "20%"); // 20%
|
|
DEFINE_mInt32(spill_gc_interval_ms, "2000"); // 2s
|
|
DEFINE_mInt32(spill_gc_work_time_ms, "2000"); // 2s
|
|
DEFINE_Int32(spill_io_thread_pool_thread_num, "-1");
|
|
DEFINE_Validator(spill_io_thread_pool_thread_num, [](const int config) -> bool {
|
|
if (config == -1) {
|
|
CpuInfo::init();
|
|
spill_io_thread_pool_thread_num = std::max(48, CpuInfo::num_cores() * 2);
|
|
}
|
|
return true;
|
|
});
|
|
DEFINE_Int32(spill_io_thread_pool_queue_size, "102400");
|
|
|
|
DEFINE_mBool(check_segment_when_build_rowset_meta, "false");
|
|
|
|
DEFINE_mInt32(max_s3_client_retry, "10");
|
|
|
|
DEFINE_mInt32(s3_read_base_wait_time_ms, "100");
|
|
DEFINE_mInt32(s3_read_max_wait_time_ms, "800");
|
|
|
|
// ca_cert_file is in this path by default, Normally no modification is required
|
|
// ca cert default path is different from different OS
|
|
DEFINE_mString(ca_cert_file_paths,
|
|
"/etc/pki/tls/certs/ca-bundle.crt;/etc/ssl/certs/ca-certificates.crt;"
|
|
"/etc/ssl/ca-bundle.pem");
|
|
|
|
/** Table sink configurations(currently contains only external table types) **/
|
|
// Minimum data processed to scale writers in exchange when non partition writing
|
|
DEFINE_mInt64(table_sink_non_partition_write_scaling_data_processed_threshold,
|
|
"26214400"); // 25MB
|
|
// Minimum data processed to trigger skewed partition rebalancing in exchange when partition writing
|
|
DEFINE_mInt64(table_sink_partition_write_min_data_processed_rebalance_threshold,
|
|
"26214400"); // 25MB
|
|
// Minimum partition data processed to rebalance writers in exchange when partition writing
|
|
DEFINE_mInt64(table_sink_partition_write_min_partition_data_processed_rebalance_threshold,
|
|
"15728640"); // 15MB
|
|
// Maximum processed partition nums of per writer when partition writing
|
|
DEFINE_mInt32(table_sink_partition_write_max_partition_nums_per_writer, "128");
|
|
|
|
/** Hive sink configurations **/
|
|
DEFINE_mInt64(hive_sink_max_file_size, "1073741824"); // 1GB
|
|
|
|
/** Iceberg sink configurations **/
|
|
DEFINE_mInt64(iceberg_sink_max_file_size, "1073741824"); // 1GB
|
|
|
|
DEFINE_mInt32(thrift_client_open_num_tries, "1");
|
|
|
|
DEFINE_mBool(ignore_schema_change_check, "false");
|
|
|
|
//JVM monitoring enable. To prevent be from crashing due to jvm compatibility issues. The default setting is off.
|
|
DEFINE_Bool(enable_jvm_monitor, "false");
|
|
|
|
// Skip loading stale rowset meta when initializing `TabletMeta` from protobuf
|
|
DEFINE_mBool(skip_loading_stale_rowset_meta, "false");
|
|
|
|
DEFINE_Bool(enable_file_logger, "true");
|
|
|
|
// The minimum row group size when exporting Parquet files. default 128MB
|
|
DEFINE_Int64(min_row_group_size, "134217728");
|
|
|
|
// The time out milliseconds for remote fetch schema RPC, default 60s
|
|
DEFINE_mInt64(fetch_remote_schema_rpc_timeout_ms, "60000");
|
|
|
|
DEFINE_mInt64(compaction_memory_bytes_limit, "1073741824");
|
|
|
|
DEFINE_mInt64(compaction_batch_size, "-1");
|
|
|
|
// If set to false, the parquet reader will not use page index to filter data.
|
|
// This is only for debug purpose, in case sometimes the page index
|
|
// filter wrong data.
|
|
DEFINE_mBool(enable_parquet_page_index, "false");
|
|
|
|
DEFINE_mBool(ignore_not_found_file_in_external_table, "true");
|
|
|
|
// Tablet meta size limit after serialization, 1.5GB
|
|
DEFINE_mInt64(tablet_meta_serialize_size_limit, "1610612736");
|
|
// Protobuf supports a maximum of 2GB, so the size of the tablet meta after serialization must be less than 2GB
|
|
// 1717986918 = 2GB * 0.8
|
|
DEFINE_Validator(tablet_meta_serialize_size_limit,
|
|
[](const int64_t config) -> bool { return config < 1717986918; });
|
|
|
|
DEFINE_mInt64(pipeline_task_leakage_detect_period_secs, "60");
|
|
DEFINE_mInt32(snappy_compression_block_size, "262144");
|
|
DEFINE_mInt32(lz4_compression_block_size, "262144");
|
|
|
|
DEFINE_mBool(enable_pipeline_task_leakage_detect, "false");
|
|
DEFINE_Bool(force_regenerate_rowsetid_on_start_error, "false");
|
|
DEFINE_mBool(enable_sleep_between_delete_cumu_compaction, "false");
|
|
|
|
DEFINE_mInt32(compaction_num_per_round, "1");
|
|
|
|
// clang-format off
|
|
#ifdef BE_TEST
|
|
// test s3
|
|
DEFINE_String(test_s3_resource, "resource");
|
|
DEFINE_String(test_s3_ak, "ak");
|
|
DEFINE_String(test_s3_sk, "sk");
|
|
DEFINE_String(test_s3_endpoint, "endpoint");
|
|
DEFINE_String(test_s3_region, "region");
|
|
DEFINE_String(test_s3_bucket, "bucket");
|
|
DEFINE_String(test_s3_prefix, "prefix");
|
|
#endif
|
|
// clang-format on
|
|
|
|
std::map<std::string, Register::Field>* Register::_s_field_map = nullptr;
|
|
std::map<std::string, std::function<bool()>>* RegisterConfValidator::_s_field_validator = nullptr;
|
|
std::map<std::string, std::string>* full_conf_map = nullptr;
|
|
|
|
std::mutex custom_conf_lock;
|
|
|
|
std::mutex mutable_string_config_lock;
|
|
|
|
// trim string
|
|
std::string& trim(std::string& s) {
|
|
// rtrim
|
|
s.erase(std::find_if(s.rbegin(), s.rend(), [](unsigned char c) { return !std::isspace(c); })
|
|
.base(),
|
|
s.end());
|
|
// ltrim
|
|
s.erase(s.begin(),
|
|
std::find_if(s.begin(), s.end(), [](unsigned char c) { return !std::isspace(c); }));
|
|
return s;
|
|
}
|
|
|
|
// split string by '='
|
|
void splitkv(const std::string& s, std::string& k, std::string& v) {
|
|
const char sep = '=';
|
|
int start = 0;
|
|
int end = 0;
|
|
if ((end = s.find(sep, start)) != std::string::npos) {
|
|
k = s.substr(start, end - start);
|
|
v = s.substr(end + 1);
|
|
} else {
|
|
k = s;
|
|
v = "";
|
|
}
|
|
}
|
|
|
|
// replace env variables
|
|
bool replaceenv(std::string& s) {
|
|
std::size_t pos = 0;
|
|
std::size_t start = 0;
|
|
while ((start = s.find("${", pos)) != std::string::npos) {
|
|
std::size_t end = s.find('}', start + 2);
|
|
if (end == std::string::npos) {
|
|
return false;
|
|
}
|
|
std::string envkey = s.substr(start + 2, end - start - 2);
|
|
const char* envval = std::getenv(envkey.c_str());
|
|
if (envval == nullptr) {
|
|
return false;
|
|
}
|
|
s.erase(start, end - start + 1);
|
|
s.insert(start, envval);
|
|
pos = start + strlen(envval);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, bool& retval);
|
|
bool strtox(const std::string& valstr, int16_t& retval);
|
|
bool strtox(const std::string& valstr, int32_t& retval);
|
|
bool strtox(const std::string& valstr, int64_t& retval);
|
|
bool strtox(const std::string& valstr, double& retval);
|
|
bool strtox(const std::string& valstr, std::string& retval);
|
|
|
|
template <typename T>
|
|
bool strtox(const std::string& valstr, std::vector<T>& retval) {
|
|
std::stringstream ss(valstr);
|
|
std::string item;
|
|
T t;
|
|
while (std::getline(ss, item, ',')) {
|
|
if (!strtox(trim(item), t)) {
|
|
return false;
|
|
}
|
|
retval.push_back(t);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, bool& retval) {
|
|
if (valstr == "true") {
|
|
retval = true;
|
|
} else if (valstr == "false") {
|
|
retval = false;
|
|
} else {
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
template <typename T>
|
|
bool strtointeger(const std::string& valstr, T& retval) {
|
|
if (valstr.length() == 0) {
|
|
return false; // empty-string is only allowed for string type.
|
|
}
|
|
char* end;
|
|
errno = 0;
|
|
const char* valcstr = valstr.c_str();
|
|
int64_t ret64 = strtoll(valcstr, &end, 10);
|
|
if (errno || end != valcstr + strlen(valcstr)) {
|
|
return false; // bad parse
|
|
}
|
|
T tmp = retval;
|
|
retval = static_cast<T>(ret64);
|
|
if (retval != ret64) {
|
|
retval = tmp;
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, int16_t& retval) {
|
|
return strtointeger(valstr, retval);
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, int32_t& retval) {
|
|
return strtointeger(valstr, retval);
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, int64_t& retval) {
|
|
return strtointeger(valstr, retval);
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, double& retval) {
|
|
if (valstr.length() == 0) {
|
|
return false; // empty-string is only allowed for string type.
|
|
}
|
|
char* end = nullptr;
|
|
errno = 0;
|
|
const char* valcstr = valstr.c_str();
|
|
retval = strtod(valcstr, &end);
|
|
if (errno || end != valcstr + strlen(valcstr)) {
|
|
return false; // bad parse
|
|
}
|
|
return true;
|
|
}
|
|
|
|
bool strtox(const std::string& valstr, std::string& retval) {
|
|
retval = valstr;
|
|
return true;
|
|
}
|
|
|
|
template <typename T>
|
|
bool convert(const std::string& value, T& retval) {
|
|
std::string valstr(value);
|
|
trim(valstr);
|
|
if (!replaceenv(valstr)) {
|
|
return false;
|
|
}
|
|
return strtox(valstr, retval);
|
|
}
|
|
|
|
// load conf file
|
|
bool Properties::load(const char* conf_file, bool must_exist) {
|
|
// if conf_file is null, use the empty props
|
|
if (conf_file == nullptr) {
|
|
return true;
|
|
}
|
|
|
|
// open the conf file
|
|
std::ifstream input(conf_file);
|
|
if (!input.is_open()) {
|
|
if (must_exist) {
|
|
std::cerr << "config::load() failed to open the file:" << conf_file << std::endl;
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
// load properties
|
|
std::string line;
|
|
std::string key;
|
|
std::string value;
|
|
line.reserve(512);
|
|
while (input) {
|
|
// read one line at a time
|
|
std::getline(input, line);
|
|
|
|
// remove left and right spaces
|
|
trim(line);
|
|
|
|
// ignore comments
|
|
if (line.empty() || line[0] == '#') {
|
|
continue;
|
|
}
|
|
|
|
// read key and value
|
|
splitkv(line, key, value);
|
|
trim(key);
|
|
trim(value);
|
|
|
|
// insert into file_conf_map
|
|
file_conf_map[key] = value;
|
|
}
|
|
|
|
// close the conf file
|
|
input.close();
|
|
|
|
return true;
|
|
}
|
|
|
|
template <typename T>
|
|
bool Properties::get_or_default(const char* key, const char* defstr, T& retval, bool* is_retval_set,
|
|
std::string& rawval) const {
|
|
const auto& it = file_conf_map.find(std::string(key));
|
|
std::string valstr;
|
|
if (it == file_conf_map.end()) {
|
|
if (defstr == nullptr) {
|
|
// Not found in conf map, and no default value need to be set, just return
|
|
*is_retval_set = false;
|
|
return true;
|
|
} else {
|
|
valstr = std::string(defstr);
|
|
}
|
|
} else {
|
|
valstr = it->second;
|
|
}
|
|
rawval = valstr;
|
|
*is_retval_set = true;
|
|
return convert(valstr, retval);
|
|
}
|
|
|
|
void Properties::set(const std::string& key, const std::string& val) {
|
|
file_conf_map.emplace(key, val);
|
|
}
|
|
|
|
void Properties::set_force(const std::string& key, const std::string& val) {
|
|
file_conf_map[key] = val;
|
|
}
|
|
|
|
Status Properties::dump(const std::string& conffile) {
|
|
std::string conffile_tmp = conffile + ".tmp";
|
|
io::FileWriterPtr file_writer;
|
|
RETURN_IF_ERROR(io::global_local_filesystem()->create_file(conffile_tmp, &file_writer));
|
|
RETURN_IF_ERROR(file_writer->append("# THIS IS AN AUTO GENERATED CONFIG FILE.\n"));
|
|
RETURN_IF_ERROR(file_writer->append(
|
|
"# You can modify this file manually, and the configurations in this file\n"));
|
|
RETURN_IF_ERROR(file_writer->append("# will overwrite the configurations in be.conf\n\n"));
|
|
|
|
for (auto const& iter : file_conf_map) {
|
|
RETURN_IF_ERROR(file_writer->append(iter.first));
|
|
RETURN_IF_ERROR(file_writer->append(" = "));
|
|
RETURN_IF_ERROR(file_writer->append(iter.second));
|
|
RETURN_IF_ERROR(file_writer->append("\n"));
|
|
}
|
|
|
|
RETURN_IF_ERROR(file_writer->close());
|
|
|
|
return io::global_local_filesystem()->rename(conffile_tmp, conffile);
|
|
}
|
|
|
|
template <typename T>
|
|
std::ostream& operator<<(std::ostream& out, const std::vector<T>& v) {
|
|
size_t last = v.size() - 1;
|
|
for (size_t i = 0; i < v.size(); ++i) {
|
|
out << v[i];
|
|
if (i != last) {
|
|
out << ", ";
|
|
}
|
|
}
|
|
return out;
|
|
}
|
|
|
|
#define SET_FIELD(FIELD, TYPE, FILL_CONF_MAP, SET_TO_DEFAULT) \
|
|
if (strcmp((FIELD).type, #TYPE) == 0) { \
|
|
TYPE new_value = TYPE(); \
|
|
bool is_newval_set = false; \
|
|
std::string raw_value; \
|
|
if (!props.get_or_default((FIELD).name, ((SET_TO_DEFAULT) ? (FIELD).defval : nullptr), \
|
|
new_value, &is_newval_set, raw_value)) { \
|
|
std::cerr << "config field error: " << (FIELD).name << " = \"" << raw_value << '"' \
|
|
<< std::endl; \
|
|
return false; \
|
|
} \
|
|
if (!is_newval_set) { \
|
|
continue; \
|
|
} \
|
|
TYPE& ref_conf_value = *reinterpret_cast<TYPE*>((FIELD).storage); \
|
|
TYPE old_value = ref_conf_value; \
|
|
ref_conf_value = new_value; \
|
|
if (RegisterConfValidator::_s_field_validator != nullptr) { \
|
|
auto validator = RegisterConfValidator::_s_field_validator->find((FIELD).name); \
|
|
if (validator != RegisterConfValidator::_s_field_validator->end() && \
|
|
!(validator->second)()) { \
|
|
ref_conf_value = old_value; \
|
|
std::cerr << "validate " << (FIELD).name << "=" << new_value << " failed" \
|
|
<< std::endl; \
|
|
return false; \
|
|
} \
|
|
} \
|
|
if (FILL_CONF_MAP) { \
|
|
std::ostringstream oss; \
|
|
oss << ref_conf_value; \
|
|
(*full_conf_map)[(FIELD).name] = oss.str(); \
|
|
} \
|
|
continue; \
|
|
}
|
|
|
|
// init conf fields
|
|
bool init(const char* conf_file, bool fill_conf_map, bool must_exist, bool set_to_default) {
|
|
Properties props;
|
|
// load properties file
|
|
if (!props.load(conf_file, must_exist)) {
|
|
return false;
|
|
}
|
|
// fill full_conf_map ?
|
|
if (fill_conf_map && full_conf_map == nullptr) {
|
|
full_conf_map = new std::map<std::string, std::string>();
|
|
}
|
|
|
|
// set conf fields
|
|
for (const auto& it : *Register::_s_field_map) {
|
|
SET_FIELD(it.second, bool, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, int16_t, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, int32_t, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, int64_t, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, double, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::string, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::vector<bool>, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::vector<int16_t>, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::vector<int32_t>, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::vector<int64_t>, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::vector<double>, fill_conf_map, set_to_default);
|
|
SET_FIELD(it.second, std::vector<std::string>, fill_conf_map, set_to_default);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
#define UPDATE_FIELD(FIELD, VALUE, TYPE, PERSIST) \
|
|
if (strcmp((FIELD).type, #TYPE) == 0) { \
|
|
TYPE new_value; \
|
|
if (!convert((VALUE), new_value)) { \
|
|
return Status::Error<ErrorCode::INVALID_ARGUMENT, false>("convert '{}' as {} failed", \
|
|
VALUE, #TYPE); \
|
|
} \
|
|
TYPE& ref_conf_value = *reinterpret_cast<TYPE*>((FIELD).storage); \
|
|
TYPE old_value = ref_conf_value; \
|
|
if (RegisterConfValidator::_s_field_validator != nullptr) { \
|
|
auto validator = RegisterConfValidator::_s_field_validator->find((FIELD).name); \
|
|
if (validator != RegisterConfValidator::_s_field_validator->end() && \
|
|
!(validator->second)()) { \
|
|
ref_conf_value = old_value; \
|
|
return Status::Error<ErrorCode::INVALID_ARGUMENT, false>("validate {}={} failed", \
|
|
(FIELD).name, new_value); \
|
|
} \
|
|
} \
|
|
ref_conf_value = new_value; \
|
|
if (full_conf_map != nullptr) { \
|
|
std::ostringstream oss; \
|
|
oss << new_value; \
|
|
(*full_conf_map)[(FIELD).name] = oss.str(); \
|
|
} \
|
|
if (PERSIST) { \
|
|
RETURN_IF_ERROR(persist_config(std::string((FIELD).name), VALUE)); \
|
|
} \
|
|
update_config(std::string((FIELD).name), VALUE); \
|
|
return Status::OK(); \
|
|
}
|
|
|
|
// write config to be_custom.conf
|
|
// the caller need to make sure that the given config is valid
|
|
Status persist_config(const std::string& field, const std::string& value) {
|
|
// lock to make sure only one thread can modify the be_custom.conf
|
|
std::lock_guard<std::mutex> l(custom_conf_lock);
|
|
|
|
static const std::string conffile = config::custom_config_dir + "/be_custom.conf";
|
|
|
|
Properties tmp_props;
|
|
if (!tmp_props.load(conffile.c_str(), false)) {
|
|
LOG(WARNING) << "failed to load " << conffile;
|
|
return Status::InternalError("failed to load conf file: {}", conffile);
|
|
}
|
|
|
|
tmp_props.set_force(field, value);
|
|
return tmp_props.dump(conffile);
|
|
}
|
|
|
|
Status set_config(const std::string& field, const std::string& value, bool need_persist,
|
|
bool force) {
|
|
auto it = Register::_s_field_map->find(field);
|
|
if (it == Register::_s_field_map->end()) {
|
|
return Status::Error<ErrorCode::NOT_FOUND, false>("'{}' is not found", field);
|
|
}
|
|
|
|
if (!force && !it->second.valmutable) {
|
|
return Status::Error<ErrorCode::NOT_IMPLEMENTED_ERROR, false>(
|
|
"'{}' is not support to modify", field);
|
|
}
|
|
|
|
UPDATE_FIELD(it->second, value, bool, need_persist);
|
|
UPDATE_FIELD(it->second, value, int16_t, need_persist);
|
|
UPDATE_FIELD(it->second, value, int32_t, need_persist);
|
|
UPDATE_FIELD(it->second, value, int64_t, need_persist);
|
|
UPDATE_FIELD(it->second, value, double, need_persist);
|
|
{
|
|
// add lock to ensure thread safe
|
|
std::lock_guard<std::mutex> lock(mutable_string_config_lock);
|
|
UPDATE_FIELD(it->second, value, std::string, need_persist);
|
|
}
|
|
|
|
// The other types are not thread safe to change dynamically.
|
|
return Status::Error<ErrorCode::NOT_IMPLEMENTED_ERROR, false>(
|
|
"'{}' is type of '{}' which is not support to modify", field, it->second.type);
|
|
}
|
|
|
|
void update_config(const std::string& field, const std::string& value) {
|
|
if ("sys_log_level" == field) {
|
|
// update log level
|
|
update_logging(field, value);
|
|
}
|
|
}
|
|
|
|
Status set_fuzzy_configs() {
|
|
std::unordered_map<std::string, std::string> fuzzy_field_and_value;
|
|
std::shared_ptr<std::mt19937_64> generator(new std::mt19937_64());
|
|
generator->seed(std::random_device()());
|
|
std::uniform_int_distribution<int64_t> distribution(0, 100);
|
|
|
|
// if have set enable_fuzzy_mode=true in be.conf, will fuzzy those field and values
|
|
fuzzy_field_and_value["disable_storage_page_cache"] =
|
|
((distribution(*generator) % 2) == 0) ? "true" : "false";
|
|
fuzzy_field_and_value["enable_system_metrics"] =
|
|
((distribution(*generator) % 2) == 0) ? "true" : "false";
|
|
fuzzy_field_and_value["enable_set_in_bitmap_value"] =
|
|
((distribution(*generator) % 2) == 0) ? "true" : "false";
|
|
fuzzy_field_and_value["enable_shrink_memory"] =
|
|
((distribution(*generator) % 2) == 0) ? "true" : "false";
|
|
|
|
fmt::memory_buffer buf;
|
|
for (auto& it : fuzzy_field_and_value) {
|
|
const auto& field = it.first;
|
|
const auto& value = it.second;
|
|
RETURN_IF_ERROR(set_config(field, value, false, true));
|
|
fmt::format_to(buf, "{}={}, ", field, value);
|
|
}
|
|
LOG(INFO) << fmt::format("FUZZY MODE IN BE: those variables have been changed: ({}).",
|
|
fmt::to_string(buf));
|
|
return Status::OK();
|
|
}
|
|
|
|
std::mutex* get_mutable_string_config_lock() {
|
|
return &mutable_string_config_lock;
|
|
}
|
|
|
|
std::vector<std::vector<std::string>> get_config_info() {
|
|
std::vector<std::vector<std::string>> configs;
|
|
std::lock_guard<std::mutex> lock(mutable_string_config_lock);
|
|
for (const auto& it : *full_conf_map) {
|
|
auto field_it = Register::_s_field_map->find(it.first);
|
|
if (field_it == Register::_s_field_map->end()) {
|
|
continue;
|
|
}
|
|
|
|
std::vector<std::string> _config;
|
|
_config.push_back(it.first);
|
|
|
|
std::string config_val = it.second;
|
|
// For compatibility, this PR #32933 change the log dir's config logic,
|
|
// and deprecate the `sys_log_dir` config.
|
|
if (it.first == "sys_log_dir" && config_val == "") {
|
|
config_val = fmt::format("{}/log", std::getenv("DORIS_HOME"));
|
|
}
|
|
|
|
_config.emplace_back(field_it->second.type);
|
|
if (0 == strcmp(field_it->second.type, "bool")) {
|
|
_config.emplace_back(config_val == "1" ? "true" : "false");
|
|
} else {
|
|
_config.push_back(config_val);
|
|
}
|
|
_config.emplace_back(field_it->second.valmutable ? "true" : "false");
|
|
|
|
configs.push_back(_config);
|
|
}
|
|
return configs;
|
|
}
|
|
|
|
} // namespace doris::config
|