| // Licensed to the Apache Software Foundation (ASF) under one |
| // or more contributor license agreements. See the NOTICE file |
| // distributed with this work for additional information |
| // regarding copyright ownership. The ASF licenses this file |
| // to you under the Apache License, Version 2.0 (the |
| // "License"); you may not use this file except in compliance |
| // with the License. You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, |
| // software distributed under the License is distributed on an |
| // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| // KIND, either express or implied. See the License for the |
| // specific language governing permissions and limitations |
| // under the License. |
| |
| #include <fmt/core.h> |
| #include <gflags/gflags.h> |
| #include <stdint.h> |
| |
| #include <algorithm> |
| #include <cctype> |
| // IWYU pragma: no_include <bthread/errno.h> |
| #include <lz4/lz4hc.h> |
| |
| #include <cerrno> // IWYU pragma: keep |
| #include <cstdlib> |
| #include <cstring> |
| #include <fstream> // IWYU pragma: keep |
| #include <functional> |
| #include <iostream> |
| #include <map> |
| #include <memory> |
| #include <mutex> |
| #include <random> |
| #include <string> |
| #include <utility> |
| #include <vector> |
| |
| #include "cloud/config.h" |
| #include "common/config.h" |
| #include "common/logging.h" |
| #include "common/status.h" |
| #include "io/fs/file_writer.h" |
| #include "io/fs/local_file_system.h" |
| #include "util/cpu_info.h" |
| |
| namespace doris::config { |
| |
| // Dir of custom config file |
| DEFINE_String(custom_config_dir, "${DORIS_HOME}/conf"); |
| |
| // Dir of jdbc drivers |
| DEFINE_String(jdbc_drivers_dir, "${DORIS_HOME}/plugins/jdbc_drivers"); |
| |
| // cluster id |
| DEFINE_Int32(cluster_id, "-1"); |
| // port on which BackendService is exported |
| DEFINE_Int32(be_port, "9060"); |
| |
| // port for brpc |
| DEFINE_Int32(brpc_port, "8060"); |
| |
| DEFINE_Int32(arrow_flight_sql_port, "8050"); |
| |
| // If the external client cannot directly access priority_networks, set public_host to be accessible |
| // to external client. |
| // There are usually two usage scenarios: |
| // 1. in production environment, it is often inconvenient to expose Doris BE nodes to the external network. |
| // However, a reverse proxy (such as Nginx) can be added to all Doris BE nodes, and the external client will be |
| // randomly routed to a Doris BE node when connecting to Nginx. set public_host to the host of Nginx. |
| // 2. if priority_networks is an internal network IP, and BE node has its own independent external IP, |
| // but Doris currently does not support modifying priority_networks, setting public_host to the real external IP. |
| DEFINE_mString(public_host, ""); |
| |
| // If the BE node is connected to the external network through a reverse proxy like Nginx |
| // and need to use Arrow Flight SQL, should add a server in Nginx to reverse proxy |
| // `Nginx:arrow_flight_sql_proxy_port` to `BE_priority_networks:arrow_flight_sql_port`. For example: |
| // upstream arrowflight { |
| // server 10.16.10.8:8069; |
| // server 10.16.10.8:8068; |
| //} |
| // server { |
| // listen 8167 http2; |
| // listen [::]:8167 http2; |
| // server_name doris.arrowflight.com; |
| // } |
| DEFINE_Int32(arrow_flight_sql_proxy_port, "-1"); |
| |
| // the number of bthreads for brpc, the default value is set to -1, |
| // which means the number of bthreads is #cpu-cores |
| DEFINE_Int32(brpc_num_threads, "256"); |
| // the time of brpc server keep idle connection, setting this value too small may cause rpc between backends to fail, |
| // the default value is set to -1, which means never close idle connection. |
| DEFINE_Int32(brpc_idle_timeout_sec, "-1"); |
| |
| // Declare a selection strategy for those servers have many ips. |
| // Note that there should at most one ip match this list. |
| // this is a list in semicolon-delimited format, in CIDR notation, e.g. 10.10.10.0/24 |
| // If no ip match this rule, will choose one randomly. |
| DEFINE_String(priority_networks, ""); |
| |
| // memory mode |
| // performance or compact |
| DEFINE_String(memory_mode, "moderate"); |
| |
| DEFINE_mBool(enable_use_cgroup_memory_info, "true"); |
| |
| // process memory limit specified as number of bytes |
| // ('<int>[bB]?'), megabytes ('<float>[mM]'), gigabytes ('<float>[gG]'), |
| // or percentage of the physical memory ('<int>%'). |
| // defaults to bytes if no unit is given" |
| // must larger than 0. and if larger than physical memory size, |
| // it will be set to physical memory size. |
| DEFINE_String(mem_limit, "90%"); |
| |
| // Soft memory limit as a fraction of hard memory limit. |
| DEFINE_Double(soft_mem_limit_frac, "0.9"); |
| |
| // Cache capacity reduce mem limit as a fraction of soft mem limit. |
| DEFINE_mDouble(cache_capacity_reduce_mem_limit_frac, "0.7"); |
| |
| // Schema change memory limit as a fraction of soft memory limit. |
| DEFINE_Double(schema_change_mem_limit_frac, "0.6"); |
| |
| // Many modern allocators (for example, tcmalloc) do not do a mremap for |
| // realloc, even in case of large enough chunks of memory. Although this allows |
| // you to increase performance and reduce memory consumption during realloc. |
| // To fix this, we do mremap manually if the chunk of memory is large enough. |
| // |
| // The threshold (128 MB, 128 * (1ULL << 20)) is chosen quite large, since changing the address |
| // space is very slow, especially in the case of a large number of threads. We |
| // expect that the set of operations mmap/something to do/mremap can only be |
| // performed about 1000 times per second. |
| // |
| // P.S. This is also required, because tcmalloc can not allocate a chunk of |
| // memory greater than 16 GB. |
| DEFINE_mInt64(mmap_threshold, "134217728"); // bytes |
| |
| // When hash table capacity is greater than 2^double_grow_degree(default 2G), grow when 75% of the capacity is satisfied. |
| // Increase can reduce the number of hash table resize, but may waste more memory. |
| DEFINE_mInt32(hash_table_double_grow_degree, "31"); |
| |
| DEFINE_mInt32(max_fill_rate, "2"); |
| |
| DEFINE_mInt32(double_resize_threshold, "23"); |
| |
| // The maximum low water mark of the system `/proc/meminfo/MemAvailable`, Unit byte, default -1. |
| // if it is -1, then low water mark = min(MemTotal - MemLimit, MemTotal * 5%), which is 3.2G on a 64G machine. |
| // Turn up max. more memory buffers will be reserved for Memory GC. |
| // Turn down max. will use as much memory as possible. |
| // note that: `max_` prefix should be removed, but keep it for compatibility. |
| DEFINE_Int64(max_sys_mem_available_low_water_mark_bytes, "-1"); |
| |
| DEFINE_Int64(memtable_limiter_reserved_memory_bytes, "838860800"); |
| |
| // The size of the memory that gc wants to release each time, as a percentage of the mem limit. |
| DEFINE_mString(process_minor_gc_size, "5%"); |
| DEFINE_mString(process_full_gc_size, "10%"); |
| |
| // gc will release cache, cancel task, and task will wait for gc to release memory, |
| // default gc strategy is conservative, if you want to exclude the interference of gc, let it be true |
| DEFINE_mBool(disable_memory_gc, "false"); |
| |
| // for the query being canceled, |
| // if (current time - cancel start time) < revoke_memory_max_tolerance_ms, the query memory is counted in `freed_memory`, |
| // and the query memory is expected to be released soon. |
| // if > revoke_memory_max_tolerance_ms, the query memory will not be counted in `freed_memory`, |
| // and the query may be blocked during the cancel process. skip this query and continue to cancel other queries. |
| DEFINE_mInt64(revoke_memory_max_tolerance_ms, "3000"); |
| |
| DEFINE_mBool(enable_stacktrace, "true"); |
| |
| DEFINE_mInt64(stacktrace_in_alloc_large_memory_bytes, "2147483647"); // 2GB -1 |
| |
| DEFINE_mInt64(crash_in_alloc_large_memory_bytes, "-1"); |
| |
| // The actual meaning of this parameter is `debug_memory`. |
| // 1. crash in memory tracker inaccurate, if memory tracker value is inaccurate, BE will crash. |
| // usually used in test environments, default value is false. |
| // 2. print more memory logs. |
| DEFINE_mBool(crash_in_memory_tracker_inaccurate, "false"); |
| |
| // default is true. if any memory tracking in Orphan mem tracker will report error. |
| // !! not modify the default value of this conf!! otherwise memory errors cannot be detected in time. |
| // allocator free memory not need to check, because when the thread memory tracker label is Orphan, |
| // use the tracker saved in Allocator. |
| DEFINE_mBool(enable_memory_orphan_check, "true"); |
| |
| // The maximum time a thread waits for full GC. Currently only query will wait for full gc. |
| DEFINE_mInt32(thread_wait_gc_max_milliseconds, "1000"); |
| |
| DEFINE_mInt64(pre_serialize_keys_limit_bytes, "16777216"); |
| |
| // the port heartbeat service used |
| DEFINE_Int32(heartbeat_service_port, "9050"); |
| // the count of heart beat service |
| DEFINE_Int32(heartbeat_service_thread_count, "1"); |
| // the count of thread to create table |
| DEFINE_Int32(create_tablet_worker_count, "3"); |
| // the count of thread to drop table |
| DEFINE_Int32(drop_tablet_worker_count, "3"); |
| // the count of thread to batch load |
| DEFINE_Int32(push_worker_count_normal_priority, "3"); |
| // the count of thread to high priority batch load |
| DEFINE_Int32(push_worker_count_high_priority, "3"); |
| // the count of thread to publish version |
| DEFINE_Int32(publish_version_worker_count, "8"); |
| // the count of tablet thread to publish version |
| DEFINE_Int32(tablet_publish_txn_max_thread, "32"); |
| // the timeout of EnginPublishVersionTask |
| DEFINE_Int32(publish_version_task_timeout_s, "8"); |
| // the count of thread to calc delete bitmap |
| DEFINE_Int32(calc_delete_bitmap_max_thread, "32"); |
| // the count of thread to calc delete bitmap worker, only used for cloud |
| DEFINE_Int32(calc_delete_bitmap_worker_count, "8"); |
| // the count of thread to calc tablet delete bitmap task, only used for cloud |
| DEFINE_Int32(calc_tablet_delete_bitmap_task_max_thread, "32"); |
| // the count of thread to clear transaction task |
| DEFINE_Int32(clear_transaction_task_worker_count, "1"); |
| // the count of thread to delete |
| DEFINE_Int32(delete_worker_count, "3"); |
| // the count of thread to alter table |
| DEFINE_Int32(alter_tablet_worker_count, "3"); |
| // the count of thread to alter index |
| DEFINE_Int32(alter_index_worker_count, "3"); |
| // the count of thread to clone |
| DEFINE_Int32(clone_worker_count, "3"); |
| // the count of thread to clone |
| DEFINE_Int32(storage_medium_migrate_count, "1"); |
| // the count of thread to check consistency |
| DEFINE_Int32(check_consistency_worker_count, "1"); |
| // the count of thread to upload |
| DEFINE_Int32(upload_worker_count, "1"); |
| // the count of thread to download |
| DEFINE_Int32(download_worker_count, "1"); |
| DEFINE_Int32(num_query_ctx_map_partitions, "128"); |
| // the count of thread to make snapshot |
| DEFINE_Int32(make_snapshot_worker_count, "5"); |
| // the count of thread to release snapshot |
| DEFINE_Int32(release_snapshot_worker_count, "5"); |
| // report random wait a little time to avoid FE receiving multiple be reports at the same time. |
| // do not set it to false for production environment |
| DEFINE_mBool(report_random_wait, "true"); |
| // the interval time(seconds) for agent report tasks signature to FE |
| DEFINE_mInt32(report_task_interval_seconds, "10"); |
| // the interval time(seconds) for agent report disk state to FE |
| DEFINE_mInt32(report_disk_state_interval_seconds, "60"); |
| // the interval time(seconds) for agent report olap table to FE |
| DEFINE_mInt32(report_tablet_interval_seconds, "60"); |
| // the max download speed(KB/s) |
| DEFINE_mInt32(max_download_speed_kbps, "50000"); |
| // download low speed limit(KB/s) |
| DEFINE_mInt32(download_low_speed_limit_kbps, "50"); |
| // download low speed time(seconds) |
| DEFINE_mInt32(download_low_speed_time, "300"); |
| // whether to download small files in batch |
| DEFINE_mBool(enable_batch_download, "true"); |
| // whether to check md5sum when download |
| DEFINE_mBool(enable_download_md5sum_check, "false"); |
| // download binlog meta timeout, default 30s |
| DEFINE_mInt32(download_binlog_meta_timeout_ms, "30000"); |
| |
| DEFINE_String(sys_log_dir, ""); |
| DEFINE_String(user_function_dir, "${DORIS_HOME}/lib/udf"); |
| // INFO, WARNING, ERROR, FATAL |
| DEFINE_mString(sys_log_level, "INFO"); |
| // TIME-DAY, TIME-HOUR, SIZE-MB-nnn |
| DEFINE_String(sys_log_roll_mode, "SIZE-MB-1024"); |
| // log roll num |
| DEFINE_Int32(sys_log_roll_num, "10"); |
| // verbose log |
| DEFINE_Strings(sys_log_verbose_modules, ""); |
| // verbose log level |
| DEFINE_Int32(sys_log_verbose_level, "10"); |
| // verbose log FLAGS_v |
| DEFINE_Int32(sys_log_verbose_flags_v, "-1"); |
| // log buffer level |
| DEFINE_String(log_buffer_level, ""); |
| // log enable custom date time format |
| DEFINE_Bool(sys_log_enable_custom_date_time_format, "false"); |
| // log custom date time format (https://en.cppreference.com/w/cpp/io/manip/put_time) |
| DEFINE_String(sys_log_custom_date_time_format, "%Y-%m-%d %H:%M:%S"); |
| // log custom date time milliseconds format (fmt::format) |
| DEFINE_String(sys_log_custom_date_time_ms_format, ",{:03d}"); |
| |
| // number of threads available to serve backend execution requests |
| DEFINE_Int32(be_service_threads, "64"); |
| |
| // The pipeline task has a high concurrency, therefore reducing its report frequency |
| DEFINE_mInt32(pipeline_status_report_interval, "10"); |
| DEFINE_mInt32(pipeline_task_exec_time_slice, "100"); |
| // number of scanner thread pool size for olap table |
| // and the min thread num of remote scanner thread pool |
| DEFINE_Int32(doris_scanner_thread_pool_thread_num, "-1"); |
| DEFINE_Validator(doris_scanner_thread_pool_thread_num, [](const int config) -> bool { |
| if (config == -1) { |
| CpuInfo::init(); |
| doris_scanner_thread_pool_thread_num = std::max(48, CpuInfo::num_cores() * 2); |
| } |
| return true; |
| }); |
| DEFINE_Int32(doris_scanner_min_thread_pool_thread_num, "8"); |
| DEFINE_Int32(remote_split_source_batch_size, "1000"); |
| DEFINE_Int32(doris_max_remote_scanner_thread_pool_thread_num, "-1"); |
| // number of olap scanner thread pool queue size |
| DEFINE_Int32(doris_scanner_thread_pool_queue_size, "102400"); |
| // default thrift client connect timeout(in seconds) |
| DEFINE_mInt32(thrift_connect_timeout_seconds, "3"); |
| DEFINE_mInt32(fetch_rpc_timeout_seconds, "30"); |
| |
| // default thrift client retry interval (in milliseconds) |
| DEFINE_mInt64(thrift_client_retry_interval_ms, "1000"); |
| // max message size of thrift request |
| // default: 100 * 1024 * 1024 |
| DEFINE_mInt64(thrift_max_message_size, "104857600"); |
| // max bytes number for single scan range, used in segmentv2 |
| DEFINE_mInt32(doris_scan_range_max_mb, "1024"); |
| // single read execute fragment row number |
| DEFINE_mInt32(doris_scanner_row_num, "16384"); |
| // single read execute fragment row bytes |
| DEFINE_mInt32(doris_scanner_row_bytes, "10485760"); |
| // single read execute fragment max run time millseconds |
| DEFINE_mInt32(doris_scanner_max_run_time_ms, "1000"); |
| // (Advanced) Maximum size of per-query receive-side buffer |
| DEFINE_mInt32(exchg_node_buffer_size_bytes, "20485760"); |
| DEFINE_mInt32(exchg_buffer_queue_capacity_factor, "64"); |
| |
| // memory_limitation_per_thread_for_schema_change_bytes unit bytes |
| DEFINE_mInt64(memory_limitation_per_thread_for_schema_change_bytes, "2147483648"); |
| |
| DEFINE_mInt32(cache_prune_interval_sec, "10"); |
| DEFINE_mInt32(cache_periodic_prune_stale_sweep_sec, "60"); |
| // the clean interval of tablet lookup cache |
| DEFINE_mInt32(tablet_lookup_cache_stale_sweep_time_sec, "30"); |
| DEFINE_mInt32(point_query_row_cache_stale_sweep_time_sec, "300"); |
| DEFINE_mInt32(disk_stat_monitor_interval, "5"); |
| DEFINE_mInt32(unused_rowset_monitor_interval, "30"); |
| DEFINE_mInt32(quering_rowsets_evict_interval, "30"); |
| DEFINE_String(storage_root_path, "${DORIS_HOME}/storage"); |
| DEFINE_mString(broken_storage_path, ""); |
| |
| // Config is used to check incompatible old format hdr_ format |
| // whether doris uses strict way. When config is true, process will log fatal |
| // and exit. When config is false, process will only log warning. |
| DEFINE_Bool(storage_strict_check_incompatible_old_format, "true"); |
| |
| // BE process will exit if the percentage of error disk reach this value. |
| DEFINE_mInt32(max_percentage_of_error_disk, "100"); |
| DEFINE_mInt32(default_num_rows_per_column_file_block, "1024"); |
| // pending data policy |
| DEFINE_mInt32(pending_data_expire_time_sec, "1800"); |
| // inc_rowset snapshot rs sweep time interval |
| DEFINE_mInt32(tablet_rowset_stale_sweep_time_sec, "600"); |
| // tablet stale rowset sweep by threshold size |
| DEFINE_Bool(tablet_rowset_stale_sweep_by_size, "false"); |
| DEFINE_mInt32(tablet_rowset_stale_sweep_threshold_size, "100"); |
| // garbage sweep policy |
| DEFINE_Int32(max_garbage_sweep_interval, "3600"); |
| DEFINE_Int32(min_garbage_sweep_interval, "180"); |
| DEFINE_mInt32(garbage_sweep_batch_size, "100"); |
| DEFINE_mInt32(snapshot_expire_time_sec, "172800"); |
| // It is only a recommended value. When the disk space is insufficient, |
| // the file storage period under trash dose not have to comply with this parameter. |
| DEFINE_mInt32(trash_file_expire_time_sec, "0"); |
| // minimum file descriptor number |
| // modify them upon necessity |
| DEFINE_Int32(min_file_descriptor_number, "60000"); |
| DEFINE_mBool(disable_segment_cache, "false"); |
| DEFINE_String(row_cache_mem_limit, "20%"); |
| |
| // Cache for storage page size |
| DEFINE_String(storage_page_cache_limit, "20%"); |
| // Shard size for page cache, the value must be power of two. |
| // It's recommended to set it to a value close to the number of BE cores in order to reduce lock contentions. |
| DEFINE_Int32(storage_page_cache_shard_size, "256"); |
| // Percentage for index page cache |
| // all storage page cache will be divided into data_page_cache and index_page_cache |
| DEFINE_Int32(index_page_cache_percentage, "10"); |
| // whether to disable page cache feature in storage |
| DEFINE_mBool(disable_storage_page_cache, "false"); |
| // whether to disable row cache feature in storage |
| DEFINE_mBool(disable_storage_row_cache, "true"); |
| // whether to disable pk page cache feature in storage |
| DEFINE_Bool(disable_pk_storage_page_cache, "false"); |
| |
| // Cache for mow primary key storage page size |
| DEFINE_String(pk_storage_page_cache_limit, "10%"); |
| // data page size for primary key index |
| DEFINE_Int32(primary_key_data_page_size, "32768"); |
| |
| DEFINE_mInt32(data_page_cache_stale_sweep_time_sec, "300"); |
| DEFINE_mInt32(index_page_cache_stale_sweep_time_sec, "600"); |
| DEFINE_mInt32(pk_index_page_cache_stale_sweep_time_sec, "600"); |
| |
| DEFINE_Bool(enable_low_cardinality_optimize, "true"); |
| DEFINE_Bool(enable_low_cardinality_cache_code, "true"); |
| |
| // be policy |
| // whether check compaction checksum |
| DEFINE_mBool(enable_compaction_checksum, "false"); |
| // whether disable automatic compaction task |
| DEFINE_mBool(disable_auto_compaction, "false"); |
| // whether enable vertical compaction |
| DEFINE_mBool(enable_vertical_compaction, "true"); |
| // whether enable ordered data compaction |
| DEFINE_mBool(enable_ordered_data_compaction, "true"); |
| // In vertical compaction, column number for every group |
| DEFINE_mInt32(vertical_compaction_num_columns_per_group, "5"); |
| // In vertical compaction, max memory usage for row_source_buffer |
| DEFINE_Int32(vertical_compaction_max_row_source_memory_mb, "1024"); |
| // In vertical compaction, max dest segment file size |
| DEFINE_mInt64(vertical_compaction_max_segment_size, "1073741824"); |
| |
| // If enabled, segments will be flushed column by column |
| DEFINE_mBool(enable_vertical_segment_writer, "true"); |
| |
| // In ordered data compaction, min segment size for input rowset |
| DEFINE_mInt32(ordered_data_compaction_min_segment_size, "10485760"); |
| |
| // This config can be set to limit thread number in compaction thread pool. |
| DEFINE_mInt32(max_base_compaction_threads, "4"); |
| DEFINE_mInt32(max_cumu_compaction_threads, "-1"); |
| DEFINE_mInt32(max_single_replica_compaction_threads, "-1"); |
| |
| DEFINE_Bool(enable_base_compaction_idle_sched, "true"); |
| DEFINE_mInt64(base_compaction_min_rowset_num, "5"); |
| DEFINE_mInt64(base_compaction_max_compaction_score, "20"); |
| DEFINE_mDouble(base_compaction_min_data_ratio, "0.3"); |
| DEFINE_mInt64(base_compaction_dup_key_max_file_size_mbytes, "1024"); |
| |
| DEFINE_Bool(enable_skip_tablet_compaction, "true"); |
| DEFINE_mInt32(skip_tablet_compaction_second, "10"); |
| |
| // output rowset of cumulative compaction total disk size exceed this config size, |
| // this rowset will be given to base compaction, unit is m byte. |
| DEFINE_mInt64(compaction_promotion_size_mbytes, "1024"); |
| |
| // output rowset of cumulative compaction total disk size exceed this config ratio of |
| // base rowset's total disk size, this rowset will be given to base compaction. The value must be between |
| // 0 and 1. |
| DEFINE_mDouble(compaction_promotion_ratio, "0.05"); |
| |
| // the smallest size of rowset promotion. When the rowset is less than this config, this |
| // rowset will be not given to base compaction. The unit is m byte. |
| DEFINE_mInt64(compaction_promotion_min_size_mbytes, "128"); |
| |
| // When output rowset of cumulative compaction total version count (end_version - start_version) |
| // exceed this config count, the rowset will be moved to base compaction |
| // NOTE: this config will work for unique key merge-on-write table only, to reduce version count |
| // related cost on delete bitmap more effectively. |
| DEFINE_mInt64(compaction_promotion_version_count, "1000"); |
| |
| // The lower bound size to do cumulative compaction. When total disk size of candidate rowsets is less than |
| // this size, size_based policy may not do to cumulative compaction. The unit is m byte. |
| DEFINE_mInt64(compaction_min_size_mbytes, "64"); |
| |
| // cumulative compaction policy: min and max delta file's number |
| DEFINE_mInt64(cumulative_compaction_min_deltas, "5"); |
| DEFINE_mInt64(cumulative_compaction_max_deltas, "1000"); |
| DEFINE_mInt32(cumulative_compaction_max_deltas_factor, "10"); |
| |
| // This config can be set to limit thread number in multiget thread pool. |
| DEFINE_mInt32(multi_get_max_threads, "10"); |
| |
| // The upper limit of "permits" held by all compaction tasks. This config can be set to limit memory consumption for compaction. |
| DEFINE_mInt64(total_permits_for_compaction_score, "1000000"); |
| |
| // sleep interval in ms after generated compaction tasks |
| DEFINE_mInt32(generate_compaction_tasks_interval_ms, "100"); |
| |
| // sleep interval in second after update replica infos |
| DEFINE_mInt32(update_replica_infos_interval_seconds, "60"); |
| |
| // Compaction task number per disk. |
| // Must be greater than 2, because Base compaction and Cumulative compaction have at least one thread each. |
| DEFINE_mInt32(compaction_task_num_per_disk, "4"); |
| // compaction thread num for fast disk(typically .SSD), must be greater than 2. |
| DEFINE_mInt32(compaction_task_num_per_fast_disk, "8"); |
| DEFINE_Validator(compaction_task_num_per_disk, |
| [](const int config) -> bool { return config >= 2; }); |
| DEFINE_Validator(compaction_task_num_per_fast_disk, |
| [](const int config) -> bool { return config >= 2; }); |
| DEFINE_Validator(low_priority_compaction_task_num_per_disk, |
| [](const int config) -> bool { return config >= 2; }); |
| |
| // How many rounds of cumulative compaction for each round of base compaction when compaction tasks generation. |
| DEFINE_mInt32(cumulative_compaction_rounds_for_each_base_compaction_round, "9"); |
| // Minimum number of threads required in the thread pool to activate the large cumu compaction delay strategy. |
| // The delay strategy is only applied when the thread pool has at least this many threads. |
| // Default -1 means disable. |
| DEFINE_mInt32(large_cumu_compaction_task_min_thread_num, "-1"); |
| // Maximum size threshold (in bytes) for input rowsets. Compaction tasks with input size |
| // exceeding this threshold will be delayed when thread pool is near capacity. Default 512MB. |
| DEFINE_mInt32(large_cumu_compaction_task_bytes_threshold, "536870912"); |
| // Maximum row count threshold for compaction input. Compaction tasks with row count |
| // exceeding this threshold will be delayed when thread pool is near capacity. Default 1 million. |
| DEFINE_mInt32(large_cumu_compaction_task_row_num_threshold, "1000000"); |
| |
| // Not compact the invisible versions, but with some limitations: |
| // if not timeout, keep no more than compaction_keep_invisible_version_max_count versions; |
| // if timeout, keep no more than compaction_keep_invisible_version_min_count versions. |
| DEFINE_mInt32(compaction_keep_invisible_version_timeout_sec, "1800"); |
| DEFINE_mInt32(compaction_keep_invisible_version_min_count, "50"); |
| DEFINE_mInt32(compaction_keep_invisible_version_max_count, "500"); |
| |
| // Threshold to logging compaction trace, in seconds. |
| DEFINE_mInt32(base_compaction_trace_threshold, "60"); |
| DEFINE_mInt32(cumulative_compaction_trace_threshold, "10"); |
| DEFINE_mBool(disable_compaction_trace_log, "true"); |
| |
| // Interval to picking rowset to compact, in seconds |
| DEFINE_mInt64(pick_rowset_to_compact_interval_sec, "86400"); |
| |
| // Compaction priority schedule |
| DEFINE_mBool(enable_compaction_priority_scheduling, "true"); |
| DEFINE_mInt32(low_priority_compaction_task_num_per_disk, "2"); |
| DEFINE_mInt32(low_priority_compaction_score_threshold, "200"); |
| |
| // Thread count to do tablet meta checkpoint, -1 means use the data directories count. |
| DEFINE_Int32(max_meta_checkpoint_threads, "-1"); |
| |
| // Threshold to logging agent task trace, in seconds. |
| DEFINE_mInt32(agent_task_trace_threshold_sec, "2"); |
| |
| // This config can be set to limit thread number in tablet migration thread pool. |
| DEFINE_Int32(min_tablet_migration_threads, "1"); |
| DEFINE_Int32(max_tablet_migration_threads, "1"); |
| |
| DEFINE_mInt32(finished_migration_tasks_size, "10000"); |
| // If size less than this, the remaining rowsets will be force to complete |
| DEFINE_mInt32(migration_remaining_size_threshold_mb, "10"); |
| // If the task runs longer than this time, the task will be terminated, in seconds. |
| // timeout = std::max(migration_task_timeout_secs, tablet size / 1MB/s) |
| DEFINE_mInt32(migration_task_timeout_secs, "300"); |
| // timeout for try_lock migration lock |
| DEFINE_Int64(migration_lock_timeout_ms, "1000"); |
| |
| // Port to start debug webserver on |
| DEFINE_Int32(webserver_port, "8040"); |
| // Https enable flag |
| DEFINE_Bool(enable_https, "false"); |
| // Path of certificate |
| DEFINE_String(ssl_certificate_path, ""); |
| // Path of private key |
| DEFINE_String(ssl_private_key_path, ""); |
| // Whether to check authorization |
| DEFINE_Bool(enable_all_http_auth, "false"); |
| // Number of webserver workers |
| DEFINE_Int32(webserver_num_workers, "128"); |
| |
| DEFINE_Bool(enable_single_replica_load, "true"); |
| // Number of download workers for single replica load |
| DEFINE_Int32(single_replica_load_download_num_workers, "64"); |
| |
| // Used for mini Load. mini load data file will be removed after this time. |
| DEFINE_Int64(load_data_reserve_hours, "4"); |
| // log error log will be removed after this time |
| DEFINE_mInt64(load_error_log_reserve_hours, "48"); |
| // error log size limit, default 200MB |
| DEFINE_mInt64(load_error_log_limit_bytes, "209715200"); |
| |
| DEFINE_Int32(brpc_heavy_work_pool_threads, "-1"); |
| DEFINE_Int32(brpc_light_work_pool_threads, "-1"); |
| DEFINE_Int32(brpc_heavy_work_pool_max_queue_size, "-1"); |
| DEFINE_Int32(brpc_light_work_pool_max_queue_size, "-1"); |
| DEFINE_mBool(enable_bthread_transmit_block, "true"); |
| DEFINE_Int32(brpc_arrow_flight_work_pool_threads, "-1"); |
| DEFINE_Int32(brpc_arrow_flight_work_pool_max_queue_size, "-1"); |
| |
| //Enable brpc builtin services, see: |
| //https://brpc.apache.org/docs/server/basics/#disable-built-in-services-completely |
| DEFINE_Bool(enable_brpc_builtin_services, "true"); |
| |
| // Enable brpc connection check |
| DEFINE_Bool(enable_brpc_connection_check, "false"); |
| |
| DEFINE_mInt64(brpc_connection_check_timeout_ms, "10000"); |
| |
| // The maximum amount of data that can be processed by a stream load |
| DEFINE_mInt64(streaming_load_max_mb, "102400"); |
| // Some data formats, such as JSON, cannot be streamed. |
| // Therefore, it is necessary to limit the maximum number of |
| // such data when using stream load to prevent excessive memory consumption. |
| DEFINE_mInt64(streaming_load_json_max_mb, "100"); |
| // the alive time of a TabletsChannel. |
| // If the channel does not receive any data till this time, |
| // the channel will be removed. |
| DEFINE_mInt32(streaming_load_rpc_max_alive_time_sec, "1200"); |
| // the timeout of a rpc to open the tablet writer in remote BE. |
| // short operation time, can set a short timeout |
| DEFINE_Int32(tablet_writer_open_rpc_timeout_sec, "60"); |
| // You can ignore brpc error '[E1011]The server is overcrowded' when writing data. |
| DEFINE_mBool(tablet_writer_ignore_eovercrowded, "true"); |
| DEFINE_mInt32(slave_replica_writer_rpc_timeout_sec, "60"); |
| // Whether to enable stream load record function, the default is false. |
| // False: disable stream load record |
| DEFINE_mBool(enable_stream_load_record, "false"); |
| // batch size of stream load record reported to FE |
| DEFINE_mInt32(stream_load_record_batch_size, "50"); |
| // expire time of stream load record in rocksdb. |
| DEFINE_Int32(stream_load_record_expire_time_secs, "28800"); |
| // time interval to clean expired stream load records |
| DEFINE_mInt64(clean_stream_load_record_interval_secs, "1800"); |
| // enable stream load commit txn on BE directly, bypassing FE. Only for cloud. |
| DEFINE_mBool(enable_stream_load_commit_txn_on_be, "false"); |
| // The buffer size to store stream table function schema info |
| DEFINE_Int64(stream_tvf_buffer_size, "1048576"); // 1MB |
| |
| // OlapTableSink sender's send interval, should be less than the real response time of a tablet writer rpc. |
| // You may need to lower the speed when the sink receiver bes are too busy. |
| DEFINE_mInt32(olap_table_sink_send_interval_microseconds, "1000"); |
| DEFINE_mDouble(olap_table_sink_send_interval_auto_partition_factor, "0.001"); |
| |
| // Fragment thread pool |
| DEFINE_Int32(fragment_mgr_asynic_work_pool_thread_num_min, "16"); |
| DEFINE_Int32(fragment_mgr_asynic_work_pool_thread_num_max, "512"); |
| DEFINE_Int32(fragment_mgr_asynic_work_pool_queue_size, "4096"); |
| |
| // Control the number of disks on the machine. If 0, this comes from the system settings. |
| DEFINE_Int32(num_disks, "0"); |
| // The read size is the size of the reads sent to os. |
| // There is a trade off of latency and throughout, trying to keep disks busy but |
| // not introduce seeks. The literature seems to agree that with 8 MB reads, random |
| // io and sequential io perform similarly. |
| DEFINE_Int32(read_size, "8388608"); // 8 * 1024 * 1024, Read Size (in bytes) |
| DEFINE_Int32(min_buffer_size, "1024"); // 1024, The minimum read buffer size (in bytes) |
| |
| // for pprof |
| DEFINE_String(pprof_profile_dir, "${DORIS_HOME}/log"); |
| // for jeprofile in jemalloc |
| DEFINE_mString(jeprofile_dir, "${DORIS_HOME}/log"); |
| DEFINE_mBool(enable_je_purge_dirty_pages, "true"); |
| DEFINE_mInt32(je_dirty_decay_ms, "5000"); |
| |
| // to forward compatibility, will be removed later |
| DEFINE_mBool(enable_token_check, "true"); |
| |
| // to open/close system metrics |
| DEFINE_Bool(enable_system_metrics, "true"); |
| |
| // Number of cores Doris will used, this will effect only when it's greater than 0. |
| // Otherwise, Doris will use all cores returned from "/proc/cpuinfo". |
| DEFINE_Int32(num_cores, "0"); |
| |
| // When BE start, If there is a broken disk, BE process will exit by default. |
| // Otherwise, we will ignore the broken disk, |
| DEFINE_Bool(ignore_broken_disk, "false"); |
| |
| // Sleep time in milliseconds between memory maintenance iterations |
| DEFINE_mInt32(memory_maintenance_sleep_time_ms, "50"); |
| |
| // Memory gc are expensive, wait a while to avoid too frequent. |
| DEFINE_mInt32(memory_gc_sleep_time_ms, "500"); |
| |
| // max write buffer size before flush, default 200MB |
| DEFINE_mInt64(write_buffer_size, "209715200"); |
| // max buffer size used in memtable for the aggregated table, default 400MB |
| DEFINE_mInt64(write_buffer_size_for_agg, "419430400"); |
| // max parallel flush task per memtable writer |
| DEFINE_mInt32(memtable_flush_running_count_limit, "2"); |
| |
| // maximum sleep time to wait for memory when writing or flushing memtable. |
| DEFINE_mInt32(memtable_wait_for_memory_sleep_time_s, "300"); |
| |
| DEFINE_Int32(load_process_max_memory_limit_percent, "50"); // 50% |
| |
| // If the memory consumption of load jobs exceed load_process_max_memory_limit, |
| // all load jobs will hang there to wait for memtable flush. We should have a |
| // soft limit which can trigger the memtable flush for the load channel who |
| // consumes lagest memory size before we reach the hard limit. The soft limit |
| // might avoid all load jobs hang at the same time. |
| DEFINE_Int32(load_process_soft_mem_limit_percent, "80"); |
| |
| // If load memory consumption is within load_process_safe_mem_permit_percent, |
| // memtable memory limiter will do nothing. |
| DEFINE_Int32(load_process_safe_mem_permit_percent, "5"); |
| |
| // If there are a lot of memtable memory, then wait them flush finished. |
| DEFINE_mDouble(load_max_wg_active_memtable_percent, "0.6"); |
| |
| // result buffer cancelled time (unit: second) |
| DEFINE_mInt32(result_buffer_cancelled_interval_time, "300"); |
| |
| // arrow flight result sink buffer rows size, default 4096 * 8 |
| DEFINE_mInt32(arrow_flight_result_sink_buffer_size_rows, "32768"); |
| // The timeout for ADBC Client to wait for data using arrow flight reader. |
| // If the query is very complex and no result is generated after this time, consider increasing this timeout. |
| DEFINE_mInt32(arrow_flight_reader_brpc_controller_timeout_ms, "300000"); |
| |
| // the increased frequency of priority for remaining tasks in BlockingPriorityQueue |
| DEFINE_mInt32(priority_queue_remaining_tasks_increased_frequency, "512"); |
| |
| // sync tablet_meta when modifying meta |
| DEFINE_mBool(sync_tablet_meta, "false"); |
| |
| // sync when closing a file writer |
| DEFINE_mBool(sync_file_on_close, "true"); |
| |
| // default thrift rpc timeout ms |
| DEFINE_mInt32(thrift_rpc_timeout_ms, "60000"); |
| |
| // txn commit rpc timeout |
| DEFINE_mInt32(txn_commit_rpc_timeout_ms, "180000"); |
| |
| // If set to true, metric calculator will run |
| DEFINE_Bool(enable_metric_calculator, "true"); |
| |
| // max consumer num in one data consumer group, for routine load |
| DEFINE_mInt32(max_consumer_num_per_group, "3"); |
| |
| // the max size of thread pool for routine load task. |
| // this should be larger than FE config 'max_routine_load_task_num_per_be' (default 5) |
| DEFINE_Int32(max_routine_load_thread_pool_size, "1024"); |
| |
| // max external scan cache batch count, means cache max_memory_cache_batch_count * batch_size row |
| // default is 20, batch_size's default value is 1024 means 20 * 1024 rows will be cached |
| DEFINE_mInt32(max_memory_sink_batch_count, "20"); |
| |
| // This configuration is used for the context gc thread schedule period |
| // note: unit is minute, default is 5min |
| DEFINE_mInt32(scan_context_gc_interval_min, "5"); |
| |
| // es scroll keep-alive |
| DEFINE_String(es_scroll_keepalive, "5m"); |
| |
| // HTTP connection timeout for es |
| DEFINE_mInt32(es_http_timeout_ms, "5000"); |
| |
| // the max client cache number per each host |
| // There are variety of client cache in BE, but currently we use the |
| // same cache size configuration. |
| // TODO(cmy): use different config to set different client cache if necessary. |
| DEFINE_Int32(max_client_cache_size_per_host, "10"); |
| |
| DEFINE_Int32(max_master_fe_client_cache_size, "10"); |
| |
| // Dir to save files downloaded by SmallFileMgr |
| DEFINE_String(small_file_dir, "${DORIS_HOME}/lib/small_file/"); |
| // path gc |
| DEFINE_Bool(path_gc_check, "true"); |
| DEFINE_mInt32(path_gc_check_interval_second, "86400"); |
| DEFINE_mInt32(path_gc_check_step, "1000"); |
| DEFINE_mInt32(path_gc_check_step_interval_ms, "10"); |
| |
| // The following 2 configs limit the max usage of disk capacity of a data dir. |
| // If both of these 2 threshold reached, no more data can be writen into that data dir. |
| // The percent of max used capacity of a data dir |
| DEFINE_mInt32(storage_flood_stage_usage_percent, "90"); // 90% |
| // The min bytes that should be left of a data dir |
| DEFINE_mInt64(storage_flood_stage_left_capacity_bytes, "1073741824"); // 1GB |
| // number of thread for flushing memtable per store |
| DEFINE_Int32(flush_thread_num_per_store, "6"); |
| // number of thread for flushing memtable per store, for high priority load task |
| DEFINE_Int32(high_priority_flush_thread_num_per_store, "6"); |
| // number of threads = min(flush_thread_num_per_store * num_store, |
| // max_flush_thread_num_per_cpu * num_cpu) |
| DEFINE_Int32(max_flush_thread_num_per_cpu, "4"); |
| |
| DEFINE_mInt32(wg_flush_thread_num_per_store, "6"); |
| DEFINE_mInt32(wg_flush_thread_num_per_cpu, "4"); |
| |
| // config for tablet meta checkpoint |
| DEFINE_mInt32(tablet_meta_checkpoint_min_new_rowsets_num, "10"); |
| DEFINE_mInt32(tablet_meta_checkpoint_min_interval_secs, "600"); |
| DEFINE_Int32(generate_tablet_meta_checkpoint_tasks_interval_secs, "600"); |
| |
| // config for default rowset type |
| // Valid configs: ALPHA, BETA |
| DEFINE_String(default_rowset_type, "BETA"); |
| |
| // Maximum size of a single message body in all protocols |
| DEFINE_Int64(brpc_max_body_size, "3147483648"); |
| DEFINE_Int64(brpc_socket_max_unwritten_bytes, "-1"); |
| // TODO(zxy): expect to be true in v1.3 |
| // Whether to embed the ProtoBuf Request serialized string together with Tuple/Block data into |
| // Controller Attachment and send it through http brpc when the length of the Tuple/Block data |
| // is greater than 1.8G. This is to avoid the error of Request length overflow (2G). |
| DEFINE_mBool(transfer_large_data_by_brpc, "true"); |
| |
| // max number of txns for every txn_partition_map in txn manager |
| // this is a self protection to avoid too many txns saving in manager |
| DEFINE_mInt64(max_runnings_transactions_per_txn_map, "2000"); |
| |
| // tablet_map_lock shard size, the value is 2^n, n=0,1,2,3,4 |
| // this is a an enhancement for better performance to manage tablet |
| DEFINE_Int32(tablet_map_shard_size, "256"); |
| |
| // txn_map_lock shard size, the value is 2^n, n=0,1,2,3,4 |
| // this is a an enhancement for better performance to manage txn |
| DEFINE_Int32(txn_map_shard_size, "1024"); |
| |
| // txn_lock shard size, the value is 2^n, n=0,1,2,3,4 |
| // this is a an enhancement for better performance to commit and publish txn |
| DEFINE_Int32(txn_shard_size, "1024"); |
| |
| // Whether to continue to start be when load tablet from header failed. |
| DEFINE_Bool(ignore_load_tablet_failure, "false"); |
| |
| // Whether to continue to start be when load tablet from header failed. |
| DEFINE_mBool(ignore_rowset_stale_unconsistent_delete, "false"); |
| |
| // Set max cache's size of query results, the unit is M byte |
| DEFINE_Int32(query_cache_max_size_mb, "256"); |
| |
| // Cache memory is pruned when reach query_cache_max_size_mb + query_cache_elasticity_size_mb |
| DEFINE_Int32(query_cache_elasticity_size_mb, "128"); |
| |
| // Maximum number of cache partitions corresponding to a SQL |
| DEFINE_Int32(query_cache_max_partition_count, "1024"); |
| |
| // Maximum number of version of a tablet. If the version num of a tablet exceed limit, |
| // the load process will reject new incoming load job of this tablet. |
| // This is to avoid too many version num. |
| DEFINE_mInt32(max_tablet_version_num, "2000"); |
| |
| // Frontend mainly use two thrift sever type: THREAD_POOL, THREADED_SELECTOR. if fe use THREADED_SELECTOR model for thrift server, |
| // the thrift_server_type_of_fe should be set THREADED_SELECTOR to make be thrift client to fe constructed with TFramedTransport |
| DEFINE_String(thrift_server_type_of_fe, "THREAD_POOL"); |
| |
| // disable zone map index when page row is too few |
| DEFINE_mInt32(zone_map_row_num_threshold, "20"); |
| |
| // aws sdk log level |
| // Off = 0, |
| // Fatal = 1, |
| // Error = 2, |
| // Warn = 3, |
| // Info = 4, |
| // Debug = 5, |
| // Trace = 6 |
| DEFINE_Int32(aws_log_level, "2"); |
| |
| // the buffer size when read data from remote storage like s3 |
| DEFINE_mInt32(remote_storage_read_buffer_mb, "16"); |
| |
| // The minimum length when TCMalloc Hook consumes/releases MemTracker, consume size |
| // smaller than this value will continue to accumulate. specified as number of bytes. |
| // Decreasing this value will increase the frequency of consume/release. |
| // Increasing this value will cause MemTracker statistics to be inaccurate. |
| DEFINE_mInt32(mem_tracker_consume_min_size_bytes, "1048576"); |
| |
| // The version information of the tablet will be stored in the memory |
| // in an adjacency graph data structure. |
| // And as the new version is written and the old version is deleted, |
| // the data structure will begin to have empty vertex with no edge associations(orphan vertex). |
| // This config is used to control that when the proportion of orphan vertex is greater than the threshold, |
| // the adjacency graph will be rebuilt to ensure that the data structure will not expand indefinitely. |
| // This config usually only needs to be modified during testing. |
| // In most cases, it does not need to be modified. |
| DEFINE_mDouble(tablet_version_graph_orphan_vertex_ratio, "0.1"); |
| |
| // share delta writers when memtable_on_sink_node = true |
| DEFINE_Bool(share_delta_writers, "true"); |
| // timeout for open load stream rpc in ms |
| DEFINE_Int64(open_load_stream_timeout_ms, "60000"); // 60s |
| // enable write background when using brpc stream |
| DEFINE_mBool(enable_brpc_stream_write_background, "true"); |
| |
| // brpc streaming max_buf_size in bytes |
| DEFINE_Int64(load_stream_max_buf_size, "20971520"); // 20MB |
| // brpc streaming messages_in_batch |
| DEFINE_Int32(load_stream_messages_in_batch, "128"); |
| // brpc streaming StreamWait seconds on EAGAIN |
| DEFINE_Int32(load_stream_eagain_wait_seconds, "600"); |
| // max tasks per flush token in load stream |
| DEFINE_Int32(load_stream_flush_token_max_tasks, "15"); |
| // max wait flush token time in load stream |
| DEFINE_Int32(load_stream_max_wait_flush_token_time_ms, "600000"); |
| // number of send batch thread pool size |
| DEFINE_Int32(send_batch_thread_pool_thread_num, "64"); |
| // number of send batch thread pool queue size |
| DEFINE_Int32(send_batch_thread_pool_queue_size, "102400"); |
| |
| // Limit the number of segment of a newly created rowset. |
| // The newly created rowset may to be compacted after loading, |
| // so if there are too many segment in a rowset, the compaction process |
| // will run out of memory. |
| // When doing compaction, each segment may take at least 1MB buffer. |
| DEFINE_mInt32(max_segment_num_per_rowset, "1000"); |
| DEFINE_mInt32(segment_compression_threshold_kb, "256"); |
| |
| // Time to clean up useless JDBC connection pool cache |
| DEFINE_mInt32(jdbc_connection_pool_cache_clear_time_sec, "28800"); |
| |
| // Global bitmap cache capacity for aggregation cache, size in bytes |
| DEFINE_Int64(delete_bitmap_agg_cache_capacity, "104857600"); |
| // The default delete bitmap cache is set to 100MB, |
| // which can be insufficient and cause performance issues when the amount of user data is large. |
| // To mitigate the problem of an inadequate cache, |
| // we will take the larger of 1.0% of the total memory and 100MB as the delete bitmap cache size. |
| DEFINE_String(delete_bitmap_dynamic_agg_cache_limit, "1.0%"); |
| DEFINE_mInt32(delete_bitmap_agg_cache_stale_sweep_time_sec, "1800"); |
| |
| // reference https://github.com/edenhill/librdkafka/blob/master/INTRODUCTION.md#broker-version-compatibility |
| // If the dependent kafka broker version older than 0.10.0.0, |
| // the value of kafka_api_version_request should be false, and the |
| // value set by the fallback version kafka_broker_version_fallback will be used, |
| // and the valid values are: 0.9.0.x, 0.8.x.y. |
| DEFINE_String(kafka_api_version_request, "true"); |
| DEFINE_String(kafka_broker_version_fallback, "0.10.0"); |
| DEFINE_String(kafka_debug, "disable"); |
| |
| // The number of pool siz of routine load consumer. |
| // If you meet the error describe in https://github.com/edenhill/librdkafka/issues/3608 |
| // Change this size to 0 to fix it temporarily. |
| DEFINE_mInt32(routine_load_consumer_pool_size, "1024"); |
| |
| // Used in single-stream-multi-table load. When receive a batch of messages from kafka, |
| // if the size of batch is more than this threshold, we will request plans for all related tables. |
| DEFINE_Int32(multi_table_batch_plan_threshold, "200"); |
| |
| // Used in single-stream-multi-table load. When receiving a batch of messages from Kafka, |
| // if the size of the table wait for plan is more than this threshold, we will request plans for all related tables. |
| // The param is aimed to avoid requesting and executing too many plans at once. |
| // Performing small batch processing on multiple tables during the loaded process can reduce the pressure of a single RPC |
| // and improve the real-time processing of data. |
| DEFINE_Int32(multi_table_max_wait_tables, "5"); |
| |
| // When the timeout of a load task is less than this threshold, |
| // Doris treats it as a high priority task. |
| // high priority tasks use a separate thread pool for flush and do not block rpc by memory cleanup logic. |
| // this threshold is mainly used to identify routine load tasks and should not be modified if not necessary. |
| DEFINE_mInt32(load_task_high_priority_threshold_second, "600"); |
| |
| // The min timeout of load rpc (add batch, close, etc.) |
| // Because a load rpc may be blocked for a while. |
| // Increase this config may avoid rpc timeout. |
| DEFINE_mInt32(min_load_rpc_timeout_ms, "20000"); |
| |
| // use which protocol to access function service, candicate is baidu_std/h2:grpc |
| DEFINE_String(function_service_protocol, "h2:grpc"); |
| |
| // use which load balancer to select server to connect |
| DEFINE_String(rpc_load_balancer, "rr"); |
| |
| // a soft limit of string type length, the hard limit is 2GB - 4, but if too long will cause very low performance, |
| // so we set a soft limit, default is 1MB |
| DEFINE_Int32(string_type_length_soft_limit_bytes, "1048576"); |
| |
| DEFINE_Validator(string_type_length_soft_limit_bytes, |
| [](const int config) -> bool { return config > 0 && config <= 2147483643; }); |
| |
| // Threshold of reading a small file into memory |
| DEFINE_mInt32(in_memory_file_size, "1048576"); // 1MB |
| |
| // Max size of parquet page header in bytes |
| DEFINE_mInt32(parquet_header_max_size_mb, "1"); |
| // Max buffer size for parquet row group |
| DEFINE_mInt32(parquet_rowgroup_max_buffer_mb, "128"); |
| // Max buffer size for parquet chunk column |
| DEFINE_mInt32(parquet_column_max_buffer_mb, "8"); |
| DEFINE_mDouble(max_amplified_read_ratio, "0.8"); |
| DEFINE_mInt32(merged_oss_min_io_size, "1048576"); |
| DEFINE_mInt32(merged_hdfs_min_io_size, "8192"); |
| |
| // OrcReader |
| DEFINE_mInt32(orc_natural_read_size_mb, "8"); |
| DEFINE_mInt64(big_column_size_buffer, "65535"); |
| DEFINE_mInt64(small_column_size_buffer, "100"); |
| |
| // Perform the always_true check at intervals determined by runtime_filter_sampling_frequency |
| DEFINE_mInt32(runtime_filter_sampling_frequency, "64"); |
| DEFINE_mInt32(execution_max_rpc_timeout_sec, "3600"); |
| DEFINE_mBool(execution_ignore_eovercrowded, "true"); |
| // cooldown task configs |
| DEFINE_Int32(cooldown_thread_num, "5"); |
| DEFINE_mInt64(generate_cooldown_task_interval_sec, "20"); |
| DEFINE_mInt32(remove_unused_remote_files_interval_sec, "21600"); // 6h |
| DEFINE_mInt32(confirm_unused_remote_files_interval_sec, "60"); |
| DEFINE_Int32(cold_data_compaction_thread_num, "2"); |
| DEFINE_mInt32(cold_data_compaction_interval_sec, "1800"); |
| |
| DEFINE_String(tmp_file_dir, "tmp"); |
| |
| DEFINE_Int32(min_s3_file_system_thread_num, "16"); |
| DEFINE_Int32(max_s3_file_system_thread_num, "64"); |
| |
| DEFINE_Bool(enable_time_lut, "true"); |
| DEFINE_mBool(enable_simdjson_reader, "true"); |
| |
| DEFINE_mBool(enable_query_like_bloom_filter, "true"); |
| // number of s3 scanner thread pool size |
| DEFINE_Int32(doris_remote_scanner_thread_pool_thread_num, "48"); |
| // number of s3 scanner thread pool queue size |
| DEFINE_Int32(doris_remote_scanner_thread_pool_queue_size, "102400"); |
| DEFINE_mInt64(block_cache_wait_timeout_ms, "1000"); |
| |
| // limit the queue of pending batches which will be sent by a single nodechannel |
| DEFINE_mInt64(nodechannel_pending_queue_max_bytes, "67108864"); |
| |
| // The batch size for sending data by brpc streaming client |
| DEFINE_mInt64(brpc_streaming_client_batch_bytes, "262144"); |
| |
| // Max waiting time to wait the "plan fragment start" rpc. |
| // If timeout, the fragment will be cancelled. |
| // This parameter is usually only used when the FE loses connection, |
| // and the BE can automatically cancel the relevant fragment after the timeout, |
| // so as to avoid occupying the execution thread for a long time. |
| DEFINE_mInt32(max_fragment_start_wait_time_seconds, "30"); |
| |
| DEFINE_mInt32(fragment_mgr_cancel_worker_interval_seconds, "1"); |
| |
| // Node role tag for backend. Mix role is the default role, and computation role have no |
| // any tablet. |
| DEFINE_String(be_node_role, "mix"); |
| |
| // Hide webserver page for safety. |
| // Hide the be config page for webserver. |
| DEFINE_Bool(hide_webserver_config_page, "false"); |
| |
| DEFINE_Bool(enable_segcompaction, "true"); |
| |
| // Max number of segments allowed in a single segcompaction task. |
| DEFINE_mInt32(segcompaction_batch_size, "10"); |
| |
| // Max row count allowed in a single source segment, bigger segments will be skipped. |
| DEFINE_Int32(segcompaction_candidate_max_rows, "1048576"); |
| |
| // Max file size allowed in a single source segment, bigger segments will be skipped. |
| DEFINE_Int64(segcompaction_candidate_max_bytes, "104857600"); |
| |
| // Max total row count allowed in a single segcompaction task. |
| DEFINE_Int32(segcompaction_task_max_rows, "1572864"); |
| |
| // Max total file size allowed in a single segcompaction task. |
| DEFINE_Int64(segcompaction_task_max_bytes, "157286400"); |
| |
| // Global segcompaction thread pool size. |
| DEFINE_mInt32(segcompaction_num_threads, "5"); |
| |
| // enable java udf and jdbc scannode |
| DEFINE_Bool(enable_java_support, "true"); |
| |
| // enable prefetch tablets before opening |
| DEFINE_mBool(enable_prefetch_tablet, "true"); |
| |
| // Set config randomly to check more issues in github workflow |
| DEFINE_Bool(enable_fuzzy_mode, "false"); |
| |
| DEFINE_Bool(enable_debug_points, "false"); |
| |
| DEFINE_Int32(pipeline_executor_size, "0"); |
| DEFINE_Bool(enable_workload_group_for_scan, "false"); |
| DEFINE_mInt64(workload_group_scan_task_wait_timeout_ms, "10000"); |
| |
| // Whether use schema dict in backend side instead of MetaService side(cloud mode) |
| DEFINE_mBool(variant_use_cloud_schema_dict_cache, "true"); |
| DEFINE_mDouble(variant_ratio_of_defaults_as_sparse_column, "1"); |
| DEFINE_mInt64(variant_threshold_rows_to_estimate_sparse_column, "2048"); |
| DEFINE_mBool(variant_throw_exeception_on_invalid_json, "false"); |
| |
| // block file cache |
| DEFINE_Bool(enable_file_cache, "false"); |
| // format: [{"path":"/path/to/file_cache","total_size":21474836480,"query_limit":10737418240}] |
| // format: [{"path":"/path/to/file_cache","total_size":21474836480,"query_limit":10737418240},{"path":"/path/to/file_cache2","total_size":21474836480,"query_limit":10737418240}] |
| // format: {"path": "/path/to/file_cache", "total_size":53687091200, "ttl_percent":50, "normal_percent":40, "disposable_percent":5, "index_percent":5} |
| // format: [{"path": "xxx", "total_size":53687091200, "storage": "memory"}] |
| // Note1: storage is "disk" by default |
| // Note2: when the storage is "memory", the path is ignored. So you can set xxx to anything you like |
| // and doris will just reset the path to "memory" internally. |
| // In a very wierd case when your storage is disk, and the directory, by accident, is named |
| // "memory" for some reason, you should write the path as: |
| // {"path": "memory", "total_size":53687091200, "storage": "disk"} |
| // or use the default storage value: |
| // {"path": "memory", "total_size":53687091200} |
| // Both will use the directory "memory" on the disk instead of the real RAM. |
| DEFINE_String(file_cache_path, "[{\"path\":\"${DORIS_HOME}/file_cache\"}]"); |
| DEFINE_Int64(file_cache_each_block_size, "1048576"); // 1MB |
| |
| DEFINE_Bool(clear_file_cache, "false"); |
| DEFINE_Bool(enable_file_cache_query_limit, "false"); |
| DEFINE_mInt32(file_cache_enter_disk_resource_limit_mode_percent, "88"); |
| DEFINE_mInt32(file_cache_exit_disk_resource_limit_mode_percent, "80"); |
| DEFINE_mBool(enable_evict_file_cache_in_advance, "true"); |
| DEFINE_mInt32(file_cache_enter_need_evict_cache_in_advance_percent, "78"); |
| DEFINE_mInt32(file_cache_exit_need_evict_cache_in_advance_percent, "75"); |
| DEFINE_mInt32(file_cache_evict_in_advance_interval_ms, "1000"); |
| DEFINE_mInt64(file_cache_evict_in_advance_batch_bytes, "31457280"); // 30MB |
| DEFINE_mInt64(file_cache_evict_in_advance_recycle_keys_num_threshold, "1000"); |
| |
| DEFINE_mBool(enable_read_cache_file_directly, "false"); |
| DEFINE_mBool(file_cache_enable_evict_from_other_queue_by_size, "true"); |
| // If true, evict the ttl cache using LRU when full. |
| // Otherwise, only expiration can evict ttl and new data won't add to cache when full. |
| DEFINE_Bool(enable_ttl_cache_evict_using_lru, "true"); |
| DEFINE_mBool(enbale_dump_error_file, "false"); |
| // limit the max size of error log on disk |
| DEFINE_mInt64(file_cache_error_log_limit_bytes, "209715200"); // 200MB |
| DEFINE_mInt64(cache_lock_wait_long_tail_threshold_us, "30000000"); |
| DEFINE_mInt64(cache_lock_held_long_tail_threshold_us, "30000000"); |
| DEFINE_mBool(enable_file_cache_keep_base_compaction_output, "false"); |
| DEFINE_mInt64(file_cache_remove_block_qps_limit, "1000"); |
| DEFINE_mInt64(file_cache_background_gc_interval_ms, "100"); |
| DEFINE_mBool(enable_reader_dryrun_when_download_file_cache, "true"); |
| DEFINE_mInt64(file_cache_background_monitor_interval_ms, "5000"); |
| DEFINE_mInt64(file_cache_background_ttl_gc_interval_ms, "3000"); |
| DEFINE_mInt64(file_cache_background_ttl_gc_batch, "1000"); |
| |
| DEFINE_mInt32(index_cache_entry_stay_time_after_lookup_s, "1800"); |
| DEFINE_mInt32(inverted_index_cache_stale_sweep_time_sec, "600"); |
| DEFINE_mBool(enable_write_index_searcher_cache, "false"); |
| // inverted index searcher cache size |
| DEFINE_String(inverted_index_searcher_cache_limit, "10%"); |
| DEFINE_Bool(enable_inverted_index_cache_check_timestamp, "true"); |
| DEFINE_mBool(enable_inverted_index_correct_term_write, "true"); |
| DEFINE_Int32(inverted_index_fd_number_limit_percent, "20"); // 20% |
| DEFINE_Int32(inverted_index_query_cache_shards, "256"); |
| |
| // inverted index match bitmap cache size |
| DEFINE_String(inverted_index_query_cache_limit, "10%"); |
| |
| // inverted index |
| DEFINE_mDouble(inverted_index_ram_buffer_size, "512"); |
| // -1 indicates not working. |
| // Normally we should not change this, it's useful for testing. |
| DEFINE_mInt32(inverted_index_max_buffered_docs, "-1"); |
| // dict path for chinese analyzer |
| DEFINE_String(inverted_index_dict_path, "${DORIS_HOME}/dict"); |
| DEFINE_Int32(inverted_index_read_buffer_size, "4096"); |
| // tree depth for bkd index |
| DEFINE_Int32(max_depth_in_bkd_tree, "32"); |
| // index compaction |
| DEFINE_mBool(inverted_index_compaction_enable, "true"); |
| // Only for debug, do not use in production |
| DEFINE_mBool(debug_inverted_index_compaction, "false"); |
| // index by RAM directory |
| DEFINE_mBool(inverted_index_ram_dir_enable, "true"); |
| // wheather index by RAM directory when base compaction |
| DEFINE_mBool(inverted_index_ram_dir_enable_when_base_compaction, "true"); |
| // use num_broadcast_buffer blocks as buffer to do broadcast |
| DEFINE_Int32(num_broadcast_buffer, "32"); |
| |
| // max depth of expression tree allowed. |
| DEFINE_Int32(max_depth_of_expr_tree, "600"); |
| |
| // Report a tablet as bad when io errors occurs more than this value. |
| DEFINE_mInt64(max_tablet_io_errors, "-1"); |
| |
| // Report a tablet as bad when its path not found |
| DEFINE_Int32(tablet_path_check_interval_seconds, "-1"); |
| DEFINE_mInt32(tablet_path_check_batch_size, "1000"); |
| |
| // it must be larger than or equal to 5MB |
| DEFINE_mInt64(s3_write_buffer_size, "5242880"); |
| // Log interval when doing s3 upload task |
| DEFINE_mInt32(s3_file_writer_log_interval_second, "60"); |
| DEFINE_mInt64(file_cache_max_file_reader_cache_size, "1000000"); |
| DEFINE_mInt64(hdfs_write_batch_buffer_size_mb, "1"); // 1MB |
| |
| //disable shrink memory by default |
| DEFINE_mBool(enable_shrink_memory, "false"); |
| DEFINE_mInt32(schema_cache_capacity, "1024"); |
| DEFINE_mInt32(schema_cache_sweep_time_sec, "100"); |
| |
| // max number of segment cache, default -1 for backward compatibility fd_number*2/5 |
| DEFINE_Int32(segment_cache_capacity, "-1"); |
| DEFINE_Int32(segment_cache_fd_percentage, "20"); |
| DEFINE_mInt32(estimated_mem_per_column_reader, "512"); |
| DEFINE_Int32(segment_cache_memory_percentage, "5"); |
| |
| // enable feature binlog, default false |
| DEFINE_Bool(enable_feature_binlog, "false"); |
| |
| // enable set in BitmapValue |
| DEFINE_Bool(enable_set_in_bitmap_value, "true"); |
| |
| DEFINE_Int64(max_hdfs_file_handle_cache_num, "20000"); |
| DEFINE_Int32(max_hdfs_file_handle_cache_time_sec, "28800"); |
| DEFINE_Int64(max_external_file_meta_cache_num, "1000"); |
| DEFINE_mInt32(common_obj_lru_cache_stale_sweep_time_sec, "900"); |
| // Apply delete pred in cumu compaction |
| DEFINE_mBool(enable_delete_when_cumu_compaction, "false"); |
| |
| // max_write_buffer_number for rocksdb |
| DEFINE_Int32(rocksdb_max_write_buffer_number, "5"); |
| |
| DEFINE_mBool(allow_zero_date, "false"); |
| DEFINE_Bool(allow_invalid_decimalv2_literal, "false"); |
| DEFINE_mString(kerberos_ccache_path, "/tmp/"); |
| DEFINE_mString(kerberos_krb5_conf_path, "/etc/krb5.conf"); |
| // Deprecated |
| DEFINE_mInt32(kerberos_refresh_interval_second, "43200"); |
| |
| DEFINE_mString(get_stack_trace_tool, "libunwind"); |
| DEFINE_mString(dwarf_location_info_mode, "FAST"); |
| DEFINE_mBool(enable_address_sanitizers_with_stack_trace, "true"); |
| |
| // the ratio of _prefetch_size/_batch_size in AutoIncIDBuffer |
| DEFINE_mInt64(auto_inc_prefetch_size_ratio, "10"); |
| |
| // the ratio of _low_level_water_level_mark/_batch_size in AutoIncIDBuffer |
| DEFINE_mInt64(auto_inc_low_water_level_mark_size_ratio, "3"); |
| |
| // number of threads that fetch auto-inc ranges from FE |
| DEFINE_mInt64(auto_inc_fetch_thread_num, "3"); |
| // default max to 2048 connections |
| DEFINE_mInt64(lookup_connection_cache_capacity, "2048"); |
| |
| // level of compression when using LZ4_HC, whose defalut value is LZ4HC_CLEVEL_DEFAULT |
| DEFINE_mInt64(LZ4_HC_compression_level, "9"); |
| |
| DEFINE_mBool(enable_merge_on_write_correctness_check, "true"); |
| // USED FOR DEBUGING |
| // core directly if the compaction found there's duplicate key on mow table |
| DEFINE_mBool(enable_mow_compaction_correctness_check_core, "false"); |
| // USED FOR DEBUGING |
| // let compaction fail if the compaction found there's duplicate key on mow table |
| DEFINE_mBool(enable_mow_compaction_correctness_check_fail, "false"); |
| // rowid conversion correctness check when compaction for mow table |
| DEFINE_mBool(enable_rowid_conversion_correctness_check, "false"); |
| // missing rows correctness check when compaction for mow table |
| DEFINE_mBool(enable_missing_rows_correctness_check, "false"); |
| // When the number of missing versions is more than this value, do not directly |
| // retry the publish and handle it through async publish. |
| DEFINE_mInt32(mow_publish_max_discontinuous_version_num, "20"); |
| // When the size of primary keys in memory exceeds this value, finish current segment |
| // and create a new segment, used in compaction. Default 50MB. |
| DEFINE_mInt64(mow_primary_key_index_max_size_in_memory, "52428800"); |
| // When the version is not continuous for MOW table in publish phase and the gap between |
| // current txn's publishing version and the max version of the tablet exceeds this value, |
| // don't print warning log |
| DEFINE_mInt32(publish_version_gap_logging_threshold, "200"); |
| // get agg by cache for mow table |
| DEFINE_mBool(enable_mow_get_agg_by_cache, "true"); |
| // get agg correctness check for mow table |
| DEFINE_mBool(enable_mow_get_agg_correctness_check_core, "false"); |
| |
| // The secure path with user files, used in the `local` table function. |
| DEFINE_mString(user_files_secure_path, "${DORIS_HOME}"); |
| |
| DEFINE_Int32(fe_expire_duration_seconds, "60"); |
| |
| DEFINE_Int32(grace_shutdown_wait_seconds, "120"); |
| |
| DEFINE_Int16(bitmap_serialize_version, "1"); |
| |
| // group commit config |
| DEFINE_String(group_commit_wal_path, ""); |
| DEFINE_Int32(group_commit_replay_wal_retry_num, "10"); |
| DEFINE_Int32(group_commit_replay_wal_retry_interval_seconds, "5"); |
| DEFINE_Int32(group_commit_replay_wal_retry_interval_max_seconds, "1800"); |
| DEFINE_Int32(group_commit_relay_wal_threads, "10"); |
| // This config can be set to limit thread number in group commit request fragment thread pool. |
| DEFINE_Int32(group_commit_insert_threads, "10"); |
| DEFINE_Int32(group_commit_memory_rows_for_max_filter_ratio, "10000"); |
| DEFINE_Bool(wait_internal_group_commit_finish, "false"); |
| // Max size(bytes) of group commit queues, used for mem back pressure, defult 64M. |
| DEFINE_mInt32(group_commit_queue_mem_limit, "67108864"); |
| // Max size(bytes) or percentage(%) of wal disk usage, used for disk space back pressure, default 10% of the disk available space. |
| // group_commit_wal_max_disk_limit=1024 or group_commit_wal_max_disk_limit=10% can be automatically identified. |
| DEFINE_String(group_commit_wal_max_disk_limit, "10%"); |
| DEFINE_Bool(group_commit_wait_replay_wal_finish, "false"); |
| |
| DEFINE_mInt32(scan_thread_nice_value, "0"); |
| DEFINE_mInt32(tablet_schema_cache_recycle_interval, "3600"); |
| DEFINE_mInt32(tablet_schema_cache_capacity, "102400"); |
| |
| DEFINE_Bool(exit_on_exception, "false"); |
| // This config controls whether the s3 file writer would flush cache asynchronously |
| DEFINE_Bool(enable_flush_file_cache_async, "true"); |
| |
| // cgroup |
| DEFINE_String(doris_cgroup_cpu_path, ""); |
| |
| DEFINE_mBool(enable_be_proc_monitor, "false"); |
| DEFINE_mInt32(be_proc_monitor_interval_ms, "10000"); |
| |
| DEFINE_Int32(workload_group_metrics_interval_ms, "5000"); |
| |
| DEFINE_Bool(ignore_always_true_predicate_for_segment, "true"); |
| |
| // Ingest binlog work pool size, -1 is disable, 0 is hardware concurrency |
| DEFINE_Int32(ingest_binlog_work_pool_size, "-1"); |
| |
| // Ingest binlog with persistent connection |
| DEFINE_Bool(enable_ingest_binlog_with_persistent_connection, "false"); |
| |
| // Log ingest binlog elapsed threshold, -1 is disabled |
| DEFINE_mInt64(ingest_binlog_elapsed_threshold_ms, "-1"); |
| |
| // Download binlog rate limit, unit is KB/s, 0 means no limit |
| DEFINE_Int32(download_binlog_rate_limit_kbs, "0"); |
| |
| DEFINE_mInt32(buffered_reader_read_timeout_ms, "600000"); |
| |
| DEFINE_Bool(enable_snapshot_action, "false"); |
| |
| DEFINE_mInt32(variant_max_merged_tablet_schema_size, "2048"); |
| |
| DEFINE_mBool(enable_column_type_check, "true"); |
| // 128 MB |
| DEFINE_mInt64(local_exchange_buffer_mem_limit, "134217728"); |
| |
| // Default 300s, if its value <= 0, then log is disabled |
| DEFINE_mInt64(enable_debug_log_timeout_secs, "0"); |
| |
| // Tolerance for the number of partition id 0 in rowset, default 0 |
| DEFINE_Int32(ignore_invalid_partition_id_rowset_num, "0"); |
| |
| DEFINE_mInt32(report_query_statistics_interval_ms, "3000"); |
| // 30s |
| DEFINE_mInt32(query_statistics_reserve_timeout_ms, "30000"); |
| |
| DEFINE_mInt32(report_exec_status_thread_num, "5"); |
| |
| // consider two high usage disk at the same available level if they do not exceed this diff. |
| DEFINE_mDouble(high_disk_avail_level_diff_usages, "0.15"); |
| |
| // create tablet in partition random robin idx lru size, default 10000 |
| DEFINE_Int32(partition_disk_index_lru_size, "10000"); |
| // limit the storage space that query spill files can use |
| DEFINE_String(spill_storage_root_path, ""); |
| DEFINE_String(spill_storage_limit, "20%"); // 20% |
| DEFINE_mInt32(spill_gc_interval_ms, "2000"); // 2s |
| DEFINE_mInt32(spill_gc_work_time_ms, "2000"); // 2s |
| DEFINE_Int32(spill_io_thread_pool_thread_num, "-1"); |
| DEFINE_Validator(spill_io_thread_pool_thread_num, [](const int config) -> bool { |
| if (config == -1) { |
| CpuInfo::init(); |
| spill_io_thread_pool_thread_num = std::max(48, CpuInfo::num_cores() * 2); |
| } |
| return true; |
| }); |
| DEFINE_Int32(spill_io_thread_pool_queue_size, "102400"); |
| |
| // paused query in queue timeout(ms) will be resumed or canceled |
| DEFINE_Int64(spill_in_paused_queue_timeout_ms, "60000"); |
| |
| DEFINE_mBool(check_segment_when_build_rowset_meta, "false"); |
| |
| DEFINE_mBool(force_azure_blob_global_endpoint, "false"); |
| |
| DEFINE_mInt32(max_s3_client_retry, "10"); |
| DEFINE_mInt32(s3_read_base_wait_time_ms, "100"); |
| DEFINE_mInt32(s3_read_max_wait_time_ms, "800"); |
| DEFINE_mBool(enable_s3_object_check_after_upload, "true"); |
| |
| DEFINE_mBool(enable_s3_rate_limiter, "false"); |
| DEFINE_mInt64(s3_get_bucket_tokens, "1000000000000000000"); |
| DEFINE_Validator(s3_get_bucket_tokens, [](int64_t config) -> bool { return config > 0; }); |
| |
| DEFINE_mInt64(s3_get_token_per_second, "1000000000000000000"); |
| DEFINE_Validator(s3_get_token_per_second, [](int64_t config) -> bool { return config > 0; }); |
| |
| DEFINE_mInt64(s3_get_token_limit, "0"); |
| |
| DEFINE_mInt64(s3_put_bucket_tokens, "1000000000000000000"); |
| DEFINE_Validator(s3_put_bucket_tokens, [](int64_t config) -> bool { return config > 0; }); |
| |
| DEFINE_mInt64(s3_put_token_per_second, "1000000000000000000"); |
| DEFINE_Validator(s3_put_token_per_second, [](int64_t config) -> bool { return config > 0; }); |
| |
| DEFINE_mInt64(s3_put_token_limit, "0"); |
| |
| DEFINE_String(trino_connector_plugin_dir, "${DORIS_HOME}/plugins/connectors"); |
| |
| // ca_cert_file is in this path by default, Normally no modification is required |
| // ca cert default path is different from different OS |
| DEFINE_mString(ca_cert_file_paths, |
| "/etc/pki/tls/certs/ca-bundle.crt;/etc/ssl/certs/ca-certificates.crt;" |
| "/etc/ssl/ca-bundle.pem"); |
| |
| /** Table sink configurations(currently contains only external table types) **/ |
| // Minimum data processed to scale writers in exchange when non partition writing |
| DEFINE_mInt64(table_sink_non_partition_write_scaling_data_processed_threshold, |
| "26214400"); // 25MB |
| // Minimum data processed to trigger skewed partition rebalancing in exchange when partition writing |
| DEFINE_mInt64(table_sink_partition_write_min_data_processed_rebalance_threshold, |
| "26214400"); // 25MB |
| // Minimum partition data processed to rebalance writers in exchange when partition writing |
| DEFINE_mInt64(table_sink_partition_write_min_partition_data_processed_rebalance_threshold, |
| "15728640"); // 15MB |
| // Maximum processed partition nums of per writer when partition writing |
| DEFINE_mInt32(table_sink_partition_write_max_partition_nums_per_writer, "128"); |
| |
| /** Hive sink configurations **/ |
| DEFINE_mInt64(hive_sink_max_file_size, "1073741824"); // 1GB |
| |
| /** Iceberg sink configurations **/ |
| DEFINE_mInt64(iceberg_sink_max_file_size, "1073741824"); // 1GB |
| |
| DEFINE_mInt32(thrift_client_open_num_tries, "1"); |
| |
| DEFINE_Bool(enable_index_compaction, "false"); |
| |
| // http scheme in S3Client to use. E.g. http or https |
| DEFINE_String(s3_client_http_scheme, "http"); |
| DEFINE_Validator(s3_client_http_scheme, [](const std::string& config) -> bool { |
| return config == "http" || config == "https"; |
| }); |
| |
| DEFINE_mBool(ignore_schema_change_check, "false"); |
| |
| DEFINE_mInt64(string_overflow_size, "4294967295"); // std::numic_limits<uint32_t>::max() |
| |
| // The min thread num for BufferedReaderPrefetchThreadPool |
| DEFINE_Int64(num_buffered_reader_prefetch_thread_pool_min_thread, "16"); |
| // The max thread num for BufferedReaderPrefetchThreadPool |
| DEFINE_Int64(num_buffered_reader_prefetch_thread_pool_max_thread, "64"); |
| // The min thread num for S3FileUploadThreadPool |
| DEFINE_Int64(num_s3_file_upload_thread_pool_min_thread, "16"); |
| // The max thread num for S3FileUploadThreadPool |
| DEFINE_Int64(num_s3_file_upload_thread_pool_max_thread, "64"); |
| // The maximum jvm heap usage ratio for hdfs write workload |
| DEFINE_mDouble(max_hdfs_wirter_jni_heap_usage_ratio, "0.5"); |
| // The sleep milliseconds duration when hdfs write exceeds the maximum usage |
| DEFINE_mInt64(hdfs_jni_write_sleep_milliseconds, "300"); |
| // The max retry times when hdfs write failed |
| DEFINE_mInt64(hdfs_jni_write_max_retry_time, "3"); |
| |
| // The min thread num for NonBlockCloseThreadPool |
| DEFINE_Int64(min_nonblock_close_thread_num, "12"); |
| // The max thread num for NonBlockCloseThreadPool |
| DEFINE_Int64(max_nonblock_close_thread_num, "64"); |
| // The possibility that mem allocator throws an exception during memory allocation |
| // This config is for test usage, be careful when changing it. |
| DEFINE_mDouble(mem_alloc_fault_probability, "0.0"); |
| // The time out milliseconds for remote fetch schema RPC, default 60s |
| DEFINE_mInt64(fetch_remote_schema_rpc_timeout_ms, "60000"); |
| |
| DEFINE_Int64(s3_file_system_local_upload_buffer_size, "5242880"); |
| |
| //JVM monitoring enable. To prevent be from crashing due to jvm compatibility issues. The default setting is off. |
| DEFINE_Bool(enable_jvm_monitor, "false"); |
| |
| DEFINE_Int32(load_data_dirs_threads, "-1"); |
| |
| // Skip loading stale rowset meta when initializing `TabletMeta` from protobuf |
| DEFINE_mBool(skip_loading_stale_rowset_meta, "false"); |
| |
| DEFINE_Bool(enable_file_logger, "true"); |
| |
| // The minimum row group size when exporting Parquet files. default 128MB |
| DEFINE_Int64(min_row_group_size, "134217728"); |
| |
| DEFINE_mInt64(compaction_memory_bytes_limit, "1073741824"); |
| |
| DEFINE_mInt64(compaction_batch_size, "-1"); |
| |
| // If set to false, the parquet reader will not use page index to filter data. |
| // This is only for debug purpose, in case sometimes the page index |
| // filter wrong data. |
| DEFINE_mBool(enable_parquet_page_index, "false"); |
| |
| DEFINE_mBool(ignore_not_found_file_in_external_table, "true"); |
| |
| DEFINE_mBool(enable_hdfs_mem_limiter, "true"); |
| |
| DEFINE_mInt16(topn_agg_limit_multiplier, "2"); |
| |
| // Tablet meta size limit after serialization, 1.5GB |
| DEFINE_mInt64(tablet_meta_serialize_size_limit, "1610612736"); |
| // Protobuf supports a maximum of 2GB, so the size of the tablet meta after serialization must be less than 2GB |
| // 1717986918 = 2GB * 0.8 |
| DEFINE_Validator(tablet_meta_serialize_size_limit, |
| [](const int64_t config) -> bool { return config < 1717986918; }); |
| |
| DEFINE_mInt64(pipeline_task_leakage_detect_period_secs, "60"); |
| DEFINE_mInt32(snappy_compression_block_size, "262144"); |
| DEFINE_mInt32(lz4_compression_block_size, "262144"); |
| |
| DEFINE_mBool(enable_pipeline_task_leakage_detect, "false"); |
| |
| DEFINE_mInt32(check_score_rounds_num, "1000"); |
| |
| DEFINE_Int32(query_cache_size, "512"); |
| |
| DEFINE_mBool(enable_delete_bitmap_merge_on_compaction, "false"); |
| // Enable validation to check the correctness of table size. |
| DEFINE_Bool(enable_table_size_correctness_check, "false"); |
| DEFINE_Bool(force_regenerate_rowsetid_on_start_error, "false"); |
| DEFINE_mBool(enable_sleep_between_delete_cumu_compaction, "false"); |
| |
| DEFINE_mInt32(compaction_num_per_round, "4"); |
| |
| DEFINE_mInt32(check_tablet_delete_bitmap_interval_seconds, "300"); |
| DEFINE_mInt32(check_tablet_delete_bitmap_score_top_n, "10"); |
| DEFINE_mBool(enable_check_tablet_delete_bitmap_score, "true"); |
| DEFINE_mInt32(schema_dict_cache_capacity, "4096"); |
| |
| // whether to prune rows with delete sign = 1 in base compaction |
| // ATTN: this config is only for test |
| DEFINE_mBool(enable_prune_delete_sign_when_base_compaction, "true"); |
| |
| DEFINE_mBool(enable_mow_verbose_log, "false"); |
| |
| DEFINE_mInt32(tablet_sched_delay_time_ms, "5000"); |
| DEFINE_mInt32(load_trigger_compaction_version_percent, "66"); |
| DEFINE_mInt64(base_compaction_interval_seconds_since_last_operation, "86400"); |
| DEFINE_mBool(enable_compaction_pause_on_high_memory, "true"); |
| |
| DEFINE_mBool(enable_calc_delete_bitmap_between_segments_concurrently, "false"); |
| |
| DEFINE_mBool(enable_update_delete_bitmap_kv_check_core, "false"); |
| |
| // clang-format off |
| #ifdef BE_TEST |
| // test s3 |
| DEFINE_String(test_s3_resource, "resource"); |
| DEFINE_String(test_s3_ak, "ak"); |
| DEFINE_String(test_s3_sk, "sk"); |
| DEFINE_String(test_s3_endpoint, "endpoint"); |
| DEFINE_String(test_s3_region, "region"); |
| DEFINE_String(test_s3_bucket, "bucket"); |
| DEFINE_String(test_s3_prefix, "prefix"); |
| #endif |
| // clang-format on |
| |
| std::map<std::string, Register::Field>* Register::_s_field_map = nullptr; |
| std::map<std::string, std::function<bool()>>* RegisterConfValidator::_s_field_validator = nullptr; |
| std::map<std::string, std::string>* full_conf_map = nullptr; |
| |
| std::mutex custom_conf_lock; |
| |
| std::mutex mutable_string_config_lock; |
| |
| // trim string |
| std::string& trim(std::string& s) { |
| // rtrim |
| s.erase(std::find_if(s.rbegin(), s.rend(), [](unsigned char c) { return !std::isspace(c); }) |
| .base(), |
| s.end()); |
| // ltrim |
| s.erase(s.begin(), |
| std::find_if(s.begin(), s.end(), [](unsigned char c) { return !std::isspace(c); })); |
| return s; |
| } |
| |
| // split string by '=' |
| void splitkv(const std::string& s, std::string& k, std::string& v) { |
| const char sep = '='; |
| int start = 0; |
| int end = 0; |
| if ((end = s.find(sep, start)) != std::string::npos) { |
| k = s.substr(start, end - start); |
| v = s.substr(end + 1); |
| } else { |
| k = s; |
| v = ""; |
| } |
| } |
| |
| // replace env variables |
| bool replaceenv(std::string& s) { |
| std::size_t pos = 0; |
| std::size_t start = 0; |
| while ((start = s.find("${", pos)) != std::string::npos) { |
| std::size_t end = s.find('}', start + 2); |
| if (end == std::string::npos) { |
| return false; |
| } |
| std::string envkey = s.substr(start + 2, end - start - 2); |
| const char* envval = std::getenv(envkey.c_str()); |
| if (envval == nullptr) { |
| return false; |
| } |
| s.erase(start, end - start + 1); |
| s.insert(start, envval); |
| pos = start + strlen(envval); |
| } |
| return true; |
| } |
| |
| bool strtox(const std::string& valstr, bool& retval); |
| bool strtox(const std::string& valstr, int16_t& retval); |
| bool strtox(const std::string& valstr, int32_t& retval); |
| bool strtox(const std::string& valstr, int64_t& retval); |
| bool strtox(const std::string& valstr, double& retval); |
| bool strtox(const std::string& valstr, std::string& retval); |
| |
| template <typename T> |
| bool strtox(const std::string& valstr, std::vector<T>& retval) { |
| std::stringstream ss(valstr); |
| std::string item; |
| T t; |
| while (std::getline(ss, item, ',')) { |
| if (!strtox(trim(item), t)) { |
| return false; |
| } |
| retval.push_back(t); |
| } |
| return true; |
| } |
| |
| bool strtox(const std::string& valstr, bool& retval) { |
| if (valstr == "true") { |
| retval = true; |
| } else if (valstr == "false") { |
| retval = false; |
| } else { |
| return false; |
| } |
| return true; |
| } |
| |
| template <typename T> |
| bool strtointeger(const std::string& valstr, T& retval) { |
| if (valstr.length() == 0) { |
| return false; // empty-string is only allowed for string type. |
| } |
| char* end; |
| errno = 0; |
| const char* valcstr = valstr.c_str(); |
| int64_t ret64 = strtoll(valcstr, &end, 10); |
| if (errno || end != valcstr + strlen(valcstr)) { |
| return false; // bad parse |
| } |
| T tmp = retval; |
| retval = static_cast<T>(ret64); |
| if (retval != ret64) { |
| retval = tmp; |
| return false; |
| } |
| return true; |
| } |
| |
| bool strtox(const std::string& valstr, int16_t& retval) { |
| return strtointeger(valstr, retval); |
| } |
| |
| bool strtox(const std::string& valstr, int32_t& retval) { |
| return strtointeger(valstr, retval); |
| } |
| |
| bool strtox(const std::string& valstr, int64_t& retval) { |
| return strtointeger(valstr, retval); |
| } |
| |
| bool strtox(const std::string& valstr, double& retval) { |
| if (valstr.length() == 0) { |
| return false; // empty-string is only allowed for string type. |
| } |
| char* end = nullptr; |
| errno = 0; |
| const char* valcstr = valstr.c_str(); |
| retval = strtod(valcstr, &end); |
| if (errno || end != valcstr + strlen(valcstr)) { |
| return false; // bad parse |
| } |
| return true; |
| } |
| |
| bool strtox(const std::string& valstr, std::string& retval) { |
| retval = valstr; |
| return true; |
| } |
| |
| template <typename T> |
| bool convert(const std::string& value, T& retval) { |
| std::string valstr(value); |
| trim(valstr); |
| if (!replaceenv(valstr)) { |
| return false; |
| } |
| return strtox(valstr, retval); |
| } |
| |
| // load conf file |
| bool Properties::load(const char* conf_file, bool must_exist) { |
| // if conf_file is null, use the empty props |
| if (conf_file == nullptr) { |
| return true; |
| } |
| |
| // open the conf file |
| std::ifstream input(conf_file); |
| if (!input.is_open()) { |
| if (must_exist) { |
| std::cerr << "config::load() failed to open the file:" << conf_file << std::endl; |
| return false; |
| } |
| return true; |
| } |
| |
| // load properties |
| std::string line; |
| std::string key; |
| std::string value; |
| line.reserve(512); |
| while (input) { |
| // read one line at a time |
| std::getline(input, line); |
| |
| // remove left and right spaces |
| trim(line); |
| |
| // ignore comments |
| if (line.empty() || line[0] == '#') { |
| continue; |
| } |
| |
| // read key and value |
| splitkv(line, key, value); |
| trim(key); |
| trim(value); |
| |
| // insert into file_conf_map |
| file_conf_map[key] = value; |
| } |
| |
| // close the conf file |
| input.close(); |
| |
| return true; |
| } |
| |
| template <typename T> |
| bool Properties::get_or_default(const char* key, const char* defstr, T& retval, bool* is_retval_set, |
| std::string& rawval) const { |
| const auto& it = file_conf_map.find(std::string(key)); |
| std::string valstr; |
| if (it == file_conf_map.end()) { |
| if (defstr == nullptr) { |
| // Not found in conf map, and no default value need to be set, just return |
| *is_retval_set = false; |
| return true; |
| } else { |
| valstr = std::string(defstr); |
| } |
| } else { |
| valstr = it->second; |
| } |
| rawval = valstr; |
| *is_retval_set = true; |
| return convert(valstr, retval); |
| } |
| |
| void Properties::set(const std::string& key, const std::string& val) { |
| file_conf_map.emplace(key, val); |
| } |
| |
| void Properties::set_force(const std::string& key, const std::string& val) { |
| file_conf_map[key] = val; |
| } |
| |
| Status Properties::dump(const std::string& conffile) { |
| std::string conffile_tmp = conffile + ".tmp"; |
| io::FileWriterPtr file_writer; |
| RETURN_IF_ERROR(io::global_local_filesystem()->create_file(conffile_tmp, &file_writer)); |
| RETURN_IF_ERROR(file_writer->append("# THIS IS AN AUTO GENERATED CONFIG FILE.\n")); |
| RETURN_IF_ERROR(file_writer->append( |
| "# You can modify this file manually, and the configurations in this file\n")); |
| RETURN_IF_ERROR(file_writer->append("# will overwrite the configurations in be.conf\n\n")); |
| |
| for (auto const& iter : file_conf_map) { |
| RETURN_IF_ERROR(file_writer->append(iter.first)); |
| RETURN_IF_ERROR(file_writer->append(" = ")); |
| RETURN_IF_ERROR(file_writer->append(iter.second)); |
| RETURN_IF_ERROR(file_writer->append("\n")); |
| } |
| |
| RETURN_IF_ERROR(file_writer->close()); |
| |
| return io::global_local_filesystem()->rename(conffile_tmp, conffile); |
| } |
| |
| template <typename T> |
| std::ostream& operator<<(std::ostream& out, const std::vector<T>& v) { |
| size_t last = v.size() - 1; |
| for (size_t i = 0; i < v.size(); ++i) { |
| out << v[i]; |
| if (i != last) { |
| out << ", "; |
| } |
| } |
| return out; |
| } |
| |
| #define SET_FIELD(FIELD, TYPE, FILL_CONF_MAP, SET_TO_DEFAULT) \ |
| if (strcmp((FIELD).type, #TYPE) == 0) { \ |
| TYPE new_value = TYPE(); \ |
| bool is_newval_set = false; \ |
| std::string raw_value; \ |
| if (!props.get_or_default((FIELD).name, ((SET_TO_DEFAULT) ? (FIELD).defval : nullptr), \ |
| new_value, &is_newval_set, raw_value)) { \ |
| std::cerr << "config field error: " << (FIELD).name << " = \"" << raw_value << '"' \ |
| << std::endl; \ |
| return false; \ |
| } \ |
| if (!is_newval_set) { \ |
| continue; \ |
| } \ |
| TYPE& ref_conf_value = *reinterpret_cast<TYPE*>((FIELD).storage); \ |
| TYPE old_value = ref_conf_value; \ |
| ref_conf_value = new_value; \ |
| if (RegisterConfValidator::_s_field_validator != nullptr) { \ |
| auto validator = RegisterConfValidator::_s_field_validator->find((FIELD).name); \ |
| if (validator != RegisterConfValidator::_s_field_validator->end() && \ |
| !(validator->second)()) { \ |
| ref_conf_value = old_value; \ |
| std::cerr << "validate " << (FIELD).name << "=" << new_value << " failed" \ |
| << std::endl; \ |
| return false; \ |
| } \ |
| } \ |
| if (FILL_CONF_MAP) { \ |
| std::ostringstream oss; \ |
| oss << ref_conf_value; \ |
| (*full_conf_map)[(FIELD).name] = oss.str(); \ |
| } \ |
| continue; \ |
| } |
| |
| // init conf fields |
| bool init(const char* conf_file, bool fill_conf_map, bool must_exist, bool set_to_default) { |
| Properties props; |
| // load properties file |
| if (!props.load(conf_file, must_exist)) { |
| return false; |
| } |
| // fill full_conf_map ? |
| if (fill_conf_map && full_conf_map == nullptr) { |
| full_conf_map = new std::map<std::string, std::string>(); |
| } |
| |
| // set conf fields |
| for (const auto& it : *Register::_s_field_map) { |
| SET_FIELD(it.second, bool, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, int16_t, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, int32_t, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, int64_t, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, double, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::string, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::vector<bool>, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::vector<int16_t>, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::vector<int32_t>, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::vector<int64_t>, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::vector<double>, fill_conf_map, set_to_default); |
| SET_FIELD(it.second, std::vector<std::string>, fill_conf_map, set_to_default); |
| } |
| |
| if (config::is_cloud_mode()) { |
| auto st = config::set_config("enable_file_cache", "true", true, true); |
| LOG(INFO) << "set config enable_file_cache " |
| << "true" |
| << " " << st; |
| } |
| |
| return true; |
| } |
| |
| #define UPDATE_FIELD(FIELD, VALUE, TYPE, PERSIST) \ |
| if (strcmp((FIELD).type, #TYPE) == 0) { \ |
| TYPE new_value; \ |
| if (!convert((VALUE), new_value)) { \ |
| return Status::Error<ErrorCode::INVALID_ARGUMENT, false>("convert '{}' as {} failed", \ |
| VALUE, #TYPE); \ |
| } \ |
| TYPE& ref_conf_value = *reinterpret_cast<TYPE*>((FIELD).storage); \ |
| TYPE old_value = ref_conf_value; \ |
| if (RegisterConfValidator::_s_field_validator != nullptr) { \ |
| auto validator = RegisterConfValidator::_s_field_validator->find((FIELD).name); \ |
| if (validator != RegisterConfValidator::_s_field_validator->end() && \ |
| !(validator->second)()) { \ |
| ref_conf_value = old_value; \ |
| return Status::Error<ErrorCode::INVALID_ARGUMENT, false>("validate {}={} failed", \ |
| (FIELD).name, new_value); \ |
| } \ |
| } \ |
| ref_conf_value = new_value; \ |
| if (full_conf_map != nullptr) { \ |
| std::ostringstream oss; \ |
| oss << new_value; \ |
| (*full_conf_map)[(FIELD).name] = oss.str(); \ |
| } \ |
| if (PERSIST) { \ |
| RETURN_IF_ERROR(persist_config(std::string((FIELD).name), VALUE)); \ |
| } \ |
| update_config(std::string((FIELD).name), VALUE); \ |
| return Status::OK(); \ |
| } |
| |
| // write config to be_custom.conf |
| // the caller need to make sure that the given config is valid |
| Status persist_config(const std::string& field, const std::string& value) { |
| // lock to make sure only one thread can modify the be_custom.conf |
| std::lock_guard<std::mutex> l(custom_conf_lock); |
| |
| static const std::string conffile = config::custom_config_dir + "/be_custom.conf"; |
| |
| Properties tmp_props; |
| if (!tmp_props.load(conffile.c_str(), false)) { |
| LOG(WARNING) << "failed to load " << conffile; |
| return Status::InternalError("failed to load conf file: {}", conffile); |
| } |
| |
| tmp_props.set_force(field, value); |
| return tmp_props.dump(conffile); |
| } |
| |
| Status set_config(const std::string& field, const std::string& value, bool need_persist, |
| bool force) { |
| auto it = Register::_s_field_map->find(field); |
| if (it == Register::_s_field_map->end()) { |
| return Status::Error<ErrorCode::NOT_FOUND, false>("'{}' is not found", field); |
| } |
| |
| if (!force && !it->second.valmutable) { |
| return Status::Error<ErrorCode::NOT_IMPLEMENTED_ERROR, false>( |
| "'{}' is not support to modify", field); |
| } |
| |
| UPDATE_FIELD(it->second, value, bool, need_persist); |
| UPDATE_FIELD(it->second, value, int16_t, need_persist); |
| UPDATE_FIELD(it->second, value, int32_t, need_persist); |
| UPDATE_FIELD(it->second, value, int64_t, need_persist); |
| UPDATE_FIELD(it->second, value, double, need_persist); |
| { |
| // add lock to ensure thread safe |
| std::lock_guard<std::mutex> lock(mutable_string_config_lock); |
| UPDATE_FIELD(it->second, value, std::string, need_persist); |
| } |
| |
| // The other types are not thread safe to change dynamically. |
| return Status::Error<ErrorCode::NOT_IMPLEMENTED_ERROR, false>( |
| "'{}' is type of '{}' which is not support to modify", field, it->second.type); |
| } |
| |
| void update_config(const std::string& field, const std::string& value) { |
| if ("sys_log_level" == field) { |
| // update log level |
| update_logging(field, value); |
| } |
| } |
| |
| Status set_fuzzy_configs() { |
| std::unordered_map<std::string, std::string> fuzzy_field_and_value; |
| std::shared_ptr<std::mt19937_64> generator(new std::mt19937_64()); |
| generator->seed(std::random_device()()); |
| std::uniform_int_distribution<int64_t> distribution(0, 100); |
| |
| // if have set enable_fuzzy_mode=true in be.conf, will fuzzy those field and values |
| fuzzy_field_and_value["disable_storage_page_cache"] = |
| ((distribution(*generator) % 2) == 0) ? "true" : "false"; |
| fuzzy_field_and_value["enable_system_metrics"] = |
| ((distribution(*generator) % 2) == 0) ? "true" : "false"; |
| fuzzy_field_and_value["enable_set_in_bitmap_value"] = |
| ((distribution(*generator) % 2) == 0) ? "true" : "false"; |
| fuzzy_field_and_value["enable_shrink_memory"] = |
| ((distribution(*generator) % 2) == 0) ? "true" : "false"; |
| fuzzy_field_and_value["string_overflow_size"] = |
| ((distribution(*generator) % 2) == 0) ? "10" : "4294967295"; |
| |
| fmt::memory_buffer buf; |
| for (auto& it : fuzzy_field_and_value) { |
| const auto& field = it.first; |
| const auto& value = it.second; |
| RETURN_IF_ERROR(set_config(field, value, false, true)); |
| fmt::format_to(buf, "{}={}, ", field, value); |
| } |
| LOG(INFO) << fmt::format("FUZZY MODE IN BE: those variables have been changed: ({}).", |
| fmt::to_string(buf)); |
| return Status::OK(); |
| } |
| |
| std::mutex* get_mutable_string_config_lock() { |
| return &mutable_string_config_lock; |
| } |
| |
| std::vector<std::vector<std::string>> get_config_info() { |
| std::vector<std::vector<std::string>> configs; |
| std::lock_guard<std::mutex> lock(mutable_string_config_lock); |
| for (const auto& it : *full_conf_map) { |
| auto field_it = Register::_s_field_map->find(it.first); |
| if (field_it == Register::_s_field_map->end()) { |
| continue; |
| } |
| |
| std::vector<std::string> _config; |
| _config.push_back(it.first); |
| |
| std::string config_val = it.second; |
| // For compatibility, this PR #32933 change the log dir's config logic, |
| // and deprecate the `sys_log_dir` config. |
| if (it.first == "sys_log_dir" && config_val == "") { |
| config_val = fmt::format("{}/log", std::getenv("DORIS_HOME")); |
| } |
| |
| _config.emplace_back(field_it->second.type); |
| if (0 == strcmp(field_it->second.type, "bool")) { |
| _config.emplace_back(config_val == "1" ? "true" : "false"); |
| } else { |
| _config.push_back(config_val); |
| } |
| _config.emplace_back(field_it->second.valmutable ? "true" : "false"); |
| |
| configs.push_back(_config); |
| } |
| return configs; |
| } |
| |
| } // namespace doris::config |