| // Licensed to the Apache Software Foundation (ASF) under one |
| // or more contributor license agreements. See the NOTICE file |
| // distributed with this work for additional information |
| // regarding copyright ownership. The ASF licenses this file |
| // to you under the Apache License, Version 2.0 (the |
| // "License"); you may not use this file except in compliance |
| // with the License. You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, |
| // software distributed under the License is distributed on an |
| // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| // KIND, either express or implied. See the License for the |
| // specific language governing permissions and limitations |
| // under the License. |
| |
| #include "scanner_scheduler.h" |
| |
| #include <algorithm> |
| #include <cstdint> |
| #include <functional> |
| #include <list> |
| #include <memory> |
| #include <ostream> |
| #include <string> |
| #include <typeinfo> |
| #include <utility> |
| #include <vector> |
| |
| #include "common/compiler_util.h" // IWYU pragma: keep |
| #include "common/config.h" |
| #include "common/logging.h" |
| #include "olap/tablet.h" |
| #include "runtime/exec_env.h" |
| #include "runtime/runtime_state.h" |
| #include "runtime/thread_context.h" |
| #include "util/async_io.h" // IWYU pragma: keep |
| #include "util/blocking_queue.hpp" |
| #include "util/cpu_info.h" |
| #include "util/defer_op.h" |
| #include "util/doris_metrics.h" |
| #include "util/runtime_profile.h" |
| #include "util/thread.h" |
| #include "util/threadpool.h" |
| #include "util/work_thread_pool.hpp" |
| #include "vec/core/block.h" |
| #include "vec/exec/scan/new_olap_scanner.h" // IWYU pragma: keep |
| #include "vec/exec/scan/scanner_context.h" |
| #include "vec/exec/scan/vscan_node.h" |
| #include "vec/exec/scan/vscanner.h" |
| #include "vfile_scanner.h" |
| |
| namespace doris::vectorized { |
| |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(local_scan_thread_pool_queue_size, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(local_scan_thread_pool_thread_num, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(remote_scan_thread_pool_queue_size, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(remote_scan_thread_pool_thread_num, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(limited_scan_thread_pool_queue_size, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(limited_scan_thread_pool_thread_num, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(group_local_scan_thread_pool_queue_size, MetricUnit::NOUNIT); |
| DEFINE_GAUGE_METRIC_PROTOTYPE_2ARG(group_local_scan_thread_pool_thread_num, MetricUnit::NOUNIT); |
| |
| ScannerScheduler::ScannerScheduler() = default; |
| |
| ScannerScheduler::~ScannerScheduler() { |
| if (!_is_init) { |
| return; |
| } |
| |
| for (int i = 0; i < QUEUE_NUM; i++) { |
| delete _pending_queues[i]; |
| } |
| delete[] _pending_queues; |
| _deregister_metrics(); |
| } |
| |
| void ScannerScheduler::stop() { |
| if (!_is_init) { |
| return; |
| } |
| |
| for (int i = 0; i < QUEUE_NUM; i++) { |
| _pending_queues[i]->shutdown(); |
| } |
| |
| _is_closed = true; |
| |
| _scheduler_pool->shutdown(); |
| _local_scan_thread_pool->shutdown(); |
| _remote_scan_thread_pool->shutdown(); |
| _limited_scan_thread_pool->shutdown(); |
| |
| _scheduler_pool->wait(); |
| _local_scan_thread_pool->join(); |
| _remote_scan_thread_pool->join(); |
| _limited_scan_thread_pool->wait(); |
| |
| LOG(INFO) << "ScannerScheduler stopped"; |
| } |
| |
| Status ScannerScheduler::init(ExecEnv* env) { |
| // 1. scheduling thread pool and scheduling queues |
| static_cast<void>(ThreadPoolBuilder("SchedulingThreadPool") |
| .set_min_threads(QUEUE_NUM) |
| .set_max_threads(QUEUE_NUM) |
| .build(&_scheduler_pool)); |
| |
| _pending_queues = new BlockingQueue<std::shared_ptr<ScannerContext>>*[QUEUE_NUM]; |
| for (int i = 0; i < QUEUE_NUM; i++) { |
| _pending_queues[i] = new BlockingQueue<std::shared_ptr<ScannerContext>>(INT32_MAX); |
| static_cast<void>(_scheduler_pool->submit_func([this, i] { this->_schedule_thread(i); })); |
| } |
| |
| // 2. local scan thread pool |
| _local_scan_thread_pool = std::make_unique<PriorityThreadPool>( |
| config::doris_scanner_thread_pool_thread_num, |
| config::doris_scanner_thread_pool_queue_size, "local_scan"); |
| |
| // 3. remote scan thread pool |
| _remote_thread_pool_max_size = config::doris_max_remote_scanner_thread_pool_thread_num != -1 |
| ? config::doris_max_remote_scanner_thread_pool_thread_num |
| : std::max(512, CpuInfo::num_cores() * 10); |
| _remote_thread_pool_max_size = |
| std::max(_remote_thread_pool_max_size, config::doris_scanner_thread_pool_thread_num); |
| _remote_scan_thread_pool = std::make_unique<PriorityThreadPool>( |
| _remote_thread_pool_max_size, config::doris_remote_scanner_thread_pool_queue_size, |
| "RemoteScanThreadPool"); |
| |
| // 4. limited scan thread pool |
| static_cast<void>(ThreadPoolBuilder("LimitedScanThreadPool") |
| .set_min_threads(config::doris_scanner_thread_pool_thread_num) |
| .set_max_threads(config::doris_scanner_thread_pool_thread_num) |
| .set_max_queue_size(config::doris_scanner_thread_pool_queue_size) |
| .build(&_limited_scan_thread_pool)); |
| _register_metrics(); |
| _is_init = true; |
| return Status::OK(); |
| } |
| |
| Status ScannerScheduler::submit(std::shared_ptr<ScannerContext> ctx) { |
| if (ctx->done()) { |
| return Status::EndOfFile("ScannerContext is done"); |
| } |
| ctx->queue_idx = (_queue_idx++ % QUEUE_NUM); |
| if (!_pending_queues[ctx->queue_idx]->blocking_put(ctx)) { |
| return Status::InternalError("failed to submit scanner context to scheduler"); |
| } |
| return Status::OK(); |
| } |
| |
| std::unique_ptr<ThreadPoolToken> ScannerScheduler::new_limited_scan_pool_token( |
| ThreadPool::ExecutionMode mode, int max_concurrency) { |
| return _limited_scan_thread_pool->new_token(mode, max_concurrency); |
| } |
| |
| void ScannerScheduler::_schedule_thread(int queue_id) { |
| BlockingQueue<std::shared_ptr<ScannerContext>>* queue = _pending_queues[queue_id]; |
| while (!_is_closed) { |
| std::shared_ptr<ScannerContext> ctx; |
| bool ok = queue->blocking_get(&ctx); |
| if (!ok) { |
| // maybe closed |
| continue; |
| } |
| |
| _schedule_scanners(ctx); |
| // If ctx is done, no need to schedule it again. |
| // But should notice that there may still scanners running in scanner pool. |
| } |
| } |
| |
| void ScannerScheduler::_schedule_scanners(std::shared_ptr<ScannerContext> ctx) { |
| auto task_lock = ctx->task_exec_ctx(); |
| if (task_lock == nullptr) { |
| LOG(INFO) << "could not lock task execution context, query " << ctx->debug_string() |
| << " maybe finished"; |
| return; |
| } |
| MonotonicStopWatch watch; |
| watch.reset(); |
| watch.start(); |
| ctx->incr_num_ctx_scheduling(1); |
| |
| if (ctx->done()) { |
| return; |
| } |
| |
| std::list<std::weak_ptr<ScannerDelegate>> this_run; |
| ctx->get_next_batch_of_scanners(&this_run); |
| if (this_run.empty()) { |
| // There will be 2 cases when this_run is empty: |
| // 1. The blocks queue reaches limit. |
| // The consumer will continue scheduling the ctx. |
| // 2. All scanners are running. |
| // There running scanner will schedule the ctx after they are finished. |
| // So here we just return to stop scheduling ctx. |
| return; |
| } |
| |
| ctx->inc_num_running_scanners(this_run.size()); |
| |
| // Submit scanners to thread pool |
| // TODO(cmy): How to handle this "nice"? |
| int nice = 1; |
| auto iter = this_run.begin(); |
| if (ctx->thread_token != nullptr) { |
| // TODO llj tg how to treat this? |
| while (iter != this_run.end()) { |
| std::shared_ptr<ScannerDelegate> scanner_delegate = (*iter).lock(); |
| if (scanner_delegate == nullptr) { |
| // Has to ++, or there is a dead loop |
| iter++; |
| continue; |
| } |
| scanner_delegate->_scanner->start_wait_worker_timer(); |
| auto s = ctx->thread_token->submit_func([this, scanner_ref = *iter, ctx]() { |
| this->_scanner_scan(this, ctx, scanner_ref); |
| }); |
| if (s.ok()) { |
| iter++; |
| } else { |
| ctx->set_status_on_error(s); |
| break; |
| } |
| } |
| } else { |
| while (iter != this_run.end()) { |
| std::shared_ptr<ScannerDelegate> scanner_delegate = (*iter).lock(); |
| if (scanner_delegate == nullptr) { |
| // Has to ++, or there is a dead loop |
| iter++; |
| continue; |
| } |
| scanner_delegate->_scanner->start_wait_worker_timer(); |
| TabletStorageType type = scanner_delegate->_scanner->get_storage_type(); |
| bool ret = false; |
| if (type == TabletStorageType::STORAGE_TYPE_LOCAL) { |
| if (auto* scan_sche = ctx->get_simple_scan_scheduler()) { |
| auto work_func = [this, scanner_ref = *iter, ctx]() { |
| this->_scanner_scan(this, ctx, scanner_ref); |
| }; |
| SimplifiedScanTask simple_scan_task = {work_func, ctx}; |
| ret = scan_sche->get_scan_queue()->try_put(simple_scan_task); |
| } else { |
| PriorityThreadPool::Task task; |
| task.work_function = [this, scanner_ref = *iter, ctx]() { |
| this->_scanner_scan(this, ctx, scanner_ref); |
| }; |
| task.priority = nice; |
| ret = _local_scan_thread_pool->offer(task); |
| } |
| } else { |
| PriorityThreadPool::Task task; |
| task.work_function = [this, scanner_ref = *iter, ctx]() { |
| this->_scanner_scan(this, ctx, scanner_ref); |
| }; |
| task.priority = nice; |
| ret = _remote_scan_thread_pool->offer(task); |
| } |
| if (ret) { |
| iter++; |
| } else { |
| ctx->set_status_on_error( |
| Status::InternalError("failed to submit scanner to scanner pool")); |
| break; |
| } |
| } |
| } |
| ctx->incr_ctx_scheduling_time(watch.elapsed_time()); |
| } |
| |
| void ScannerScheduler::_scanner_scan(ScannerScheduler* scheduler, |
| std::shared_ptr<ScannerContext> ctx, |
| std::weak_ptr<ScannerDelegate> scanner_ref) { |
| auto task_lock = ctx->task_exec_ctx(); |
| if (task_lock == nullptr) { |
| // LOG(WARNING) << "could not lock task execution context, query " << print_id(_query_id) |
| // << " maybe finished"; |
| return; |
| } |
| //LOG_EVERY_N(INFO, 100) << "start running scanner from ctx " << ctx->debug_string(); |
| // will release scanner if it is the last one, task lock is hold here, to ensure |
| // that scanner could call scannode's method during deconstructor |
| std::shared_ptr<ScannerDelegate> scanner_delegate = scanner_ref.lock(); |
| auto& scanner = scanner_delegate->_scanner; |
| if (scanner_delegate == nullptr) { |
| return; |
| } |
| SCOPED_ATTACH_TASK(scanner->runtime_state()); |
| // for cpu hard limit, thread name should not be reset |
| if (ctx->_should_reset_thread_name) { |
| Thread::set_self_name("_scanner_scan"); |
| } |
| |
| #ifndef __APPLE__ |
| // The configuration item is used to lower the priority of the scanner thread, |
| // typically employed to ensure CPU scheduling for write operations. |
| if (config::scan_thread_nice_value != 0 && scanner->get_name() != VFileScanner::NAME) { |
| Thread::set_thread_nice_value(); |
| } |
| #endif |
| scanner->update_wait_worker_timer(); |
| scanner->start_scan_cpu_timer(); |
| Status status = Status::OK(); |
| bool eos = false; |
| RuntimeState* state = ctx->state(); |
| DCHECK(nullptr != state); |
| if (!scanner->is_init()) { |
| status = scanner->init(); |
| if (!status.ok()) { |
| ctx->set_status_on_error(status); |
| eos = true; |
| } |
| } |
| if (!eos && !scanner->is_open()) { |
| status = scanner->open(state); |
| if (!status.ok()) { |
| ctx->set_status_on_error(status); |
| eos = true; |
| } |
| scanner->set_opened(); |
| } |
| |
| static_cast<void>(scanner->try_append_late_arrival_runtime_filter()); |
| |
| // Because we use thread pool to scan data from storage. One scanner can't |
| // use this thread too long, this can starve other query's scanner. So, we |
| // need yield this thread when we do enough work. However, OlapStorage read |
| // data in pre-aggregate mode, then we can't use storage returned data to |
| // judge if we need to yield. So we record all raw data read in this round |
| // scan, if this exceeds row number or bytes threshold, we yield this thread. |
| std::vector<vectorized::BlockUPtr> blocks; |
| int64_t raw_bytes_read = 0; |
| int64_t raw_bytes_threshold = config::doris_scanner_row_bytes; |
| int num_rows_in_block = 0; |
| |
| // Only set to true when ctx->done() return true. |
| // Use this flag because we need distinguish eos from `should_stop`. |
| // If eos is true, we still need to return blocks, |
| // but is should_stop is true, no need to return blocks |
| bool should_stop = false; |
| // Has to wait at least one full block, or it will cause a lot of schedule task in priority |
| // queue, it will affect query latency and query concurrency for example ssb 3.3. |
| auto should_do_scan = [&, batch_size = state->batch_size(), |
| time = state->wait_full_block_schedule_times()]() { |
| if (raw_bytes_read < raw_bytes_threshold) { |
| return true; |
| } else if (num_rows_in_block < batch_size) { |
| return raw_bytes_read < raw_bytes_threshold * time; |
| } |
| return false; |
| }; |
| |
| while (!eos && should_do_scan()) { |
| // TODO llj task group should should_yield? |
| if (UNLIKELY(ctx->done())) { |
| // No need to set status on error here. |
| // Because done() maybe caused by "should_stop" |
| should_stop = true; |
| break; |
| } |
| |
| BlockUPtr block = ctx->get_free_block(); |
| |
| status = scanner->get_block_after_projects(state, block.get(), &eos); |
| // The VFileScanner for external table may try to open not exist files, |
| // Because FE file cache for external table may out of date. |
| // So, NOT_FOUND for VFileScanner is not a fail case. |
| // Will remove this after file reader refactor. |
| if (!status.ok() && (scanner->get_name() != doris::vectorized::VFileScanner::NAME || |
| (scanner->get_name() == doris::vectorized::VFileScanner::NAME && |
| !status.is<ErrorCode::NOT_FOUND>()))) { |
| LOG(WARNING) << "Scan thread read VScanner failed: " << status.to_string(); |
| break; |
| } |
| VLOG_ROW << "VScanNode input rows: " << block->rows() << ", eos: " << eos; |
| if (status.is<ErrorCode::NOT_FOUND>()) { |
| // The only case in this "if" branch is external table file delete and fe cache has not been updated yet. |
| // Set status to OK. |
| status = Status::OK(); |
| eos = true; |
| } |
| |
| raw_bytes_read += block->allocated_bytes(); |
| num_rows_in_block += block->rows(); |
| if (UNLIKELY(block->rows() == 0)) { |
| ctx->return_free_block(std::move(block)); |
| } else { |
| if (!blocks.empty() && blocks.back()->rows() + block->rows() <= state->batch_size()) { |
| vectorized::MutableBlock mutable_block(blocks.back().get()); |
| static_cast<void>(mutable_block.merge(*block)); |
| blocks.back().get()->set_columns(std::move(mutable_block.mutable_columns())); |
| ctx->return_free_block(std::move(block)); |
| } else { |
| blocks.push_back(std::move(block)); |
| } |
| } |
| } // end for while |
| |
| // if we failed, check status. |
| if (UNLIKELY(!status.ok())) { |
| // _transfer_done = true; |
| ctx->set_status_on_error(status); |
| eos = true; |
| blocks.clear(); |
| } else if (should_stop) { |
| // No need to return blocks because of should_stop, just delete them |
| blocks.clear(); |
| } else if (!blocks.empty()) { |
| ctx->append_blocks_to_queue(blocks); |
| } |
| |
| scanner->update_scan_cpu_timer(); |
| if (eos || should_stop) { |
| scanner->mark_to_need_to_close(); |
| } |
| ctx->push_back_scanner_and_reschedule(scanner_delegate); |
| } |
| |
| void ScannerScheduler::_register_metrics() { |
| REGISTER_HOOK_METRIC(local_scan_thread_pool_queue_size, |
| [this]() { return _local_scan_thread_pool->get_queue_size(); }); |
| REGISTER_HOOK_METRIC(local_scan_thread_pool_thread_num, |
| [this]() { return _local_scan_thread_pool->get_active_threads(); }); |
| REGISTER_HOOK_METRIC(remote_scan_thread_pool_queue_size, |
| [this]() { return _remote_scan_thread_pool->get_queue_size(); }); |
| REGISTER_HOOK_METRIC(remote_scan_thread_pool_thread_num, |
| [this]() { return _remote_scan_thread_pool->get_active_threads(); }); |
| REGISTER_HOOK_METRIC(limited_scan_thread_pool_queue_size, |
| [this]() { return _limited_scan_thread_pool->get_queue_size(); }); |
| REGISTER_HOOK_METRIC(limited_scan_thread_pool_thread_num, |
| [this]() { return _limited_scan_thread_pool->num_threads(); }); |
| } |
| |
| void ScannerScheduler::_deregister_metrics() { |
| DEREGISTER_HOOK_METRIC(local_scan_thread_pool_queue_size); |
| DEREGISTER_HOOK_METRIC(local_scan_thread_pool_thread_num); |
| DEREGISTER_HOOK_METRIC(remote_scan_thread_pool_queue_size); |
| DEREGISTER_HOOK_METRIC(remote_scan_thread_pool_thread_num); |
| DEREGISTER_HOOK_METRIC(limited_scan_thread_pool_queue_size); |
| DEREGISTER_HOOK_METRIC(limited_scan_thread_pool_thread_num); |
| DEREGISTER_HOOK_METRIC(group_local_scan_thread_pool_queue_size); |
| DEREGISTER_HOOK_METRIC(group_local_scan_thread_pool_thread_num); |
| } |
| } // namespace doris::vectorized |