blob: 85a44fbf6f7d634e94ca785103c63efd4bdabd62 [file] [log] [blame]
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
#include "cloud/cloud_internal_service.h"
#include <bthread/countdown_event.h>
#include <algorithm>
#include <thread>
#include "cloud/cloud_storage_engine.h"
#include "cloud/cloud_tablet.h"
#include "cloud/cloud_tablet_mgr.h"
#include "cloud/cloud_warm_up_manager.h"
#include "cloud/config.h"
#include "io/cache/block_file_cache.h"
#include "io/cache/block_file_cache_downloader.h"
#include "io/cache/block_file_cache_factory.h"
#include "runtime/thread_context.h"
#include "runtime/workload_management/io_throttle.h"
#include "util/async_io.h"
#include "util/debug_points.h"
namespace doris {
#include "common/compile_check_avoid_begin.h"
#include "common/compile_check_begin.h"
bvar::Adder<uint64_t> g_file_cache_get_by_peer_num("file_cache_get_by_peer_num");
bvar::Adder<uint64_t> g_file_cache_get_by_peer_blocks_num("file_cache_get_by_peer_blocks_num");
bvar::Adder<uint64_t> g_file_cache_get_by_peer_success_num("file_cache_get_by_peer_success_num");
bvar::Adder<uint64_t> g_file_cache_get_by_peer_failed_num("file_cache_get_by_peer_failed_num");
bvar::LatencyRecorder g_file_cache_get_by_peer_server_latency(
"file_cache_get_by_peer_server_latency");
bvar::LatencyRecorder g_file_cache_get_by_peer_read_cache_file_latency(
"file_cache_get_by_peer_read_cache_file_latency");
bvar::LatencyRecorder g_cloud_internal_service_get_file_cache_meta_by_tablet_id_latency(
"cloud_internal_service_get_file_cache_meta_by_tablet_id_latency");
CloudInternalServiceImpl::CloudInternalServiceImpl(CloudStorageEngine& engine, ExecEnv* exec_env)
: PInternalService(exec_env), _engine(engine) {}
CloudInternalServiceImpl::~CloudInternalServiceImpl() = default;
void CloudInternalServiceImpl::alter_vault_sync(google::protobuf::RpcController* controller,
const doris::PAlterVaultSyncRequest* request,
PAlterVaultSyncResponse* response,
google::protobuf::Closure* done) {
LOG(INFO) << "alter be to sync vault info from Meta Service";
// If the vaults containing hdfs vault then it would try to create hdfs connection using jni
// which would acuiqre one thread local jniEnv. But bthread context can't guarantee that the brpc
// worker thread wouldn't do bthread switch between worker threads.
bool ret = _heavy_work_pool.try_offer([this, done]() {
brpc::ClosureGuard closure_guard(done);
_engine.sync_storage_vault();
});
if (!ret) {
brpc::ClosureGuard closure_guard(done);
LOG(WARNING) << "fail to offer alter_vault_sync request to the work pool, pool="
<< _heavy_work_pool.get_info();
}
}
FileCacheType cache_type_to_pb(io::FileCacheType type) {
switch (type) {
case io::FileCacheType::TTL:
return FileCacheType::TTL;
case io::FileCacheType::INDEX:
return FileCacheType::INDEX;
case io::FileCacheType::NORMAL:
return FileCacheType::NORMAL;
default:
DCHECK(false);
}
return FileCacheType::NORMAL;
}
void CloudInternalServiceImpl::get_file_cache_meta_by_tablet_id(
google::protobuf::RpcController* controller [[maybe_unused]],
const PGetFileCacheMetaRequest* request, PGetFileCacheMetaResponse* response,
google::protobuf::Closure* done) {
brpc::ClosureGuard closure_guard(done);
if (!config::enable_file_cache) {
LOG_WARNING("try to access tablet file cache meta, but file cache not enabled");
return;
}
auto begin_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::steady_clock::now().time_since_epoch())
.count();
std::ostringstream tablet_ids_stream;
int count = 0;
for (const auto& tablet_id : request->tablet_ids()) {
tablet_ids_stream << tablet_id << ", ";
count++;
if (count >= 10) {
break;
}
}
LOG(INFO) << "warm up get meta from this be, tablets num=" << request->tablet_ids().size()
<< ", first 10 tablet_ids=[ " << tablet_ids_stream.str() << " ]";
for (const auto& tablet_id : request->tablet_ids()) {
auto res = _engine.tablet_mgr().get_tablet(tablet_id);
if (!res.has_value()) {
LOG(ERROR) << "failed to get tablet: " << tablet_id
<< " err msg: " << res.error().msg();
continue;
}
CloudTabletSPtr tablet = std::move(res.value());
auto st = tablet->sync_rowsets();
if (!st) {
// just log failed, try it best
LOG(WARNING) << "failed to sync rowsets: " << tablet_id
<< " err msg: " << st.to_string();
}
auto rowsets = tablet->get_snapshot_rowset();
auto add_meta = [&](PGetFileCacheMetaResponse* resp, int64_t tablet_id,
const std::string& rowset_id, int32_t segment_id,
const std::string& file_name,
const std::tuple<int64_t, int64_t, io::FileCacheType, int64_t>& tuple,
const RowsetSharedPtr& rowset, bool is_index) {
FileCacheBlockMeta* meta = resp->add_file_cache_block_metas();
meta->set_tablet_id(tablet_id);
meta->set_rowset_id(rowset_id);
meta->set_segment_id(segment_id);
meta->set_file_name(file_name);
if (!is_index) {
// .dat
meta->set_file_size(rowset->rowset_meta()->segment_file_size(segment_id));
meta->set_file_type(doris::FileType::SEGMENT_FILE);
} else {
// .idx
const auto& idx_file_info =
rowset->rowset_meta()->inverted_index_file_info(segment_id);
meta->set_file_size(idx_file_info.has_index_size() ? idx_file_info.index_size()
: -1);
meta->set_file_type(doris::FileType::INVERTED_INDEX_FILE);
}
meta->set_offset(std::get<0>(tuple));
meta->set_size(std::get<1>(tuple));
meta->set_cache_type(cache_type_to_pb(std::get<2>(tuple)));
meta->set_expiration_time(std::get<3>(tuple));
};
auto process_file_for_segment = [&](PGetFileCacheMetaResponse* resp,
const RowsetSharedPtr& rowset, int64_t tablet_id,
const std::string& rowset_id, int32_t segment_id,
bool is_inedex) {
const char* extension = is_inedex ? ".idx" : ".dat";
std::string file_name = fmt::format("{}_{}{}", rowset_id, segment_id, extension);
auto cache_key = io::BlockFileCache::hash(file_name);
auto* cache = io::FileCacheFactory::instance()->get_by_path(cache_key);
if (!cache) return;
auto segments_meta = cache->get_hot_blocks_meta(cache_key);
for (const auto& tuple : segments_meta) {
add_meta(resp, tablet_id, rowset_id, segment_id, file_name, tuple, rowset,
is_inedex);
}
};
for (const RowsetSharedPtr& rowset : rowsets) {
std::string rowset_id = rowset->rowset_id().to_string();
for (int32_t segment_id = 0; segment_id < rowset->num_segments(); ++segment_id) {
process_file_for_segment(response, rowset, tablet_id, rowset_id, segment_id, false);
process_file_for_segment(response, rowset, tablet_id, rowset_id, segment_id, true);
}
}
}
auto end_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::steady_clock::now().time_since_epoch())
.count();
g_cloud_internal_service_get_file_cache_meta_by_tablet_id_latency << (end_ts - begin_ts);
LOG(INFO) << "get file cache meta by tablet ids = [ " << tablet_ids_stream.str() << " ] took "
<< end_ts - begin_ts << " us";
VLOG_DEBUG << "get file cache meta by tablet id request=" << request->DebugString()
<< ", response=" << response->DebugString();
}
namespace {
// Helper functions for fetch_peer_data
Status handle_peer_file_range_request(const std::string& path, PFetchPeerDataResponse* response) {
// Read specific range [file_offset, file_offset+file_size) across cached blocks
auto datas = io::FileCacheFactory::instance()->get_cache_data_by_path(path);
for (auto& cb : datas) {
*(response->add_datas()) = std::move(cb);
}
return Status::OK();
}
void set_error_response(PFetchPeerDataResponse* response, const std::string& error_msg) {
response->mutable_status()->add_error_msgs(error_msg);
response->mutable_status()->set_status_code(TStatusCode::INTERNAL_ERROR);
}
Status read_file_block(const std::shared_ptr<io::FileBlock>& file_block, size_t file_size,
doris::CacheBlockPB* output) {
std::string data;
// ATTN: calculate the rightmost boundary value of the block, due to inaccurate current block meta information.
// see CachedRemoteFileReader::read_at_impl for more details.
// Ensure file_size >= file_block->offset() to avoid underflow
if (file_size < file_block->offset()) {
LOG(WARNING) << "file_size (" << file_size << ") < file_block->offset("
<< file_block->offset() << ")";
return Status::InternalError<false>("file_size less than block offset");
}
size_t read_size = std::min(static_cast<size_t>(file_size - file_block->offset()),
file_block->range().size());
data.resize(read_size);
auto begin_read_file_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::steady_clock::now().time_since_epoch())
.count();
SCOPED_ATTACH_TASK(ExecEnv::GetInstance()->s3_file_buffer_tracker());
Slice slice(data.data(), data.size());
Status read_st = file_block->read(slice, /*read_offset=*/0);
auto end_read_file_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::steady_clock::now().time_since_epoch())
.count();
g_file_cache_get_by_peer_read_cache_file_latency << (end_read_file_ts - begin_read_file_ts);
if (read_st.ok()) {
output->set_block_offset(static_cast<int64_t>(file_block->offset()));
output->set_block_size(static_cast<int64_t>(read_size));
output->set_data(std::move(data));
return Status::OK();
} else {
g_file_cache_get_by_peer_failed_num << 1;
LOG(WARNING) << "read cache block failed: " << read_st;
return read_st;
}
}
Status handle_peer_file_cache_block_request(const PFetchPeerDataRequest* request,
PFetchPeerDataResponse* response) {
const auto& path = request->path();
auto hash = io::BlockFileCache::hash(path);
auto* cache = io::FileCacheFactory::instance()->get_by_path(hash);
if (cache == nullptr) {
g_file_cache_get_by_peer_failed_num << 1;
set_error_response(response, "can't get file cache instance");
return Status::InternalError<false>("can't get file cache instance");
}
io::CacheContext ctx {};
io::ReadStatistics local_stats;
ctx.stats = &local_stats;
for (const auto& cb_req : request->cache_req()) {
size_t offset = static_cast<size_t>(std::max<int64_t>(0, cb_req.block_offset()));
size_t size = static_cast<size_t>(std::max<int64_t>(0, cb_req.block_size()));
auto holder = cache->get_or_set(hash, offset, size, ctx);
for (auto& fb : holder.file_blocks) {
if (fb->state() != io::FileBlock::State::DOWNLOADED) {
g_file_cache_get_by_peer_failed_num << 1;
LOG(WARNING) << "read cache block failed, state=" << fb->state();
set_error_response(response, "read cache file error");
return Status::InternalError<false>("cache block not downloaded");
}
g_file_cache_get_by_peer_blocks_num << 1;
doris::CacheBlockPB* out = response->add_datas();
Status read_status = read_file_block(fb, request->file_size(), out);
if (!read_status.ok()) {
set_error_response(response, "read cache file error");
return read_status;
}
}
}
return Status::OK();
}
} // namespace
void CloudInternalServiceImpl::fetch_peer_data(google::protobuf::RpcController* controller
[[maybe_unused]],
const PFetchPeerDataRequest* request,
PFetchPeerDataResponse* response,
google::protobuf::Closure* done) {
bool ret = _heavy_work_pool.try_offer([request, response, done]() {
brpc::ClosureGuard closure_guard(done);
g_file_cache_get_by_peer_num << 1;
if (!config::enable_file_cache) {
LOG_WARNING("try to access file cache data, but file cache not enabled");
return;
}
auto begin_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::steady_clock::now().time_since_epoch())
.count();
const auto type = request->type();
const auto& path = request->path();
response->mutable_status()->set_status_code(TStatusCode::OK);
Status status = Status::OK();
if (type == PFetchPeerDataRequest_Type_PEER_FILE_RANGE) {
status = handle_peer_file_range_request(path, response);
} else if (type == PFetchPeerDataRequest_Type_PEER_FILE_CACHE_BLOCK) {
status = handle_peer_file_cache_block_request(request, response);
}
if (!status.ok()) {
LOG(WARNING) << "fetch peer data failed: " << status.to_string();
set_error_response(response, status.to_string());
}
DBUG_EXECUTE_IF("CloudInternalServiceImpl::fetch_peer_data_slower", {
int st_us = dp->param<int>("sleep", 1000);
LOG_WARNING("CloudInternalServiceImpl::fetch_peer_data_slower").tag("sleep", st_us);
bthread_usleep(st_us);
});
auto end_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::steady_clock::now().time_since_epoch())
.count();
g_file_cache_get_by_peer_server_latency << (end_ts - begin_ts);
g_file_cache_get_by_peer_success_num << 1;
VLOG_DEBUG << "fetch cache request=" << request->DebugString()
<< ", response=" << response->DebugString();
});
if (!ret) {
brpc::ClosureGuard closure_guard(done);
LOG(WARNING) << "fail to offer fetch peer data request to the work pool, pool="
<< _heavy_work_pool.get_info();
}
}
#include "common/compile_check_end.h"
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_submitted_segment_num(
"file_cache_event_driven_warm_up_submitted_segment_num");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_finished_segment_num(
"file_cache_event_driven_warm_up_finished_segment_num");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_failed_segment_num(
"file_cache_event_driven_warm_up_failed_segment_num");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_submitted_segment_size(
"file_cache_event_driven_warm_up_submitted_segment_size");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_finished_segment_size(
"file_cache_event_driven_warm_up_finished_segment_size");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_failed_segment_size(
"file_cache_event_driven_warm_up_failed_segment_size");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_submitted_index_num(
"file_cache_event_driven_warm_up_submitted_index_num");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_finished_index_num(
"file_cache_event_driven_warm_up_finished_index_num");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_failed_index_num(
"file_cache_event_driven_warm_up_failed_index_num");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_submitted_index_size(
"file_cache_event_driven_warm_up_submitted_index_size");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_finished_index_size(
"file_cache_event_driven_warm_up_finished_index_size");
bvar::Adder<uint64_t> g_file_cache_event_driven_warm_up_failed_index_size(
"file_cache_event_driven_warm_up_failed_index_size");
bvar::Status<int64_t> g_file_cache_warm_up_rowset_last_handle_unix_ts(
"file_cache_warm_up_rowset_last_handle_unix_ts", 0);
bvar::Status<int64_t> g_file_cache_warm_up_rowset_last_finish_unix_ts(
"file_cache_warm_up_rowset_last_finish_unix_ts", 0);
bvar::LatencyRecorder g_file_cache_warm_up_rowset_latency("file_cache_warm_up_rowset_latency");
bvar::LatencyRecorder g_file_cache_warm_up_rowset_request_to_handle_latency(
"file_cache_warm_up_rowset_request_to_handle_latency");
bvar::LatencyRecorder g_file_cache_warm_up_rowset_handle_to_finish_latency(
"file_cache_warm_up_rowset_handle_to_finish_latency");
bvar::Adder<uint64_t> g_file_cache_warm_up_rowset_slow_count(
"file_cache_warm_up_rowset_slow_count");
bvar::Adder<uint64_t> g_file_cache_warm_up_rowset_request_to_handle_slow_count(
"file_cache_warm_up_rowset_request_to_handle_slow_count");
bvar::Adder<uint64_t> g_file_cache_warm_up_rowset_handle_to_finish_slow_count(
"file_cache_warm_up_rowset_handle_to_finish_slow_count");
bvar::Adder<uint64_t> g_file_cache_warm_up_rowset_wait_for_compaction_num(
"file_cache_warm_up_rowset_wait_for_compaction_num");
bvar::Adder<uint64_t> g_file_cache_warm_up_rowset_wait_for_compaction_timeout_num(
"file_cache_warm_up_rowset_wait_for_compaction_timeout_num");
void CloudInternalServiceImpl::warm_up_rowset(google::protobuf::RpcController* controller
[[maybe_unused]],
const PWarmUpRowsetRequest* request,
PWarmUpRowsetResponse* response,
google::protobuf::Closure* done) {
brpc::ClosureGuard closure_guard(done);
std::shared_ptr<bthread::CountdownEvent> wait = nullptr;
timespec due_time;
if (request->has_sync_wait_timeout_ms() && request->sync_wait_timeout_ms() > 0) {
g_file_cache_warm_up_rowset_wait_for_compaction_num << 1;
wait = std::make_shared<bthread::CountdownEvent>(0);
VLOG_DEBUG << "sync_wait_timeout: " << request->sync_wait_timeout_ms() << " ms";
due_time = butil::milliseconds_from_now(request->sync_wait_timeout_ms());
}
for (auto& rs_meta_pb : request->rowset_metas()) {
RowsetMeta rs_meta;
rs_meta.init_from_pb(rs_meta_pb);
auto storage_resource = rs_meta.remote_storage_resource();
if (!storage_resource) {
LOG(WARNING) << storage_resource.error();
continue;
}
int64_t tablet_id = rs_meta.tablet_id();
auto rowset_id = rs_meta.rowset_id();
bool local_only = !(request->has_skip_existence_check() && request->skip_existence_check());
auto res = _engine.tablet_mgr().get_tablet(tablet_id, /* warmup_data = */ false,
/* sync_delete_bitmap = */ true,
/* sync_stats = */ nullptr,
/* local_only = */ local_only);
if (!res.has_value()) {
LOG_WARNING("Warm up error ").tag("tablet_id", tablet_id).error(res.error());
if (res.error().msg().find("local_only=true") != std::string::npos) {
res.error().set_code(ErrorCode::TABLE_NOT_FOUND);
}
res.error().to_protobuf(response->mutable_status());
continue;
}
auto tablet = res.value();
auto tablet_meta = tablet->tablet_meta();
int64_t handle_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::system_clock::now().time_since_epoch())
.count();
g_file_cache_warm_up_rowset_last_handle_unix_ts.set_value(handle_ts);
int64_t request_ts = request->has_unix_ts_us() ? request->unix_ts_us() : 0;
g_file_cache_warm_up_rowset_request_to_handle_latency << (handle_ts - request_ts);
if (request_ts > 0 && handle_ts - request_ts > config::warm_up_rowset_slow_log_ms * 1000) {
g_file_cache_warm_up_rowset_request_to_handle_slow_count << 1;
LOG(INFO) << "warm up rowset (request to handle) took " << handle_ts - request_ts
<< " us, tablet_id: " << rs_meta.tablet_id()
<< ", rowset_id: " << rowset_id.to_string();
}
int64_t expiration_time = tablet_meta->ttl_seconds();
if (!tablet->add_rowset_warmup_state(rs_meta, WarmUpTriggerSource::EVENT_DRIVEN)) {
LOG(INFO) << "found duplicate warmup task for rowset " << rowset_id.to_string()
<< ", skip it";
continue;
}
for (int64_t segment_id = 0; segment_id < rs_meta.num_segments(); segment_id++) {
auto segment_size = rs_meta.segment_file_size(segment_id);
auto download_done = [=, version = rs_meta.version()](Status st) {
DBUG_EXECUTE_IF("CloudInternalServiceImpl::warm_up_rowset.download_segment", {
auto sleep_time = dp->param<int>("sleep", 3);
LOG_INFO("[verbose] block download for rowset={}, version={}, sleep={}",
rowset_id.to_string(), version.to_string(), sleep_time);
std::this_thread::sleep_for(std::chrono::seconds(sleep_time));
});
DBUG_EXECUTE_IF(
"CloudInternalServiceImpl::warm_up_rowset.download_segment.inject_error", {
st = Status::InternalError("injected error");
LOG_INFO("[verbose] inject error, tablet={}, rowset={}, st={}",
tablet_id, rowset_id.to_string(), st.to_string());
});
if (st.ok()) {
g_file_cache_event_driven_warm_up_finished_segment_num << 1;
g_file_cache_event_driven_warm_up_finished_segment_size << segment_size;
int64_t now_ts = std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::system_clock::now().time_since_epoch())
.count();
g_file_cache_warm_up_rowset_last_finish_unix_ts.set_value(now_ts);
g_file_cache_warm_up_rowset_latency << (now_ts - request_ts);
g_file_cache_warm_up_rowset_handle_to_finish_latency << (now_ts - handle_ts);
if (request_ts > 0 &&
now_ts - request_ts > config::warm_up_rowset_slow_log_ms * 1000) {
g_file_cache_warm_up_rowset_slow_count << 1;
LOG(INFO) << "warm up rowset took " << now_ts - request_ts
<< " us, tablet_id: " << tablet_id
<< ", rowset_id: " << rowset_id.to_string()
<< ", segment_id: " << segment_id;
}
if (now_ts - handle_ts > config::warm_up_rowset_slow_log_ms * 1000) {
g_file_cache_warm_up_rowset_handle_to_finish_slow_count << 1;
LOG(INFO) << "warm up rowset (handle to finish) took " << now_ts - handle_ts
<< " us, tablet_id: " << tablet_id
<< ", rowset_id: " << rowset_id.to_string()
<< ", segment_id: " << segment_id;
}
} else {
g_file_cache_event_driven_warm_up_failed_segment_num << 1;
g_file_cache_event_driven_warm_up_failed_segment_size << segment_size;
LOG(WARNING) << "download segment failed, tablet_id: " << tablet_id
<< " rowset_id: " << rowset_id.to_string() << ", error: " << st;
}
if (tablet->complete_rowset_segment_warmup(WarmUpTriggerSource::EVENT_DRIVEN,
rowset_id, st, 1, 0)
.trigger_source == WarmUpTriggerSource::EVENT_DRIVEN) {
VLOG_DEBUG << "warmup rowset " << version.to_string() << "("
<< rowset_id.to_string() << ") completed";
}
if (wait) {
wait->signal();
}
};
io::DownloadFileMeta download_meta {
.path = storage_resource.value()->remote_segment_path(rs_meta, segment_id),
.file_size = segment_size,
.offset = 0,
.download_size = segment_size,
.file_system = storage_resource.value()->fs,
.ctx = {.is_index_data = false,
.expiration_time = expiration_time,
.is_dryrun = config::enable_reader_dryrun_when_download_file_cache,
.is_warmup = true},
.download_done = std::move(download_done),
};
g_file_cache_event_driven_warm_up_submitted_segment_num << 1;
g_file_cache_event_driven_warm_up_submitted_segment_size << segment_size;
if (wait) {
wait->add_count();
}
_engine.file_cache_block_downloader().submit_download_task(download_meta);
auto download_inverted_index = [&, tablet](std::string index_path, uint64_t idx_size) {
auto storage_resource = rs_meta.remote_storage_resource();
auto download_done = [=, version = rs_meta.version()](Status st) {
DBUG_EXECUTE_IF(
"CloudInternalServiceImpl::warm_up_rowset.download_inverted_idx", {
auto sleep_time = dp->param<int>("sleep", 3);
LOG_INFO(
"[verbose] block download for rowset={}, inverted index "
"file={}, sleep={}",
rowset_id.to_string(), index_path, sleep_time);
std::this_thread::sleep_for(std::chrono::seconds(sleep_time));
});
if (st.ok()) {
g_file_cache_event_driven_warm_up_finished_index_num << 1;
g_file_cache_event_driven_warm_up_finished_index_size << idx_size;
int64_t now_ts =
std::chrono::duration_cast<std::chrono::microseconds>(
std::chrono::system_clock::now().time_since_epoch())
.count();
g_file_cache_warm_up_rowset_last_finish_unix_ts.set_value(now_ts);
g_file_cache_warm_up_rowset_latency << (now_ts - request_ts);
g_file_cache_warm_up_rowset_handle_to_finish_latency
<< (now_ts - handle_ts);
if (request_ts > 0 &&
now_ts - request_ts > config::warm_up_rowset_slow_log_ms * 1000) {
g_file_cache_warm_up_rowset_slow_count << 1;
LOG(INFO) << "warm up rowset took " << now_ts - request_ts
<< " us, tablet_id: " << tablet_id
<< ", rowset_id: " << rowset_id.to_string()
<< ", segment_id: " << segment_id;
}
if (now_ts - handle_ts > config::warm_up_rowset_slow_log_ms * 1000) {
g_file_cache_warm_up_rowset_handle_to_finish_slow_count << 1;
LOG(INFO) << "warm up rowset (handle to finish) took "
<< now_ts - handle_ts << " us, tablet_id: " << tablet_id
<< ", rowset_id: " << rowset_id.to_string()
<< ", segment_id: " << segment_id;
}
} else {
g_file_cache_event_driven_warm_up_failed_index_num << 1;
g_file_cache_event_driven_warm_up_failed_index_size << idx_size;
LOG(WARNING) << "download inverted index failed, tablet_id: " << tablet_id
<< " rowset_id: " << rowset_id << ", error: " << st;
}
if (tablet->complete_rowset_segment_warmup(WarmUpTriggerSource::EVENT_DRIVEN,
rowset_id, st, 0, 1)
.trigger_source == WarmUpTriggerSource::EVENT_DRIVEN) {
VLOG_DEBUG << "warmup rowset " << version.to_string() << "("
<< rowset_id.to_string() << ") completed";
}
if (wait) {
wait->signal();
}
};
io::DownloadFileMeta download_meta {
.path = io::Path(index_path),
.file_size = static_cast<int64_t>(idx_size),
.file_system = storage_resource.value()->fs,
.ctx = {.is_index_data = false, // DORIS-20877
.expiration_time = expiration_time,
.is_dryrun = config::enable_reader_dryrun_when_download_file_cache,
.is_warmup = true},
.download_done = std::move(download_done),
};
g_file_cache_event_driven_warm_up_submitted_index_num << 1;
g_file_cache_event_driven_warm_up_submitted_index_size << idx_size;
tablet->update_rowset_warmup_state_inverted_idx_num(
WarmUpTriggerSource::EVENT_DRIVEN, rowset_id, 1);
if (wait) {
wait->add_count();
}
_engine.file_cache_block_downloader().submit_download_task(download_meta);
};
// inverted index
auto schema_ptr = rs_meta.tablet_schema();
auto idx_version = schema_ptr->get_inverted_index_storage_format();
if (schema_ptr->has_inverted_index() || schema_ptr->has_ann_index()) {
if (idx_version == InvertedIndexStorageFormatPB::V1) {
auto&& inverted_index_info = rs_meta.inverted_index_file_info(segment_id);
std::unordered_map<int64_t, int64_t> index_size_map;
for (const auto& info : inverted_index_info.index_info()) {
if (info.index_file_size() != -1) {
index_size_map[info.index_id()] = info.index_file_size();
} else {
VLOG_DEBUG << "Invalid index_file_size for segment_id " << segment_id
<< ", index_id " << info.index_id();
}
}
for (const auto& index : schema_ptr->inverted_indexes()) {
auto idx_path = storage_resource.value()->remote_idx_v1_path(
rs_meta, segment_id, index->index_id(), index->get_index_suffix());
download_inverted_index(idx_path, index_size_map[index->index_id()]);
}
} else { // InvertedIndexStorageFormatPB::V2
auto&& inverted_index_info = rs_meta.inverted_index_file_info(segment_id);
int64_t idx_size = 0;
if (inverted_index_info.has_index_size()) {
idx_size = inverted_index_info.index_size();
} else {
VLOG_DEBUG << "index_size is not set for segment " << segment_id;
}
auto idx_path =
storage_resource.value()->remote_idx_v2_path(rs_meta, segment_id);
download_inverted_index(idx_path, idx_size);
}
}
}
}
if (wait && wait->timed_wait(due_time)) {
g_file_cache_warm_up_rowset_wait_for_compaction_timeout_num << 1;
LOG_WARNING("the time spent warming up {} rowsets exceeded {} ms",
request->rowset_metas().size(), request->sync_wait_timeout_ms());
}
}
bvar::Adder<uint64_t> g_file_cache_recycle_cache_finished_segment_num(
"file_cache_recycle_cache_finished_segment_num");
bvar::Adder<uint64_t> g_file_cache_recycle_cache_finished_index_num(
"file_cache_recycle_cache_finished_index_num");
void CloudInternalServiceImpl::recycle_cache(google::protobuf::RpcController* controller
[[maybe_unused]],
const PRecycleCacheRequest* request,
PRecycleCacheResponse* response,
google::protobuf::Closure* done) {
brpc::ClosureGuard closure_guard(done);
if (!config::enable_file_cache) {
return;
}
for (const auto& meta : request->cache_metas()) {
for (int64_t segment_id = 0; segment_id < meta.num_segments(); segment_id++) {
auto file_key = Segment::file_cache_key(meta.rowset_id(), segment_id);
auto* file_cache = io::FileCacheFactory::instance()->get_by_path(file_key);
file_cache->remove_if_cached_async(file_key);
g_file_cache_recycle_cache_finished_segment_num << 1;
}
// inverted index
for (const auto& file_name : meta.index_file_names()) {
auto file_key = io::BlockFileCache::hash(file_name);
auto* file_cache = io::FileCacheFactory::instance()->get_by_path(file_key);
file_cache->remove_if_cached_async(file_key);
g_file_cache_recycle_cache_finished_index_num << 1;
}
}
}
#include "common/compile_check_avoid_end.h"
} // namespace doris