blob: 51e6f38fc1e109b580aecc01186daa300b5b3afa [file] [log] [blame]
// Licensed to the Apache Software Foundation (ASF) under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing,
// software distributed under the License is distributed on an
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
// KIND, either express or implied. See the License for the
// specific language governing permissions and limitations
// under the License.
#include "recycler/checker.h"
#include <aws/s3/S3Client.h>
#include <aws/s3/model/ListObjectsV2Request.h>
#include <butil/endpoint.h>
#include <butil/strings/string_split.h>
#include <fmt/core.h>
#include <gen_cpp/cloud.pb.h>
#include <gen_cpp/olap_file.pb.h>
#include <glog/logging.h>
#include <algorithm>
#include <chrono>
#include <climits>
#include <cstdint>
#include <functional>
#include <memory>
#include <mutex>
#include <numeric>
#include <sstream>
#include <string_view>
#include <unordered_map>
#include <unordered_set>
#include <vector>
#include "common/bvars.h"
#include "common/config.h"
#include "common/defer.h"
#include "common/encryption_util.h"
#include "common/logging.h"
#include "common/util.h"
#include "cpp/sync_point.h"
#include "meta-service/meta_service.h"
#include "meta-service/meta_service_schema.h"
#include "meta-service/meta_service_tablet_stats.h"
#include "meta-store/blob_message.h"
#include "meta-store/keys.h"
#include "meta-store/txn_kv.h"
#include "snapshot/snapshot_manager.h"
#ifdef ENABLE_HDFS_STORAGE_VAULT
#include "recycler/hdfs_accessor.h"
#endif
#include "recycler/s3_accessor.h"
#include "recycler/storage_vault_accessor.h"
#ifdef UNIT_TEST
#include "../test/mock_accessor.h"
#endif
#include "recycler/recycler.h"
#include "recycler/util.h"
namespace doris::cloud {
namespace config {
extern int32_t brpc_listen_port;
extern int32_t scan_instances_interval_seconds;
extern int32_t recycle_job_lease_expired_ms;
extern int32_t recycle_concurrency;
extern std::vector<std::string> recycle_whitelist;
extern std::vector<std::string> recycle_blacklist;
extern bool enable_inverted_check;
} // namespace config
using namespace std::chrono;
Checker::Checker(std::shared_ptr<TxnKv> txn_kv) : txn_kv_(std::move(txn_kv)) {
ip_port_ = std::string(butil::my_ip_cstr()) + ":" + std::to_string(config::brpc_listen_port);
}
Checker::~Checker() {
if (!stopped()) {
stop();
}
}
int Checker::start() {
DCHECK(txn_kv_);
instance_filter_.reset(config::recycle_whitelist, config::recycle_blacklist);
// launch instance scanner
auto scanner_func = [this]() {
std::this_thread::sleep_for(
std::chrono::seconds(config::recycler_sleep_before_scheduling_seconds));
while (!stopped()) {
std::vector<InstanceInfoPB> instances;
get_all_instances(txn_kv_.get(), instances);
LOG(INFO) << "Checker get instances: " << [&instances] {
std::stringstream ss;
for (auto& i : instances) ss << ' ' << i.instance_id();
return ss.str();
}();
if (!instances.empty()) {
// enqueue instances
std::lock_guard lock(mtx_);
for (auto& instance : instances) {
if (instance_filter_.filter_out(instance.instance_id())) continue;
if (instance.status() == InstanceInfoPB::DELETED) continue;
using namespace std::chrono;
auto enqueue_time_s =
duration_cast<seconds>(system_clock::now().time_since_epoch()).count();
auto [_, success] =
pending_instance_map_.insert({instance.instance_id(), enqueue_time_s});
// skip instance already in pending queue
if (success) {
pending_instance_queue_.push_back(std::move(instance));
}
}
pending_instance_cond_.notify_all();
}
{
std::unique_lock lock(mtx_);
notifier_.wait_for(lock,
std::chrono::seconds(config::scan_instances_interval_seconds),
[&]() { return stopped(); });
}
}
};
workers_.emplace_back(scanner_func);
// Launch lease thread
workers_.emplace_back([this] { lease_check_jobs(); });
// Launch inspect thread
workers_.emplace_back([this] { inspect_instance_check_interval(); });
// launch check workers
auto checker_func = [this]() {
while (!stopped()) {
// fetch instance to check
InstanceInfoPB instance;
long enqueue_time_s = 0;
{
std::unique_lock lock(mtx_);
pending_instance_cond_.wait(lock, [&]() -> bool {
return !pending_instance_queue_.empty() || stopped();
});
if (stopped()) {
return;
}
instance = std::move(pending_instance_queue_.front());
pending_instance_queue_.pop_front();
enqueue_time_s = pending_instance_map_[instance.instance_id()];
pending_instance_map_.erase(instance.instance_id());
}
const auto& instance_id = instance.instance_id();
{
std::lock_guard lock(mtx_);
// skip instance in recycling
if (working_instance_map_.count(instance_id)) {
LOG(INFO) << "checker skip instance already working, instance_id="
<< instance_id;
continue;
}
}
auto checker = std::make_shared<InstanceChecker>(txn_kv_, instance.instance_id());
if (checker->init(instance) != 0) {
LOG(WARNING) << "failed to init instance checker, instance_id="
<< instance.instance_id();
continue;
}
std::string check_job_key;
job_check_key({instance.instance_id()}, &check_job_key);
LOG(INFO) << "checker picked instance, instance_id=" << instance.instance_id()
<< " enqueue_time_s=" << enqueue_time_s;
int ret = prepare_instance_recycle_job(txn_kv_.get(), check_job_key,
instance.instance_id(), ip_port_,
config::check_object_interval_seconds * 1000);
if (ret != 0) { // Prepare failed
LOG(WARNING) << "checker prepare job failed, instance_id=" << instance.instance_id()
<< " ret=" << ret;
continue;
} else {
std::lock_guard lock(mtx_);
working_instance_map_.emplace(instance_id, checker);
}
if (stopped()) return;
using namespace std::chrono;
auto ctime_ms =
duration_cast<milliseconds>(system_clock::now().time_since_epoch()).count();
g_bvar_checker_enqueue_cost_s.put(instance_id, ctime_ms / 1000 - enqueue_time_s);
bool success {true};
auto log_progress = [&](std::string_view stage) {
LOG(INFO) << "checker progress, instance_id=" << instance_id << " stage=" << stage;
};
log_progress("do_check");
if (int ret = checker->do_check(); ret != 0) {
success = false;
}
if (config::enable_inverted_check) {
log_progress("do_inverted_check");
if (int ret = checker->do_inverted_check(); ret != 0) {
success = false;
}
}
if (config::enable_delete_bitmap_inverted_check) {
log_progress("do_delete_bitmap_inverted_check");
if (int ret = checker->do_delete_bitmap_inverted_check(); ret != 0) {
success = false;
}
}
if (config::enable_mow_job_key_check) {
log_progress("do_mow_job_key_check");
if (int ret = checker->do_mow_job_key_check(); ret != 0) {
success = false;
}
}
if (config::enable_tablet_stats_key_check) {
log_progress("do_tablet_stats_key_check");
if (int ret = checker->do_tablet_stats_key_check(); ret != 0) {
success = false;
}
}
if (config::enable_restore_job_check) {
log_progress("do_restore_job_check");
if (int ret = checker->do_restore_job_check(); ret != 0) {
success = false;
}
}
if (config::enable_txn_key_check) {
log_progress("do_txn_key_check");
if (int ret = checker->do_txn_key_check(); ret != 0) {
success = false;
}
}
if (config::enable_meta_rowset_key_check) {
log_progress("do_meta_rowset_key_check");
if (int ret = checker->do_meta_rowset_key_check(); ret != 0) {
success = false;
}
}
if (config::enable_delete_bitmap_storage_optimize_v2_check) {
log_progress("do_delete_bitmap_storage_optimize_check_v2");
if (int ret = checker->do_delete_bitmap_storage_optimize_check(2 /*version*/);
ret != 0) {
success = false;
}
}
if (config::enable_version_key_check) {
log_progress("do_version_key_check");
if (int ret = checker->do_version_key_check(); ret != 0) {
success = false;
}
}
if (config::enable_snapshot_check) {
log_progress("do_snapshots_check");
if (int ret = checker->do_snapshots_check(); ret != 0) {
success = false;
}
}
if (config::enable_mvcc_meta_key_check) {
log_progress("do_mvcc_meta_key_check");
if (int ret = checker->do_mvcc_meta_key_check(); ret != 0) {
success = false;
}
}
if (config::enable_packed_file_check) {
log_progress("do_packed_file_check");
if (int ret = checker->do_packed_file_check(); ret != 0) {
success = false;
}
}
// If instance checker has been aborted, don't finish this job
if (!checker->stopped()) {
finish_instance_recycle_job(txn_kv_.get(), check_job_key, instance.instance_id(),
ip_port_, success, ctime_ms);
}
LOG(INFO) << "checker finished instance, instance_id=" << instance.instance_id()
<< " success=" << success;
{
std::lock_guard lock(mtx_);
working_instance_map_.erase(instance.instance_id());
}
}
};
int num_threads = config::recycle_concurrency; // FIXME: use a new config entry?
for (int i = 0; i < num_threads; ++i) {
workers_.emplace_back(checker_func);
}
return 0;
}
void Checker::stop() {
stopped_ = true;
notifier_.notify_all();
pending_instance_cond_.notify_all();
{
std::lock_guard lock(mtx_);
for (auto& [_, checker] : working_instance_map_) {
checker->stop();
}
}
for (auto& w : workers_) {
if (w.joinable()) w.join();
}
}
void Checker::lease_check_jobs() {
while (!stopped()) {
std::vector<std::string> instances;
instances.reserve(working_instance_map_.size());
{
std::lock_guard lock(mtx_);
for (auto& [id, _] : working_instance_map_) {
instances.push_back(id);
}
}
for (auto& i : instances) {
std::string check_job_key;
job_check_key({i}, &check_job_key);
int ret = lease_instance_recycle_job(txn_kv_.get(), check_job_key, i, ip_port_);
if (ret == 1) {
std::lock_guard lock(mtx_);
if (auto it = working_instance_map_.find(i); it != working_instance_map_.end()) {
it->second->stop();
}
}
}
{
std::unique_lock lock(mtx_);
notifier_.wait_for(lock,
std::chrono::milliseconds(config::recycle_job_lease_expired_ms / 3),
[&]() { return stopped(); });
}
}
}
#define LOG_CHECK_INTERVAL_ALARM LOG(WARNING) << "Err for check interval: "
void Checker::do_inspect(const InstanceInfoPB& instance) {
std::string check_job_key = job_check_key({instance.instance_id()});
std::unique_ptr<Transaction> txn;
std::string val;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG_CHECK_INTERVAL_ALARM << "failed to create txn";
return;
}
err = txn->get(check_job_key, &val);
if (err != TxnErrorCode::TXN_OK && err != TxnErrorCode::TXN_KEY_NOT_FOUND) {
LOG_CHECK_INTERVAL_ALARM << "failed to get kv, err=" << err
<< " key=" << hex(check_job_key);
return;
}
auto checker = InstanceChecker(txn_kv_, instance.instance_id());
if (checker.init(instance) != 0) {
LOG_CHECK_INTERVAL_ALARM << "failed to init instance checker, instance_id="
<< instance.instance_id();
return;
}
int64_t bucket_lifecycle_days = 0;
if (checker.get_bucket_lifecycle(&bucket_lifecycle_days) != 0) {
LOG_CHECK_INTERVAL_ALARM << "failed to get bucket lifecycle, instance_id="
<< instance.instance_id();
return;
}
DCHECK(bucket_lifecycle_days > 0);
if (bucket_lifecycle_days == INT64_MAX) {
// No s3 bucket (may all accessors are HdfsAccessor), skip inspect
return;
}
int64_t last_ctime_ms = -1;
auto job_status = JobRecyclePB::IDLE;
auto has_last_ctime = [&]() {
JobRecyclePB job_info;
if (!job_info.ParseFromString(val)) {
LOG_CHECK_INTERVAL_ALARM << "failed to parse JobRecyclePB, key=" << hex(check_job_key);
}
DCHECK(job_info.instance_id() == instance.instance_id());
if (!job_info.has_last_ctime_ms()) return false;
last_ctime_ms = job_info.last_ctime_ms();
job_status = job_info.status();
g_bvar_checker_last_success_time_ms.put(instance.instance_id(),
job_info.last_success_time_ms());
return true;
};
using namespace std::chrono;
auto now = duration_cast<milliseconds>(system_clock::now().time_since_epoch()).count();
if (err == TxnErrorCode::TXN_KEY_NOT_FOUND || !has_last_ctime()) {
// Use instance's ctime for instances that do not have job's last ctime
last_ctime_ms = instance.ctime();
}
DCHECK(now - last_ctime_ms >= 0);
int64_t expiration_ms =
bucket_lifecycle_days > config::reserved_buffer_days
? (bucket_lifecycle_days - config::reserved_buffer_days) * 86400000
: bucket_lifecycle_days * 86400000;
TEST_SYNC_POINT_CALLBACK("Checker:do_inspect", &last_ctime_ms);
if (now - last_ctime_ms >= expiration_ms) {
LOG_CHECK_INTERVAL_ALARM << "check risks, instance_id: " << instance.instance_id()
<< " last_ctime_ms: " << last_ctime_ms
<< " job_status: " << job_status
<< " bucket_lifecycle_days: " << bucket_lifecycle_days
<< " reserved_buffer_days: " << config::reserved_buffer_days
<< " expiration_ms: " << expiration_ms;
}
}
#undef LOG_CHECK_INTERVAL_ALARM
void Checker::inspect_instance_check_interval() {
while (!stopped()) {
LOG(INFO) << "start to inspect instance check interval";
std::vector<InstanceInfoPB> instances;
get_all_instances(txn_kv_.get(), instances);
for (const auto& instance : instances) {
if (instance_filter_.filter_out(instance.instance_id())) continue;
if (stopped()) return;
if (instance.status() == InstanceInfoPB::DELETED) continue;
do_inspect(instance);
}
{
std::unique_lock lock(mtx_);
notifier_.wait_for(lock, std::chrono::seconds(config::scan_instances_interval_seconds),
[&]() { return stopped(); });
}
}
}
// return 0 for success get a key, 1 for key not found, negative for error
int key_exist(TxnKv* txn_kv, std::string_view key) {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn, err=" << err;
return -1;
}
std::string val;
switch (txn->get(key, &val)) {
case TxnErrorCode::TXN_OK:
return 0;
case TxnErrorCode::TXN_KEY_NOT_FOUND:
return 1;
default:
return -1;
}
}
InstanceChecker::InstanceChecker(std::shared_ptr<TxnKv> txn_kv, const std::string& instance_id)
: txn_kv_(txn_kv), instance_id_(instance_id) {
snapshot_manager_ = std::make_shared<SnapshotManager>(txn_kv);
resource_mgr_ = std::make_shared<ResourceManager>(std::move(txn_kv));
resource_mgr_->init();
}
int InstanceChecker::init(const InstanceInfoPB& instance) {
int ret = init_obj_store_accessors(instance);
if (ret != 0) {
return ret;
}
return init_storage_vault_accessors(instance);
}
int InstanceChecker::init_obj_store_accessors(const InstanceInfoPB& instance) {
for (const auto& obj_info : instance.obj_info()) {
#ifdef UNIT_TEST
auto accessor = std::make_shared<MockAccessor>();
#else
auto s3_conf = S3Conf::from_obj_store_info(obj_info);
if (!s3_conf) {
LOG(WARNING) << "failed to init object accessor, instance_id=" << instance_id_;
return -1;
}
std::shared_ptr<S3Accessor> accessor;
int ret = S3Accessor::create(std::move(*s3_conf), &accessor);
if (ret != 0) {
LOG(WARNING) << "failed to init object accessor. instance_id=" << instance_id_
<< " resource_id=" << obj_info.id();
return ret;
}
#endif
accessor_map_.emplace(obj_info.id(), std::move(accessor));
}
return 0;
}
int InstanceChecker::init_storage_vault_accessors(const InstanceInfoPB& instance) {
if (instance.resource_ids().empty()) {
return 0;
}
FullRangeGetOptions opts(txn_kv_);
opts.prefetch = true;
auto it = txn_kv_->full_range_get(storage_vault_key({instance_id_, ""}),
storage_vault_key({instance_id_, "\xff"}), std::move(opts));
for (auto kv = it->next(); kv.has_value(); kv = it->next()) {
auto [k, v] = *kv;
StorageVaultPB vault;
if (!vault.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed storage vault, unable to deserialize key=" << hex(k);
return -1;
}
TEST_SYNC_POINT_CALLBACK("InstanceRecycler::init_storage_vault_accessors.mock_vault",
&accessor_map_, &vault);
if (vault.has_hdfs_info()) {
#ifdef ENABLE_HDFS_STORAGE_VAULT
auto accessor = std::make_shared<HdfsAccessor>(vault.hdfs_info());
int ret = accessor->init();
if (ret != 0) {
LOG(WARNING) << "failed to init hdfs accessor. instance_id=" << instance_id_
<< " resource_id=" << vault.id() << " name=" << vault.name();
return ret;
}
accessor_map_.emplace(vault.id(), std::move(accessor));
#else
LOG(ERROR) << "HDFS is disabled (via the ENABLE_HDFS_STORAGE_VAULT build option), "
<< "but HDFS storage vaults were detected";
#endif
} else if (vault.has_obj_info()) {
#ifdef UNIT_TEST
auto accessor = std::make_shared<MockAccessor>();
#else
auto s3_conf = S3Conf::from_obj_store_info(vault.obj_info());
if (!s3_conf) {
LOG(WARNING) << "failed to init object accessor, instance_id=" << instance_id_;
return -1;
}
std::shared_ptr<S3Accessor> accessor;
int ret = S3Accessor::create(std::move(*s3_conf), &accessor);
if (ret != 0) {
LOG(WARNING) << "failed to init s3 accessor. instance_id=" << instance_id_
<< " resource_id=" << vault.id() << " name=" << vault.name();
return ret;
}
#endif
accessor_map_.emplace(vault.id(), std::move(accessor));
}
}
if (!it->is_valid()) {
LOG_WARNING("failed to get storage vault kv");
return -1;
}
return 0;
}
int InstanceChecker::do_check() {
TEST_SYNC_POINT("InstanceChecker.do_check");
LOG(INFO) << "begin to check instance objects instance_id=" << instance_id_;
int check_ret = 0;
long num_scanned = 0;
long num_scanned_with_segment = 0;
long num_rowset_loss = 0;
long instance_volume = 0;
using namespace std::chrono;
auto start_time = steady_clock::now();
DORIS_CLOUD_DEFER {
auto cost = duration<float>(steady_clock::now() - start_time).count();
LOG(INFO) << "check instance objects finished, cost=" << cost
<< "s. instance_id=" << instance_id_ << " num_scanned=" << num_scanned
<< " num_scanned_with_segment=" << num_scanned_with_segment
<< " num_rowset_loss=" << num_rowset_loss
<< " instance_volume=" << instance_volume;
g_bvar_checker_num_scanned.put(instance_id_, num_scanned);
g_bvar_checker_num_scanned_with_segment.put(instance_id_, num_scanned_with_segment);
g_bvar_checker_num_check_failed.put(instance_id_, num_rowset_loss);
g_bvar_checker_check_cost_s.put(instance_id_, static_cast<long>(cost));
// FIXME(plat1ko): What if some list operation failed?
g_bvar_checker_instance_volume.put(instance_id_, instance_volume);
};
struct TabletFiles {
int64_t tablet_id {0};
std::unordered_set<std::string> files;
};
TabletFiles tablet_files_cache;
auto check_rowset_objects = [&, this](doris::RowsetMetaCloudPB& rs_meta, std::string_view key) {
if (rs_meta.num_segments() == 0) {
return;
}
bool data_loss = false;
bool segment_file_loss = false;
bool index_file_loss = false;
DORIS_CLOUD_DEFER {
if (data_loss) {
LOG(INFO) << "segment file is" << (segment_file_loss ? "" : " not") << " loss, "
<< "index file is" << (index_file_loss ? "" : " not") << " loss, "
<< "rowset.tablet_id = " << rs_meta.tablet_id();
num_rowset_loss++;
}
};
++num_scanned_with_segment;
if (tablet_files_cache.tablet_id != rs_meta.tablet_id()) {
long tablet_volume = 0;
// Clear cache
tablet_files_cache.tablet_id = 0;
tablet_files_cache.files.clear();
// Get all file paths under this tablet directory
auto find_it = accessor_map_.find(rs_meta.resource_id());
if (find_it == accessor_map_.end()) {
LOG_WARNING("resource id not found in accessor map")
.tag("resource_id", rs_meta.resource_id())
.tag("tablet_id", rs_meta.tablet_id())
.tag("rowset_id", rs_meta.rowset_id_v2());
check_ret = -1;
return;
}
std::unique_ptr<ListIterator> list_iter;
int ret = find_it->second->list_directory(tablet_path_prefix(rs_meta.tablet_id()),
&list_iter);
if (ret != 0) { // No need to log, because S3Accessor has logged this error
check_ret = -1;
return;
}
for (auto file = list_iter->next(); file.has_value(); file = list_iter->next()) {
tablet_files_cache.files.insert(std::move(file->path));
tablet_volume += file->size;
}
tablet_files_cache.tablet_id = rs_meta.tablet_id();
instance_volume += tablet_volume;
}
for (int i = 0; i < rs_meta.num_segments(); ++i) {
auto path = segment_path(rs_meta.tablet_id(), rs_meta.rowset_id_v2(), i);
// Skip check if segment is already packed into a larger file
const auto& index_map = rs_meta.packed_slice_locations();
if (index_map.find(path) != index_map.end()) {
continue;
}
if (tablet_files_cache.files.contains(path)) {
continue;
}
if (1 == key_exist(txn_kv_.get(), key)) {
// Rowset has been deleted instead of data loss
break;
}
data_loss = true;
segment_file_loss = true;
TEST_SYNC_POINT_CALLBACK("InstanceChecker.do_check1", &path);
LOG(WARNING) << "object not exist, path=" << path
<< ", rs_meta=" << rs_meta.ShortDebugString() << " key=" << hex(key);
}
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn, err=" << err;
check_ret = -1;
return;
}
TabletIndexPB tablet_index;
if (get_tablet_idx(txn_kv_.get(), instance_id_, rs_meta.tablet_id(), tablet_index) == -1) {
LOG(WARNING) << "failed to get tablet index, tablet_id= " << rs_meta.tablet_id();
check_ret = -1;
return;
}
auto tablet_schema_key =
meta_schema_key({instance_id_, tablet_index.index_id(), rs_meta.schema_version()});
ValueBuf tablet_schema_val;
err = cloud::blob_get(txn.get(), tablet_schema_key, &tablet_schema_val);
if (err != TxnErrorCode::TXN_OK) {
check_ret = -1;
LOG(WARNING) << "failed to get schema, err=" << err;
return;
}
auto* schema = rs_meta.mutable_tablet_schema();
if (!parse_schema_value(tablet_schema_val, schema)) {
LOG(WARNING) << "malformed schema value, key=" << hex(tablet_schema_key);
return;
}
std::vector<std::pair<int64_t, std::string>> index_ids;
for (const auto& i : rs_meta.tablet_schema().index()) {
if (i.has_index_type() && i.index_type() == IndexType::INVERTED) {
index_ids.emplace_back(i.index_id(), i.index_suffix_name());
}
}
if (!index_ids.empty()) {
const auto& index_map = rs_meta.packed_slice_locations();
for (int i = 0; i < rs_meta.num_segments(); ++i) {
std::vector<std::string> index_path_v;
if (rs_meta.tablet_schema().inverted_index_storage_format() ==
InvertedIndexStorageFormatPB::V1) {
for (const auto& index_id : index_ids) {
LOG(INFO) << "check inverted index, tablet_id=" << rs_meta.tablet_id()
<< " rowset_id=" << rs_meta.rowset_id_v2() << " segment_id=" << i
<< " index_id=" << index_id.first
<< " index_suffix_name=" << index_id.second;
index_path_v.emplace_back(
inverted_index_path_v1(rs_meta.tablet_id(), rs_meta.rowset_id_v2(),
i, index_id.first, index_id.second));
}
} else {
index_path_v.emplace_back(
inverted_index_path_v2(rs_meta.tablet_id(), rs_meta.rowset_id_v2(), i));
}
if (std::ranges::all_of(index_path_v, [&](const auto& idx_file_path) {
// Skip check if inverted index file is already packed into a larger file
if (index_map.find(idx_file_path) != index_map.end()) {
return true;
}
if (!tablet_files_cache.files.contains(idx_file_path)) {
LOG(INFO) << "loss index file: " << idx_file_path;
return false;
}
return true;
})) {
continue;
}
index_file_loss = true;
data_loss = true;
}
}
};
// scan visible rowsets
auto start_key = meta_rowset_key({instance_id_, 0, 0});
auto end_key = meta_rowset_key({instance_id_, INT64_MAX, 0});
std::unique_ptr<RangeGetIterator> it;
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn, err=" << err;
return -1;
}
err = txn->get(start_key, end_key, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "internal error, failed to get rowset meta, err=" << err;
return -1;
}
num_scanned += it->size();
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
if (!it->has_next()) {
start_key = k;
}
doris::RowsetMetaCloudPB rs_meta;
if (!rs_meta.ParseFromArray(v.data(), v.size())) {
++num_rowset_loss;
LOG(WARNING) << "malformed rowset meta. key=" << hex(k) << " val=" << hex(v);
continue;
}
check_rowset_objects(rs_meta, k);
}
start_key.push_back('\x00'); // Update to next smallest key for iteration
} while (it->more() && !stopped());
return num_rowset_loss > 0 ? 1 : check_ret;
}
int InstanceChecker::get_bucket_lifecycle(int64_t* lifecycle_days) {
// If there are multiple buckets, return the minimum lifecycle.
int64_t min_lifecycle_days = INT64_MAX;
int64_t tmp_liefcycle_days = 0;
for (const auto& [id, accessor] : accessor_map_) {
if (accessor->type() != AccessorType::S3) {
continue;
}
auto* s3_accessor = static_cast<S3Accessor*>(accessor.get());
if (s3_accessor->check_versioning() != 0) {
return -1;
}
if (s3_accessor->get_life_cycle(&tmp_liefcycle_days) != 0) {
return -1;
}
if (tmp_liefcycle_days < min_lifecycle_days) {
min_lifecycle_days = tmp_liefcycle_days;
}
}
*lifecycle_days = min_lifecycle_days;
return 0;
}
int InstanceChecker::do_inverted_check() {
if (accessor_map_.size() > 1) {
LOG(INFO) << "currently not support inverted check for multi accessor. instance_id="
<< instance_id_;
return 0;
}
LOG(INFO) << "begin to inverted check objects instance_id=" << instance_id_;
int check_ret = 0;
long num_scanned = 0;
long num_file_leak = 0;
using namespace std::chrono;
auto start_time = steady_clock::now();
DORIS_CLOUD_DEFER {
g_bvar_inverted_checker_num_scanned.put(instance_id_, num_scanned);
g_bvar_inverted_checker_num_check_failed.put(instance_id_, num_file_leak);
auto cost = duration<float>(steady_clock::now() - start_time).count();
LOG(INFO) << "inverted check instance objects finished, cost=" << cost
<< "s. instance_id=" << instance_id_ << " num_scanned=" << num_scanned
<< " num_file_leak=" << num_file_leak;
};
struct TabletRowsets {
int64_t tablet_id {0};
std::unordered_set<std::string> rowset_ids;
};
TabletRowsets tablet_rowsets_cache;
RowsetIndexesFormatV1 rowset_index_cache_v1;
RowsetIndexesFormatV2 rowset_index_cache_v2;
// Return 0 if check success, return 1 if file is garbage data, negative if error occurred
auto check_segment_file = [&](const std::string& obj_key) {
std::vector<std::string> str;
butil::SplitString(obj_key, '/', &str);
// data/{tablet_id}/{rowset_id}_{seg_num}.dat
if (str.size() < 3) {
// clang-format off
LOG(WARNING) << "split obj_key error, str.size() should be less than 3,"
<< " value = " << str.size();
// clang-format on
return -1;
}
int64_t tablet_id = atol(str[1].c_str());
if (tablet_id <= 0) {
LOG(WARNING) << "failed to parse tablet_id, key=" << obj_key;
return -1;
}
if (!str[2].ends_with(".dat")) {
// skip check not segment file
return 0;
}
std::string rowset_id;
if (auto pos = str.back().find('_'); pos != std::string::npos) {
rowset_id = str.back().substr(0, pos);
} else {
LOG(WARNING) << "failed to parse rowset_id, key=" << obj_key;
return -1;
}
if (tablet_rowsets_cache.tablet_id == tablet_id) {
if (tablet_rowsets_cache.rowset_ids.contains(rowset_id)) {
return 0;
} else {
LOG(WARNING) << "rowset not exists, key=" << obj_key;
return -1;
}
}
// Get all rowset id of this tablet
tablet_rowsets_cache.tablet_id = tablet_id;
tablet_rowsets_cache.rowset_ids.clear();
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_rowset_key({instance_id_, tablet_id, 0});
auto end = meta_rowset_key({instance_id_, tablet_id, INT64_MAX});
do {
TxnErrorCode err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next()) {
// recycle corresponding resources
auto [k, v] = it->next();
doris::RowsetMetaCloudPB rowset;
if (!rowset.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed rowset meta value, key=" << hex(k);
return -1;
}
tablet_rowsets_cache.rowset_ids.insert(rowset.rowset_id_v2());
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
break;
}
}
} while (it->more() && !stopped());
if (!tablet_rowsets_cache.rowset_ids.contains(rowset_id)) {
// Garbage data leak
LOG(WARNING) << "rowset should be recycled, key=" << obj_key;
return 1;
}
return 0;
};
auto check_inverted_index_file = [&](const std::string& obj_key) {
std::vector<std::string> str;
butil::SplitString(obj_key, '/', &str);
// format v1: data/{tablet_id}/{rowset_id}_{seg_num}_{idx_id}{idx_suffix}.idx
// format v2: data/{tablet_id}/{rowset_id}_{seg_num}.idx
if (str.size() < 3) {
// clang-format off
LOG(WARNING) << "split obj_key error, str.size() should be less than 3,"
<< " value = " << str.size();
// clang-format on
return -1;
}
int64_t tablet_id = atol(str[1].c_str());
if (tablet_id <= 0) {
LOG(WARNING) << "failed to parse tablet_id, key=" << obj_key;
return -1;
}
// v1: {rowset_id}_{seg_num}_{idx_id}{idx_suffix}.idx
// v2: {rowset_id}_{seg_num}.idx
std::string rowset_info = str.back();
if (!rowset_info.ends_with(".idx")) {
return 0; // Not an index file
}
InvertedIndexStorageFormatPB inverted_index_storage_format =
std::count(rowset_info.begin(), rowset_info.end(), '_') > 1
? InvertedIndexStorageFormatPB::V1
: InvertedIndexStorageFormatPB::V2;
size_t pos = rowset_info.find_last_of('_');
if (pos == std::string::npos || pos + 1 >= str.back().size() - 4) {
LOG(WARNING) << "Invalid index_id format, key=" << obj_key;
return -1;
}
if (inverted_index_storage_format == InvertedIndexStorageFormatPB::V1) {
return check_inverted_index_file_storage_format_v1(tablet_id, obj_key, rowset_info,
rowset_index_cache_v1);
} else {
return check_inverted_index_file_storage_format_v2(tablet_id, obj_key, rowset_info,
rowset_index_cache_v2);
}
};
// so we choose to skip here.
TEST_SYNC_POINT_RETURN_WITH_VALUE("InstanceChecker::do_inverted_check", (int)0);
for (auto& [_, accessor] : accessor_map_) {
std::unique_ptr<ListIterator> list_iter;
int ret = accessor->list_directory("data", &list_iter);
if (ret != 0) {
return -1;
}
for (auto file = list_iter->next(); file.has_value(); file = list_iter->next()) {
const auto& path = file->path;
if (path == "data/packed_file" || path.starts_with("data/packed_file/")) {
continue; // packed_file has dedicated check logic
}
++num_scanned;
int ret = check_segment_file(path);
if (ret != 0) {
LOG(WARNING) << "failed to check segment file, uri=" << accessor->uri()
<< " path=" << path;
if (ret == 1) {
++num_file_leak;
} else {
check_ret = -1;
}
}
ret = check_inverted_index_file(path);
if (ret != 0) {
LOG(WARNING) << "failed to check index file, uri=" << accessor->uri()
<< " path=" << path;
if (ret == 1) {
++num_file_leak;
} else {
check_ret = -1;
}
}
}
if (!list_iter->is_valid()) {
LOG(WARNING) << "failed to list data directory. uri=" << accessor->uri();
return -1;
}
}
return num_file_leak > 0 ? 1 : check_ret;
}
int InstanceChecker::traverse_mow_tablet(const std::function<int(int64_t, bool)>& check_func) {
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_rowset_key({instance_id_, 0, 0});
auto end = meta_rowset_key({instance_id_, std::numeric_limits<int64_t>::max(), 0});
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it, false, 1);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "meta" ${instance_id} "rowset" ${tablet_id} ${version} -> RowsetMetaCloudPB
auto tablet_id = std::get<int64_t>(std::get<0>(out[3]));
if (!it->has_next()) {
// Update to next smallest key for iteration
// scan for next tablet in this instance
begin = meta_rowset_key({instance_id_, tablet_id + 1, 0});
}
TabletMetaCloudPB tablet_meta;
int ret = get_tablet_meta(txn_kv_.get(), instance_id_, tablet_id, tablet_meta);
if (ret < 0) {
LOG(WARNING) << fmt::format(
"failed to get_tablet_meta in do_delete_bitmap_integrity_check(), "
"instance_id={}, tablet_id={}",
instance_id_, tablet_id);
return ret;
}
if (tablet_meta.enable_unique_key_merge_on_write()) {
// only check merge-on-write table
bool has_sequence_col = tablet_meta.schema().has_sequence_col_idx() &&
tablet_meta.schema().sequence_col_idx() != -1;
int ret = check_func(tablet_id, has_sequence_col);
if (ret < 0) {
// return immediately when encounter unexpected error,
// otherwise, we continue to check the next tablet
return ret;
}
}
}
} while (it->more() && !stopped());
return 0;
}
int InstanceChecker::traverse_rowset_delete_bitmaps(
int64_t tablet_id, std::string rowset_id,
const std::function<int(int64_t, std::string_view, int64_t, int64_t)>& callback) {
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_delete_bitmap_key({instance_id_, tablet_id, rowset_id, 0, 0});
auto end = meta_delete_bitmap_key({instance_id_, tablet_id, rowset_id,
std::numeric_limits<int64_t>::max(),
std::numeric_limits<int64_t>::max()});
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "meta" ${instance_id} "delete_bitmap" ${tablet_id} ${rowset_id} ${version} ${segment_id} -> roaringbitmap
auto version = std::get<std::int64_t>(std::get<0>(out[5]));
auto segment_id = std::get<std::int64_t>(std::get<0>(out[6]));
int ret = callback(tablet_id, rowset_id, version, segment_id);
if (ret != 0) {
return ret;
}
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
break;
}
}
} while (it->more() && !stopped());
return 0;
}
int InstanceChecker::collect_tablet_rowsets(
int64_t tablet_id, const std::function<void(const doris::RowsetMetaCloudPB&)>& collect_cb) {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_rowset_key({instance_id_, tablet_id, 0});
auto end = meta_rowset_key({instance_id_, tablet_id + 1, 0});
int64_t rowsets_num {0};
do {
TxnErrorCode err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
doris::RowsetMetaCloudPB rowset;
if (!rowset.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed rowset meta value, key=" << hex(k);
return -1;
}
++rowsets_num;
collect_cb(rowset);
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
break;
}
}
} while (it->more() && !stopped());
LOG(INFO) << fmt::format(
"[delete bitmap checker] successfully collect rowsets for instance_id={}, "
"tablet_id={}, rowsets_num={}",
instance_id_, tablet_id, rowsets_num);
return 0;
}
int InstanceChecker::do_delete_bitmap_inverted_check() {
LOG(INFO) << fmt::format(
"[delete bitmap checker] begin to do_delete_bitmap_inverted_check for instance_id={}",
instance_id_);
// number of delete bitmap keys being scanned
int64_t total_delete_bitmap_keys {0};
// number of delete bitmaps which belongs to non mow tablet
int64_t abnormal_delete_bitmaps {0};
// number of delete bitmaps which doesn't have corresponding rowset in MS
int64_t leaked_delete_bitmaps {0};
auto start_time = std::chrono::steady_clock::now();
DORIS_CLOUD_DEFER {
g_bvar_inverted_checker_leaked_delete_bitmaps.put(instance_id_, leaked_delete_bitmaps);
g_bvar_inverted_checker_abnormal_delete_bitmaps.put(instance_id_, abnormal_delete_bitmaps);
g_bvar_inverted_checker_delete_bitmaps_scanned.put(instance_id_, total_delete_bitmap_keys);
auto cost = std::chrono::duration_cast<std::chrono::milliseconds>(
std::chrono::steady_clock::now() - start_time)
.count();
if (leaked_delete_bitmaps > 0 || abnormal_delete_bitmaps > 0) {
LOG(WARNING) << fmt::format(
"[delete bitmap check fails] delete bitmap inverted check for instance_id={}, "
"cost={} ms, total_delete_bitmap_keys={}, leaked_delete_bitmaps={}, "
"abnormal_delete_bitmaps={}",
instance_id_, cost, total_delete_bitmap_keys, leaked_delete_bitmaps,
abnormal_delete_bitmaps);
} else {
LOG(INFO) << fmt::format(
"[delete bitmap checker] delete bitmap inverted check for instance_id={}, "
"passed. cost={} ms, total_delete_bitmap_keys={}",
instance_id_, cost, total_delete_bitmap_keys);
}
};
struct TabletsRowsetsCache {
int64_t tablet_id {-1};
bool enable_merge_on_write {false};
std::unordered_set<std::string> rowsets {};
std::unordered_set<std::string> pending_delete_bitmaps {};
} tablet_rowsets_cache {};
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_delete_bitmap_key({instance_id_, 0, "", 0, 0});
auto end =
meta_delete_bitmap_key({instance_id_, std::numeric_limits<int64_t>::max(), "", 0, 0});
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "meta" ${instance_id} "delete_bitmap" ${tablet_id} ${rowset_id} ${version} ${segment_id} -> roaringbitmap
auto tablet_id = std::get<int64_t>(std::get<0>(out[3]));
auto rowset_id = std::get<std::string>(std::get<0>(out[4]));
auto version = std::get<std::int64_t>(std::get<0>(out[5]));
auto segment_id = std::get<std::int64_t>(std::get<0>(out[6]));
++total_delete_bitmap_keys;
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
}
if (tablet_rowsets_cache.tablet_id == -1 ||
tablet_rowsets_cache.tablet_id != tablet_id) {
TabletMetaCloudPB tablet_meta;
int ret = get_tablet_meta(txn_kv_.get(), instance_id_, tablet_id, tablet_meta);
if (ret < 0) {
LOG(WARNING) << fmt::format(
"[delete bitmap checker] failed to get_tablet_meta in "
"do_delete_bitmap_inverted_check(), instance_id={}, tablet_id={}",
instance_id_, tablet_id);
return ret;
}
tablet_rowsets_cache.tablet_id = tablet_id;
tablet_rowsets_cache.enable_merge_on_write =
tablet_meta.enable_unique_key_merge_on_write();
tablet_rowsets_cache.rowsets.clear();
tablet_rowsets_cache.pending_delete_bitmaps.clear();
if (tablet_rowsets_cache.enable_merge_on_write) {
// only collect rowsets for merge-on-write tablet
auto collect_cb =
[&tablet_rowsets_cache](const doris::RowsetMetaCloudPB& rowset) {
tablet_rowsets_cache.rowsets.insert(rowset.rowset_id_v2());
};
ret = collect_tablet_rowsets(tablet_id, collect_cb);
if (ret < 0) {
return ret;
}
// get pending delete bitmaps
ret = get_pending_delete_bitmap_keys(
tablet_id, tablet_rowsets_cache.pending_delete_bitmaps);
if (ret < 0) {
return ret;
}
}
}
DCHECK_EQ(tablet_id, tablet_rowsets_cache.tablet_id);
if (!tablet_rowsets_cache.enable_merge_on_write) {
// clang-format off
TEST_SYNC_POINT_CALLBACK(
"InstanceChecker::do_delete_bitmap_inverted_check.get_abnormal_delete_bitmap",
&tablet_id, &rowset_id, &version, &segment_id);
// clang-format on
++abnormal_delete_bitmaps;
// log an error and continue to check the next delete bitmap
LOG(WARNING) << fmt::format(
"[delete bitmap check fails] find a delete bitmap belongs to tablet "
"which is not a merge-on-write table! instance_id={}, tablet_id={}, "
"version={}, segment_id={}",
instance_id_, tablet_id, version, segment_id);
continue;
}
if (!tablet_rowsets_cache.rowsets.contains(rowset_id) &&
!tablet_rowsets_cache.pending_delete_bitmaps.contains(std::string(k))) {
TEST_SYNC_POINT_CALLBACK(
"InstanceChecker::do_delete_bitmap_inverted_check.get_leaked_delete_bitmap",
&tablet_id, &rowset_id, &version, &segment_id);
++leaked_delete_bitmaps;
// log an error and continue to check the next delete bitmap
LOG(WARNING) << fmt::format(
"[delete bitmap check fails] can't find corresponding rowset for delete "
"bitmap instance_id={}, tablet_id={}, rowset_id={}, version={}, "
"segment_id={}",
instance_id_, tablet_id, rowset_id, version, segment_id);
}
}
} while (it->more() && !stopped());
return (leaked_delete_bitmaps > 0 || abnormal_delete_bitmaps > 0) ? 1 : 0;
}
int InstanceChecker::get_pending_delete_bitmap_keys(
int64_t tablet_id, std::unordered_set<std::string>& pending_delete_bitmaps) {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
std::string pending_key = meta_pending_delete_bitmap_key({instance_id_, tablet_id});
std::string pending_val;
err = txn->get(pending_key, &pending_val);
if (err != TxnErrorCode::TXN_OK && err != TxnErrorCode::TXN_KEY_NOT_FOUND) {
LOG(WARNING) << "failed to get pending delete bitmap kv, err=" << err;
return -1;
}
if (err == TxnErrorCode::TXN_OK) {
PendingDeleteBitmapPB pending_info;
if (!pending_info.ParseFromString(pending_val)) [[unlikely]] {
LOG(WARNING) << "failed to parse PendingDeleteBitmapPB, tablet=" << tablet_id;
return -1;
}
for (auto& delete_bitmap_key : pending_info.delete_bitmap_keys()) {
pending_delete_bitmaps.emplace(std::string(delete_bitmap_key));
}
}
return 0;
}
int InstanceChecker::check_inverted_index_file_storage_format_v1(
int64_t tablet_id, const std::string& file_path, const std::string& rowset_info,
RowsetIndexesFormatV1& rowset_index_cache_v1) {
// format v1: data/{tablet_id}/{rowset_id}_{seg_num}_{idx_id}{idx_suffix}.idx
std::string rowset_id;
int64_t segment_id;
std::string index_id_with_suffix_name;
// {rowset_id}_{seg_num}_{idx_id}{idx_suffix}.idx
std::vector<std::string> str;
butil::SplitString(rowset_info.substr(0, rowset_info.size() - 4), '_', &str);
if (str.size() < 3) {
LOG(WARNING) << "Split rowset info with '_' error, str size < 3, rowset_info = "
<< rowset_info;
return -1;
}
rowset_id = str[0];
segment_id = std::atoll(str[1].c_str());
index_id_with_suffix_name = str[2];
if (rowset_index_cache_v1.rowset_id == rowset_id) {
if (rowset_index_cache_v1.segment_ids.contains(segment_id)) {
if (auto it = rowset_index_cache_v1.index_ids.find(index_id_with_suffix_name);
it == rowset_index_cache_v1.index_ids.end()) {
// clang-format off
LOG(WARNING) << fmt::format("index_id with suffix name not found, rowset_info = {}, obj_key = {}", rowset_info, file_path);
// clang-format on
return -1;
}
} else {
// clang-format off
LOG(WARNING) << fmt::format("segment id not found, rowset_info = {}, obj_key = {}", rowset_info, file_path);
// clang-format on
return -1;
}
}
rowset_index_cache_v1.rowset_id = rowset_id;
rowset_index_cache_v1.segment_ids.clear();
rowset_index_cache_v1.index_ids.clear();
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_rowset_key({instance_id_, tablet_id, 0});
auto end = meta_rowset_key({instance_id_, tablet_id, INT64_MAX});
do {
TxnErrorCode err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next()) {
// recycle corresponding resources
auto [k, v] = it->next();
doris::RowsetMetaCloudPB rs_meta;
if (!rs_meta.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed rowset meta value, key=" << hex(k);
return -1;
}
TabletIndexPB tablet_index;
if (get_tablet_idx(txn_kv_.get(), instance_id_, rs_meta.tablet_id(), tablet_index) ==
-1) {
LOG(WARNING) << "failedt to get tablet index, tablet_id= " << rs_meta.tablet_id();
return -1;
}
auto tablet_schema_key = meta_schema_key(
{instance_id_, tablet_index.index_id(), rs_meta.schema_version()});
ValueBuf tablet_schema_val;
err = cloud::blob_get(txn.get(), tablet_schema_key, &tablet_schema_val);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get schema, err=" << err;
return -1;
}
auto* schema = rs_meta.mutable_tablet_schema();
if (!parse_schema_value(tablet_schema_val, schema)) {
LOG(WARNING) << "malformed schema value, key=" << hex(tablet_schema_key);
return -1;
}
for (size_t i = 0; i < rs_meta.num_segments(); i++) {
rowset_index_cache_v1.segment_ids.insert(i);
}
for (const auto& i : rs_meta.tablet_schema().index()) {
if (i.has_index_type() && i.index_type() == IndexType::INVERTED) {
LOG(INFO) << fmt::format(
"record index info, index_id: {}, index_suffix_name: {}", i.index_id(),
i.index_suffix_name());
rowset_index_cache_v1.index_ids.insert(
fmt::format("{}{}", i.index_id(), i.index_suffix_name()));
}
}
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
break;
}
}
} while (it->more() && !stopped());
if (!rowset_index_cache_v1.segment_ids.contains(segment_id)) {
// Garbage data leak
// clang-format off
LOG(WARNING) << "rowset_index_cache_v1.segment_ids don't contains segment_id, rowset should be recycled,"
<< " key = " << file_path
<< " segment_id = " << segment_id;
// clang-format on
return 1;
}
if (!rowset_index_cache_v1.index_ids.contains(index_id_with_suffix_name)) {
// Garbage data leak
// clang-format off
LOG(WARNING) << "rowset_index_cache_v1.index_ids don't contains index_id_with_suffix_name,"
<< " rowset with inde meta should be recycled, key=" << file_path
<< " index_id_with_suffix_name=" << index_id_with_suffix_name;
// clang-format on
return 1;
}
return 0;
}
int InstanceChecker::check_inverted_index_file_storage_format_v2(
int64_t tablet_id, const std::string& file_path, const std::string& rowset_info,
RowsetIndexesFormatV2& rowset_index_cache_v2) {
std::string rowset_id;
int64_t segment_id;
// {rowset_id}_{seg_num}.idx
std::vector<std::string> str;
butil::SplitString(rowset_info.substr(0, rowset_info.size() - 4), '_', &str);
if (str.size() < 2) {
// clang-format off
LOG(WARNING) << "Split rowset info with '_' error, str size < 2, rowset_info = " << rowset_info;
// clang-format on
return -1;
}
rowset_id = str[0];
segment_id = std::atoll(str[1].c_str());
if (rowset_index_cache_v2.rowset_id == rowset_id) {
if (!rowset_index_cache_v2.segment_ids.contains(segment_id)) {
// clang-format off
LOG(WARNING) << fmt::format("index file not found, rowset_info = {}, obj_key = {}", rowset_info, file_path);
// clang-format on
return -1;
}
}
rowset_index_cache_v2.rowset_id = rowset_id;
rowset_index_cache_v2.segment_ids.clear();
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_rowset_key({instance_id_, tablet_id, 0});
auto end = meta_rowset_key({instance_id_, tablet_id, INT64_MAX});
do {
TxnErrorCode err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get rowset kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next()) {
// recycle corresponding resources
auto [k, v] = it->next();
doris::RowsetMetaCloudPB rs_meta;
if (!rs_meta.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed rowset meta value, key=" << hex(k);
return -1;
}
for (size_t i = 0; i < rs_meta.num_segments(); i++) {
rowset_index_cache_v2.segment_ids.insert(i);
}
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
break;
}
}
} while (it->more() && !stopped());
if (!rowset_index_cache_v2.segment_ids.contains(segment_id)) {
// Garbage data leak
LOG(WARNING) << "rowset with index meta should be recycled, key=" << file_path;
return 1;
}
return 0;
}
int InstanceChecker::check_delete_bitmap_storage_optimize_v2(
int64_t tablet_id, bool has_sequence_col,
int64_t& rowsets_with_useless_delete_bitmap_version) {
// end_version: create_time
std::map<int64_t, int64_t> tablet_rowsets_map {};
// rowset_id: {start_version, end_version}
std::map<std::string, std::pair<int64_t, int64_t>> rowset_version_map;
// Get all visible rowsets of this tablet
auto collect_cb = [&](const doris::RowsetMetaCloudPB& rowset) {
if (rowset.start_version() == 0 && rowset.end_version() == 1) {
// ignore dummy rowset [0-1]
return;
}
tablet_rowsets_map[rowset.end_version()] = rowset.creation_time();
rowset_version_map[rowset.rowset_id_v2()] =
std::make_pair(rowset.start_version(), rowset.end_version());
};
if (int ret = collect_tablet_rowsets(tablet_id, collect_cb); ret != 0) {
return ret;
}
std::unordered_set<std::string> pending_delete_bitmaps;
if (auto ret = get_pending_delete_bitmap_keys(tablet_id, pending_delete_bitmaps); ret < 0) {
return ret;
}
std::unique_ptr<RangeGetIterator> it;
auto begin = meta_delete_bitmap_key({instance_id_, tablet_id, "", 0, 0});
auto end = meta_delete_bitmap_key({instance_id_, tablet_id + 1, "", 0, 0});
std::string last_rowset_id = "";
int64_t last_version = 0;
int64_t last_failed_version = 0;
std::vector<int64_t> failed_versions;
auto print_failed_versions = [&]() {
TEST_SYNC_POINT_CALLBACK(
"InstanceChecker::check_delete_bitmap_storage_optimize_v2.get_abnormal_"
"rowset",
&tablet_id, &last_rowset_id);
rowsets_with_useless_delete_bitmap_version++;
// some versions are continuous, such as [8, 9, 10, 11, 13, 17, 18]
// print as [8-11, 13, 17-18]
int64_t last_start_version = -1;
int64_t last_end_version = -1;
std::stringstream ss;
ss << "[";
for (int64_t version : failed_versions) {
if (last_start_version == -1) {
last_start_version = version;
last_end_version = version;
continue;
}
if (last_end_version + 1 == version) {
last_end_version = version;
} else {
if (last_start_version == last_end_version) {
ss << last_start_version << ", ";
} else {
ss << last_start_version << "-" << last_end_version << ", ";
}
last_start_version = version;
last_end_version = version;
}
}
if (last_start_version == last_end_version) {
ss << last_start_version;
} else {
ss << last_start_version << "-" << last_end_version;
}
ss << "]";
std::stringstream version_str;
auto it = rowset_version_map.find(last_rowset_id);
if (it != rowset_version_map.end()) {
version_str << "[" << it->second.first << "-" << it->second.second << "]";
}
LOG(WARNING) << fmt::format(
"[delete bitmap check fails] delete bitmap storage optimize v2 check fail "
"for instance_id={}, tablet_id={}, rowset_id={}, version={} found delete "
"bitmap with versions={}, size={}",
instance_id_, tablet_id, last_rowset_id, version_str.str(), ss.str(),
failed_versions.size());
};
using namespace std::chrono;
int64_t now = duration_cast<seconds>(system_clock::now().time_since_epoch()).count();
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get delete bitmap kv, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "meta" ${instance_id} "delete_bitmap" ${tablet_id} ${rowset_id} ${version} ${segment_id} -> roaringbitmap
auto rowset_id = std::get<std::string>(std::get<0>(out[4]));
auto version = std::get<std::int64_t>(std::get<0>(out[5]));
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
}
if (rowset_id == last_rowset_id && version == last_version) {
// skip the same rowset and version
continue;
}
if (rowset_id != last_rowset_id && !failed_versions.empty()) {
print_failed_versions();
last_failed_version = 0;
failed_versions.clear();
}
last_rowset_id = rowset_id;
last_version = version;
if (tablet_rowsets_map.find(version) != tablet_rowsets_map.end()) {
continue;
}
auto version_it = rowset_version_map.find(rowset_id);
if (version_it == rowset_version_map.end()) {
// checked in do_delete_bitmap_inverted_check
continue;
}
if (pending_delete_bitmaps.contains(std::string(k))) {
continue;
}
if (has_sequence_col && version >= version_it->second.first &&
version <= version_it->second.second) {
continue;
}
// there may be an interval in this situation:
// 1. finish compaction job; 2. checker; 3. finish agg and remove delete bitmap to ms
auto rowset_it = tablet_rowsets_map.upper_bound(version);
if (rowset_it == tablet_rowsets_map.end()) {
if (version != last_failed_version) {
failed_versions.push_back(version);
}
last_failed_version = version;
continue;
}
if (rowset_it->second + config::delete_bitmap_storage_optimize_v2_check_skip_seconds >=
now) {
continue;
}
if (version != last_failed_version) {
failed_versions.push_back(version);
}
last_failed_version = version;
}
} while (it->more() && !stopped());
if (!failed_versions.empty()) {
print_failed_versions();
}
LOG(INFO) << fmt::format(
"[delete bitmap checker] finish check delete bitmap storage optimize v2 for "
"instance_id={}, tablet_id={}, rowsets_num={}, "
"rowsets_with_useless_delete_bitmap_version={}",
instance_id_, tablet_id, tablet_rowsets_map.size(),
rowsets_with_useless_delete_bitmap_version);
return (rowsets_with_useless_delete_bitmap_version > 1 ? 1 : 0);
}
int InstanceChecker::do_delete_bitmap_storage_optimize_check(int version) {
if (version != 2) {
return -1;
}
int64_t total_tablets_num {0};
int64_t failed_tablets_num {0};
// for v2 check
int64_t max_rowsets_with_useless_delete_bitmap_version = 0;
int64_t tablet_id_with_max_rowsets_with_useless_delete_bitmap_version = 0;
// check that for every visible rowset, there exists at least delete one bitmap in MS
int ret = traverse_mow_tablet([&](int64_t tablet_id, bool has_sequence_col) {
++total_tablets_num;
int64_t rowsets_with_useless_delete_bitmap_version = 0;
int res = check_delete_bitmap_storage_optimize_v2(
tablet_id, has_sequence_col, rowsets_with_useless_delete_bitmap_version);
if (rowsets_with_useless_delete_bitmap_version >
max_rowsets_with_useless_delete_bitmap_version) {
max_rowsets_with_useless_delete_bitmap_version =
rowsets_with_useless_delete_bitmap_version;
tablet_id_with_max_rowsets_with_useless_delete_bitmap_version = tablet_id;
}
failed_tablets_num += (res != 0);
return res;
});
if (ret < 0) {
return ret;
}
g_bvar_max_rowsets_with_useless_delete_bitmap_version.put(
instance_id_, max_rowsets_with_useless_delete_bitmap_version);
std::stringstream ss;
ss << "[delete bitmap checker] check delete bitmap storage optimize v" << version
<< " for instance_id=" << instance_id_ << ", total_tablets_num=" << total_tablets_num
<< ", failed_tablets_num=" << failed_tablets_num
<< ". max_rowsets_with_useless_delete_bitmap_version="
<< max_rowsets_with_useless_delete_bitmap_version
<< ", tablet_id=" << tablet_id_with_max_rowsets_with_useless_delete_bitmap_version;
LOG(INFO) << ss.str();
return (failed_tablets_num > 0) ? 1 : 0;
}
int InstanceChecker::do_mow_job_key_check() {
std::unique_ptr<RangeGetIterator> it;
std::string begin = mow_tablet_job_key({instance_id_, 0, 0});
std::string end = mow_tablet_job_key({instance_id_, INT64_MAX, 0});
MowTabletJobPB mow_tablet_job;
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get mow tablet job key, err=" << err;
return -1;
}
int64_t now = duration_cast<std::chrono::seconds>(
std::chrono::system_clock::now().time_since_epoch())
.count();
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "meta" ${instance_id} "mow_tablet_job" ${table_id} ${initiator}
auto table_id = std::get<int64_t>(std::get<0>(out[3]));
auto initiator = std::get<int64_t>(std::get<0>(out[4]));
if (!mow_tablet_job.ParseFromArray(v.data(), v.size())) [[unlikely]] {
LOG(WARNING) << "failed to parse MowTabletJobPB";
return -1;
}
int64_t expiration = mow_tablet_job.expiration();
// check job key failed should meet both following two condition:
// 1. job key is expired
// 2. table lock key is not found or key is not expired
if (expiration < now - config::mow_job_key_check_expiration_diff_seconds) {
std::string lock_key =
meta_delete_bitmap_update_lock_key({instance_id_, table_id, -1});
std::string lock_val;
err = txn->get(lock_key, &lock_val);
std::string reason = "";
if (err == TxnErrorCode::TXN_KEY_NOT_FOUND) {
reason = "table lock key not found";
} else {
DeleteBitmapUpdateLockPB lock_info;
if (!lock_info.ParseFromString(lock_val)) [[unlikely]] {
LOG(WARNING) << "failed to parse DeleteBitmapUpdateLockPB";
return -1;
}
if (lock_info.expiration() > now || lock_info.lock_id() != -1) {
reason = "table lock is not expired,lock_id=" +
std::to_string(lock_info.lock_id());
}
}
if (reason != "") {
LOG(WARNING) << fmt::format(
"[compaction key check fails] mow job key check fail for "
"instance_id={}, table_id={}, initiator={}, expiration={}, now={}, "
"reason={}",
instance_id_, table_id, initiator, expiration, now, reason);
return -1;
}
}
}
begin = it->next_begin_key(); // Update to next smallest key for iteration
} while (it->more() && !stopped());
return 0;
}
int InstanceChecker::do_tablet_stats_key_check() {
int ret = 0;
int64_t nums_leak = 0;
int64_t nums_loss = 0;
int64_t nums_scanned = 0;
int64_t nums_abnormal = 0;
std::string begin = meta_tablet_key({instance_id_, 0, 0, 0, 0});
std::string end = meta_tablet_key({instance_id_, INT64_MAX, 0, 0, 0});
// inverted check tablet exists
LOG(INFO) << "begin inverted check stats_tablet_key";
ret = scan_and_handle_kv(begin, end, [&](std::string_view key, std::string_view value) {
int ret = check_stats_tablet_key_exists(key, value);
nums_scanned++;
if (ret == 1) {
nums_loss++;
}
return ret;
});
if (ret == -1) {
LOG(WARNING) << "failed to inverted check if stats tablet key exists";
return -1;
} else if (ret == 1) {
LOG(WARNING) << "stats_tablet_key loss, nums_scanned=" << nums_scanned
<< ", nums_loss=" << nums_loss;
return 1;
}
LOG(INFO) << "finish inverted check stats_tablet_key, nums_scanned=" << nums_scanned
<< ", nums_loss=" << nums_loss;
begin = stats_tablet_key({instance_id_, 0, 0, 0, 0});
end = stats_tablet_key({instance_id_, INT64_MAX, 0, 0, 0});
nums_scanned = 0;
// check tablet exists
LOG(INFO) << "begin check stats_tablet_key leaked";
ret = scan_and_handle_kv(begin, end, [&](std::string_view key, std::string_view value) {
int ret = check_stats_tablet_key_leaked(key, value);
nums_scanned++;
if (ret == 1) {
nums_leak++;
}
return ret;
});
if (ret == -1) {
LOG(WARNING) << "failed to check if stats tablet key exists";
return -1;
} else if (ret == 1) {
LOG(WARNING) << "stats_tablet_key leaked, nums_scanned=" << nums_scanned
<< ", nums_leak=" << nums_leak;
return 1;
}
LOG(INFO) << "finish check stats_tablet_key leaked, nums_scanned=" << nums_scanned
<< ", nums_leak=" << nums_leak;
begin = stats_tablet_key({instance_id_, 0, 0, 0, 0});
end = stats_tablet_key({instance_id_, INT64_MAX, 0, 0, 0});
nums_scanned = 0;
// check if key is normal
LOG(INFO) << "begin check stats_tablet_key abnormal";
ret = scan_and_handle_kv(begin, end, [&](std::string_view key, std::string_view value) {
int ret = check_stats_tablet_key(key, value);
nums_scanned++;
if (ret == 1) {
nums_abnormal++;
}
return ret;
});
if (ret == -1) {
LOG(WARNING) << "failed to check if stats tablet key exists";
return -1;
} else if (ret == 1) {
LOG(WARNING) << "stats_tablet_key abnormal, nums_scanned=" << nums_scanned
<< ", nums_abnormal=" << nums_abnormal;
return 1;
}
LOG(INFO) << "finish check stats_tablet_key, nums_scanned=" << nums_scanned
<< ", nums_abnormal=" << nums_abnormal;
return 0;
}
int InstanceChecker::check_stats_tablet_key_exists(std::string_view key, std::string_view value) {
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "meta" ${instance_id} "tablet" ${table_id} ${index_id} ${partition_id} ${tablet_id}
auto table_id = std::get<int64_t>(std::get<0>(out[3]));
auto index_id = std::get<int64_t>(std::get<0>(out[4]));
auto partition_id = std::get<int64_t>(std::get<0>(out[5]));
auto tablet_id = std::get<int64_t>(std::get<0>(out[6]));
std::string tablet_stats_key =
stats_tablet_key({instance_id_, table_id, index_id, partition_id, tablet_id});
int ret = key_exist(txn_kv_.get(), tablet_stats_key);
if (ret == 1) {
// clang-format off
LOG(WARNING) << "stats tablet key's tablet key loss,"
<< " stats tablet key=" << hex(tablet_stats_key)
<< " meta tablet key=" << hex(key);
// clang-format on
return 1;
} else if (ret == -1) {
LOG(WARNING) << "failed to check key exists, key=" << hex(tablet_stats_key);
return -1;
}
LOG(INFO) << "check stats_tablet_key_exists ok, key=" << hex(key);
return 0;
}
int InstanceChecker::check_stats_tablet_key_leaked(std::string_view key, std::string_view value) {
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "stats" ${instance_id} "tablet" ${table_id} ${index_id} ${partition_id} ${tablet_id}
auto table_id = std::get<int64_t>(std::get<0>(out[3]));
auto index_id = std::get<int64_t>(std::get<0>(out[4]));
auto partition_id = std::get<int64_t>(std::get<0>(out[5]));
auto tablet_id = std::get<int64_t>(std::get<0>(out[6]));
std::string tablet_key =
meta_tablet_key({instance_id_, table_id, index_id, partition_id, tablet_id});
int ret = key_exist(txn_kv_.get(), tablet_key);
if (ret == 1) {
// clang-format off
LOG(WARNING) << "stats tablet key's tablet key leak,"
<< " stats tablet key=" << hex(key)
<< " meta tablet key=" << hex(tablet_key);
// clang-format on
return 1;
} else if (ret == -1) {
LOG(WARNING) << "failed to check key exists, key=" << hex(tablet_key);
return -1;
}
LOG(INFO) << "check stats_tablet_key_leaked ok, key=" << hex(key);
return 0;
}
int InstanceChecker::check_stats_tablet_key(std::string_view key, std::string_view value) {
TabletStatsPB tablet_stats_pb;
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "stats" ${instance_id} "tablet" ${table_id} ${index_id} ${partition_id} ${tablet_id}
auto tablet_id = std::get<int64_t>(std::get<0>(out[6]));
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG_WARNING("failed to recycle tablet ")
.tag("tablet id", tablet_id)
.tag("instance_id", instance_id_)
.tag("reason", "failed to create txn");
return -1;
}
std::string tablet_idx_key = meta_tablet_idx_key({instance_id_, tablet_id});
std::string tablet_idx_val;
TabletIndexPB tablet_idx;
err = txn->get(tablet_idx_key, &tablet_idx_val);
if (err != TxnErrorCode::TXN_OK) {
// clang-format off
LOG(WARNING) << "failed to get tablet index key,"
<< " key=" << hex(tablet_idx_key)
<< " code=" << err;
// clang-format on
return -1;
}
tablet_idx.ParseFromString(tablet_idx_val);
MetaServiceCode code = MetaServiceCode::OK;
std::string msg;
internal_get_tablet_stats(code, msg, txn.get(), instance_id_, tablet_idx, tablet_stats_pb);
if (code != MetaServiceCode::OK) {
// clang-format off
LOG(WARNING) << "failed to get tablet stats,"
<< " code=" << code
<< " msg=" << msg;
// clang-format on
return -1;
}
GetRowsetResponse resp;
// get rowsets in tablet
internal_get_rowset(txn.get(), 0, std::numeric_limits<int64_t>::max() - 1, instance_id_,
tablet_id, code, msg, &resp);
if (code != MetaServiceCode::OK) {
LOG_WARNING("failed to get rowsets of tablet when check stats tablet key")
.tag("tablet id", tablet_id)
.tag("msg", msg)
.tag("code", code)
.tag("instance id", instance_id_);
return -1;
}
int64_t num_rows = 0;
int64_t num_rowsets = 0;
int64_t num_segments = 0;
int64_t total_data_size = 0;
for (const auto& rs_meta : resp.rowset_meta()) {
num_rows += rs_meta.num_rows();
num_rowsets++;
num_segments += rs_meta.num_segments();
total_data_size += rs_meta.total_disk_size();
}
int ret = 0;
if (tablet_stats_pb.data_size() != total_data_size) {
ret = 1;
// clang-format off
LOG(WARNING) << " tablet_stats_pb's data size is not same with all rowset total data size,"
<< " tablet_stats_pb's data size=" << tablet_stats_pb.data_size()
<< " all rowset total data size=" << total_data_size
<< " stats tablet meta=" << tablet_stats_pb.ShortDebugString();
// clang-format on
} else if (tablet_stats_pb.num_rows() != num_rows) {
ret = 1;
// clang-format off
LOG(WARNING) << " tablet_stats_pb's num_rows is not same with all rowset total num_rows,"
<< " tablet_stats_pb's num_rows=" << tablet_stats_pb.num_rows()
<< " all rowset total num_rows=" << num_rows
<< " stats tablet meta=" << tablet_stats_pb.ShortDebugString();
// clang-format on
} else if (tablet_stats_pb.num_rowsets() != num_rowsets) {
ret = 1;
// clang-format off
LOG(WARNING) << " tablet_stats_pb's num_rowsets is not same with all rowset nums,"
<< " tablet_stats_pb's num_rowsets=" << tablet_stats_pb.num_rowsets()
<< " all rowset nums=" << num_rowsets
<< " stats tablet meta=" << tablet_stats_pb.ShortDebugString();
// clang-format on
} else if (tablet_stats_pb.num_segments() != num_segments) {
ret = 1;
// clang-format off
LOG(WARNING) << " tablet_stats_pb's num_segments is not same with all rowset total num_segments,"
<< " tablet_stats_pb's num_segments=" << tablet_stats_pb.num_segments()
<< " all rowset total num_segments=" << num_segments
<< " stats tablet meta=" << tablet_stats_pb.ShortDebugString();
// clang-format on
}
return ret;
}
int InstanceChecker::scan_and_handle_kv(
std::string& start_key, const std::string& end_key,
std::function<int(std::string_view, std::string_view)> handle_kv) {
std::unique_ptr<Transaction> txn;
int ret = 0;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn";
return -1;
}
std::unique_ptr<RangeGetIterator> it;
int limit = 10000;
TEST_SYNC_POINT_CALLBACK("InstanceChecker:scan_and_handle_kv:limit", &limit);
do {
err = txn->get(start_key, end_key, &it, false, limit);
TEST_SYNC_POINT_CALLBACK("InstanceChecker:scan_and_handle_kv:get_err", &err);
if (err == TxnErrorCode::TXN_TOO_OLD) {
LOG(WARNING) << "failed to get range kv, err=txn too old, "
<< " now fallback to non snapshot scan";
err = txn_kv_->create_txn(&txn);
if (err == TxnErrorCode::TXN_OK) {
err = txn->get(start_key, end_key, &it);
}
}
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "internal error, failed to get range kv, err=" << err;
return -1;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
int handle_ret = handle_kv(k, v);
if (handle_ret == -1) {
return -1;
} else {
ret = std::max(ret, handle_ret);
}
if (!it->has_next()) {
start_key = k;
}
}
start_key = it->next_begin_key();
} while (it->more() && !stopped());
return ret;
}
int InstanceChecker::do_version_key_check() {
std::unique_ptr<RangeGetIterator> it;
std::string begin = table_version_key({instance_id_, 0, 0});
std::string end = table_version_key({instance_id_, INT64_MAX, 0});
bool check_res = true;
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get mow tablet job key, err=" << err;
return -1;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
int64_t table_version = -1;
// 0x01 "version" ${instance_id} "table" ${db_id} ${tbl_id}
if (!txn->decode_atomic_int(v, &table_version)) {
LOG(WARNING) << "malformed table version value";
return -1;
}
auto table_id = std::get<int64_t>(std::get<0>(out[4]));
auto db_id = std::get<int64_t>(std::get<0>(out[3]));
std::string partition_version_key_begin =
partition_version_key({instance_id_, db_id, table_id, 0});
std::string partition_version_key_end =
partition_version_key({instance_id_, db_id, table_id, INT64_MAX});
VersionPB partition_version_pb;
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(partition_version_key_begin, partition_version_key_end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get mow tablet job key, err=" << err;
return -1;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
// 0x01 "version" ${instance_id} "partition" ${db_id} ${tbl_id} ${partition_id}
std::string_view k1 = k;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
if (!partition_version_pb.ParseFromArray(v.data(), v.size())) [[unlikely]] {
LOG(WARNING) << "failed to parse partition VersionPB";
return -1;
}
auto partition_id = std::get<int64_t>(std::get<0>(out[5]));
int64_t partition_version = partition_version_pb.version();
if (table_version < partition_version) {
check_res = false;
LOG(WARNING)
<< "table version is less than partition version,"
<< " table_id: " << table_id << "tablet_version: " << table_version
<< " partition_id: " << partition_id
<< " partition_version: " << partition_version;
}
}
partition_version_key_begin = it->next_begin_key();
} while (it->more() && !stopped());
}
begin = it->next_begin_key(); // Update to next smallest key for iteration
} while (it->more() && !stopped());
return check_res ? 0 : -1;
}
int InstanceChecker::do_restore_job_check() {
int64_t num_prepared = 0;
int64_t num_committed = 0;
int64_t num_dropped = 0;
int64_t num_completed = 0;
int64_t num_recycling = 0;
int64_t num_cost_many_time = 0;
const int64_t COST_MANY_THRESHOLD = 3600;
using namespace std::chrono;
auto start_time = steady_clock::now();
DORIS_CLOUD_DEFER {
g_bvar_checker_restore_job_prepared_state.put(instance_id_, num_prepared);
g_bvar_checker_restore_job_committed_state.put(instance_id_, num_committed);
g_bvar_checker_restore_job_dropped_state.put(instance_id_, num_dropped);
g_bvar_checker_restore_job_completed_state.put(instance_id_, num_completed);
g_bvar_checker_restore_job_recycling_state.put(instance_id_, num_recycling);
g_bvar_checker_restore_job_cost_many_time.put(instance_id_, num_cost_many_time);
auto cost_ms =
duration_cast<std::chrono::milliseconds>(steady_clock::now() - start_time).count();
LOG(INFO) << "check instance restore jobs finished, cost=" << cost_ms
<< "ms. instance_id=" << instance_id_ << " num_prepared=" << num_prepared
<< " num_committed=" << num_committed << " num_dropped=" << num_dropped
<< " num_completed=" << num_completed << " num_recycling=" << num_recycling
<< " num_cost_many_time=" << num_cost_many_time;
};
LOG_INFO("begin to check restore jobs").tag("instance_id", instance_id_);
JobRestoreTabletKeyInfo restore_job_key_info0 {instance_id_, 0};
JobRestoreTabletKeyInfo restore_job_key_info1 {instance_id_, INT64_MAX};
std::string begin;
std::string end;
job_restore_tablet_key(restore_job_key_info0, &begin);
job_restore_tablet_key(restore_job_key_info1, &end);
std::unique_ptr<RangeGetIterator> it;
do {
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn";
return -1;
}
err = txn->get(begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get mow tablet job key, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
while (it->has_next()) {
auto [k, v] = it->next();
RestoreJobCloudPB restore_job_pb;
if (!restore_job_pb.ParseFromArray(v.data(), v.size())) {
LOG_WARNING("malformed restore job value").tag("key", hex(k));
return -1;
}
switch (restore_job_pb.state()) {
case RestoreJobCloudPB::PREPARED:
++num_prepared;
break;
case RestoreJobCloudPB::COMMITTED:
++num_committed;
break;
case RestoreJobCloudPB::DROPPED:
++num_dropped;
break;
case RestoreJobCloudPB::COMPLETED:
++num_completed;
break;
case RestoreJobCloudPB::RECYCLING:
++num_recycling;
break;
default:
break;
}
int64_t current_time = ::time(nullptr);
if ((restore_job_pb.state() == RestoreJobCloudPB::PREPARED ||
restore_job_pb.state() == RestoreJobCloudPB::COMMITTED) &&
current_time > restore_job_pb.ctime_s() + COST_MANY_THRESHOLD) {
// restore job run more than 1 hour
++num_cost_many_time;
LOG_WARNING("restore job cost too many time")
.tag("key", hex(k))
.tag("tablet_id", restore_job_pb.tablet_id())
.tag("state", restore_job_pb.state())
.tag("ctime_s", restore_job_pb.ctime_s())
.tag("mtime_s", restore_job_pb.mtime_s());
}
if (!it->has_next()) {
begin = k;
begin.push_back('\x00'); // Update to next smallest key for iteration
break;
}
}
} while (it->more() && !stopped());
return 0;
}
int InstanceChecker::check_txn_info_key(std::string_view key, std::string_view value) {
std::unordered_map<int64_t, std::string> txn_info_;
TxnLabelPB txn_label_pb;
auto handle_check_txn_label_key = [&](std::string_view key, std::string_view value) -> int {
TxnInfoPB txn_info_pb;
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "txn" ${instance_id} "txn_info" ${db_id} ${txn_id}
if (!txn_info_pb.ParseFromArray(value.data(), value.size())) {
LOG(WARNING) << "failed to parse TxnInfoPB";
return -1;
}
auto txn_id = std::get<int64_t>(std::get<0>(out[4]));
auto it = txn_info_.find(txn_id);
if (it == txn_info_.end()) {
return 0;
} else {
if (it->second != txn_info_pb.label()) {
LOG(WARNING) << "txn_info_pb's txn_label not same with txn_label_pb's txn_label,"
<< " txn_info_pb's txn_label: " << txn_info_pb.label()
<< " txn_label_pb meta: " << txn_label_pb.ShortDebugString();
return 1;
}
}
return 0;
};
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "txn" ${instance_id} "txn_label" ${db_id} ${label}
if (!txn_label_pb.ParseFromArray(value.data(), value.size() - VERSION_STAMP_LEN)) {
LOG(WARNING) << "failed to parse TxnLabelPB";
return -1;
}
auto db_id = std::get<int64_t>(std::get<0>(out[3]));
auto label = std::get<std::string>(std::get<0>(out[4]));
// txn_id -> txn_label
for (const auto& txn_id : txn_label_pb.txn_ids()) {
txn_info_.insert({txn_id, label});
}
std::string txn_info_key_begin = txn_info_key({instance_id_, db_id, 0});
std::string txn_info_key_end = txn_info_key({instance_id_, db_id, INT64_MAX});
return scan_and_handle_kv(txn_info_key_begin, txn_info_key_end,
[&](std::string_view k, std::string_view v) -> int {
return handle_check_txn_label_key(k, v);
});
}
int InstanceChecker::check_txn_label_key(std::string_view key, std::string_view value) {
TxnInfoPB txn_info_pb;
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "txn" ${instance_id} "txn_info" ${db_id} ${txn_id}
if (!txn_info_pb.ParseFromArray(value.data(), value.size())) {
LOG(WARNING) << "failed to parse TxnInfoPB";
return -1;
}
auto txn_id = std::get<int64_t>(std::get<0>(out[4]));
auto db_id = std::get<int64_t>(std::get<0>(out[3]));
auto label = txn_info_pb.label();
std::string txn_label = txn_label_key({instance_id_, db_id, label});
std::string txn_label_val;
TxnLabelPB txn_label_pb;
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn";
return -1;
}
if (txn->get(txn_label, &txn_label_val) != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get txn label key, key=" << hex(txn_label);
return -1;
}
txn_label_pb.ParseFromString(txn_label_val);
auto txn_ids = txn_label_pb.txn_ids();
if (!std::count(txn_ids.begin(), txn_ids.end(), txn_id)) {
// clang-format off txn_info_pb
LOG(WARNING) << "txn_info_pb's txn_id not found in txn_label_pb info,"
<< " txn_id: " << txn_id
<< " txn_label_pb meta: " << txn_label_pb.ShortDebugString();
// clang-format on
return 1;
}
return 0;
}
int InstanceChecker::check_txn_index_key(std::string_view key, std::string_view value) {
TxnInfoPB txn_info_pb;
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "txn" ${instance_id} "txn_info" ${db_id} ${txn_id}
if (!txn_info_pb.ParseFromArray(value.data(), value.size())) {
LOG(WARNING) << "failed to parse TxnInfoPB";
return -1;
}
auto txn_id = std::get<int64_t>(std::get<0>(out[4]));
auto db_id = std::get<int64_t>(std::get<0>(out[3]));
/// get tablet id
std::string txn_index = txn_index_key({instance_id_, txn_id});
std::string txn_index_val;
TxnIndexPB txn_index_pb;
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn";
return -1;
}
if (txn->get(txn_index, &txn_index_val) != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get txn label key, key=" << hex(txn_index);
return -1;
}
txn_index_pb.ParseFromString(txn_index_val);
if (txn_index_pb.tablet_index().db_id() != db_id) {
// clang-format off txn_info_pb
LOG(WARNING) << "txn_index_pb's db_id not same with txn_info_pb's db_id,"
<< " txn_index_pb meta: " << txn_index_pb.ShortDebugString()
<< " txn_info_pb meta: " << txn_info_pb.ShortDebugString();
// clang-format on
return 1;
}
return 0;
}
int InstanceChecker::check_txn_running_key(std::string_view key, std::string_view value) {
TxnRunningPB txn_running_pb;
int64_t current_time =
duration_cast<milliseconds>(system_clock::now().time_since_epoch()).count();
if (!txn_running_pb.ParseFromArray(value.data(), value.size())) {
LOG(WARNING) << "failed to parse TxnRunningPB";
return -1;
}
if (txn_running_pb.timeout_time() <= current_time) {
LOG(WARNING) << "txn_running_pb.timeout_time() is less than current_time,"
<< " but txn_running_key exists, "
<< " txn_running_pb meta: " << txn_running_pb.ShortDebugString();
return 1;
}
return 0;
}
int InstanceChecker::do_txn_key_check() {
int ret = 0;
// check txn info key depend on txn label key
std::string begin = txn_label_key({instance_id_, 0, ""});
std::string end = txn_label_key({instance_id_, INT64_MAX, ""});
int64_t num_scanned = 0;
int64_t num_abnormal = 0;
LOG(INFO) << "begin check txn_label_key and txn_info_key";
ret = scan_and_handle_kv(begin, end, [&, this](std::string_view k, std::string_view v) -> int {
num_scanned++;
int ret = check_txn_info_key(k, v);
if (ret == 1) {
num_abnormal++;
}
return ret;
});
if (ret == 1) {
LOG(WARNING) << "failed to check txn_info_key depending on txn_label_key, num_scanned="
<< num_scanned << ", num_abnormal=" << num_abnormal;
return 1;
} else if (ret == -1) {
LOG(WARNING) << "failed to check txn label key and txn info key";
return -1;
}
// check txn label key depend on txn info key
begin = txn_info_key({instance_id_, 0, 0});
end = txn_info_key({instance_id_, INT64_MAX, 0});
num_scanned = 0;
num_abnormal = 0;
LOG(INFO) << "begin check txn_label_key and txn_info_key";
ret = scan_and_handle_kv(begin, end, [&, this](std::string_view k, std::string_view v) -> int {
num_scanned++;
int ret = check_txn_label_key(k, v);
if (ret == 1) {
num_abnormal++;
}
return ret;
});
if (ret == 1) {
LOG(WARNING) << "failed to check txn_label_key depending on txn_info_key, num_scanned="
<< num_scanned << ", num_abnormal=" << num_abnormal;
return 1;
} else if (ret == -1) {
LOG(WARNING) << "failed to inverted check txn label key and txn info key";
return -1;
}
LOG(INFO) << "finish check txn_label_key and txn_info_key, num_scanned=" << num_scanned
<< ", num_abnormal=" << num_abnormal;
// check txn index key depend on txn info key
begin = txn_info_key({instance_id_, 0, 0});
end = txn_info_key({instance_id_, INT64_MAX, 0});
num_scanned = 0;
num_abnormal = 0;
LOG(INFO) << "begin check txn_index_key and txn_info_key";
ret = scan_and_handle_kv(begin, end, [&, this](std::string_view k, std::string_view v) -> int {
num_scanned++;
int ret = check_txn_index_key(k, v);
if (ret == 1) {
num_abnormal++;
}
return ret;
});
if (ret == 1) {
LOG(WARNING) << "failed to check txn_idx_key depending on txn_info_key, num_scanned="
<< num_scanned << ", num_abnormal=" << num_abnormal;
return 1;
} else if (ret == -1) {
LOG(WARNING) << "failed to check txn index key";
return -1;
}
LOG(INFO) << "finish check txn_index_key and txn_info_key, num_scanned=" << num_scanned
<< ", num_abnormal=" << num_abnormal;
// check txn running key
begin = txn_running_key({instance_id_, 0, 0});
end = txn_running_key({instance_id_, INT64_MAX, 0});
num_scanned = 0;
num_abnormal = 0;
LOG(INFO) << "begin check txn_running_key";
ret = scan_and_handle_kv(begin, end, [&, this](std::string_view k, std::string_view v) -> int {
num_scanned++;
int ret = check_txn_running_key(k, v);
if (ret == 1) {
num_abnormal++;
}
return ret;
});
if (ret == 1) {
LOG(WARNING) << "failed to check txn_running_key, num_scanned=" << num_scanned
<< ", num_abnormal=" << num_abnormal;
return 1;
} else if (ret == -1) {
LOG(WARNING) << "failed to check txn running key";
return -1;
}
LOG(INFO) << "finish check txn_running_key, num_scanned=" << num_scanned
<< ", num_abnormal=" << num_abnormal;
return 0;
}
int InstanceChecker::check_meta_tmp_rowset_key(std::string_view key, std::string_view value) {
TxnInfoPB txn_info_pb;
std::string_view k1 = key;
k1.remove_prefix(1);
std::vector<std::tuple<std::variant<int64_t, std::string>, int, int>> out;
decode_key(&k1, &out);
// 0x01 "txn" ${instance_id} "txn_info" ${db_id} ${txn_id}
if (!txn_info_pb.ParseFromArray(value.data(), value.size())) {
LOG(WARNING) << "failed to parse TxnInfoPB";
return -1;
}
/// get tablet id
auto txn_id = std::get<int64_t>(std::get<0>(out[4]));
std::string txn_index = txn_index_key({instance_id_, txn_id});
std::string txn_index_val;
TxnIndexPB txn_index_pb;
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to init txn";
return -1;
}
if (txn->get(txn_index, &txn_index_val) != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to get txn index key, key=" << txn_index;
return -1;
}
txn_index_pb.ParseFromString(txn_index_val);
auto tablet_id = txn_index_pb.tablet_index().tablet_id();
std::string meta_tmp_rowset_key = meta_rowset_tmp_key({instance_id_, txn_id, tablet_id});
int is_key_exist = key_exist(txn_kv_.get(), meta_tmp_rowset_key);
if (is_key_exist == 1) {
if (txn_info_pb.status() != TxnStatusPB::TXN_STATUS_VISIBLE) {
// clang-format off
LOG(INFO) << "meta tmp rowset key not exist but txn status != TXN_STATUS_VISIBLE"
<< "meta tmp rowset key=" << meta_tmp_rowset_key
<< "txn_info=" << txn_info_pb.ShortDebugString();
// clang-format on
return 1;
}
} else if (is_key_exist == 0) {
if (txn_info_pb.status() != TxnStatusPB::TXN_STATUS_PREPARED) {
// clang-format off
LOG(INFO) << "meta tmp rowset key exist but txn status != TXN_STATUS_PREPARED"
<< "meta tmp rowset key=" << meta_tmp_rowset_key
<< "txn_info=" << txn_info_pb.ShortDebugString();
// clang-format on
return 1;
}
} else {
LOG(WARNING) << "failed to get key, key=" << meta_tmp_rowset_key;
return -1;
}
return 0;
}
int InstanceChecker::check_meta_rowset_key(std::string_view key, std::string_view value) {
RowsetMetaCloudPB meta_rowset_pb;
if (!meta_rowset_pb.ParseFromArray(value.data(), value.size())) {
LOG(WARNING) << "failed to parse RowsetMetaCloudPB";
return -1;
}
std::string tablet_index_key = meta_tablet_idx_key({instance_id_, meta_rowset_pb.tablet_id()});
if (key_exist(txn_kv_.get(), tablet_index_key) == 1) {
LOG(WARNING) << "rowset's tablet id not found in fdb"
<< "tablet_index_key: " << tablet_index_key
<< "rowset meta: " << meta_rowset_pb.ShortDebugString();
return 1;
}
return 0;
}
int InstanceChecker::do_meta_rowset_key_check() {
int ret = 0;
std::string begin = meta_rowset_key({instance_id_, 0, 0});
std::string end = meta_rowset_key({instance_id_, INT64_MAX, 0});
int64_t num_scanned = 0;
int64_t num_loss = 0;
ret = scan_and_handle_kv(begin, end, [&](std::string_view k, std::string_view v) {
num_scanned++;
int ret = check_meta_rowset_key(k, v);
if (ret == 1) {
num_loss++;
}
return ret;
});
if (ret == -1) {
LOG(WARNING) << "failed to check meta rowset key,";
return -1;
} else if (ret == 1) {
LOG(WARNING) << "meta rowset key may be loss, num_scanned=" << num_scanned
<< ", num_loss=" << num_loss;
}
LOG(INFO) << "meta rowset key check finish, num_scanned=" << num_scanned
<< ", num_loss=" << num_loss;
begin = txn_info_key({instance_id_, 0, 0});
end = txn_info_key({instance_id_, INT64_MAX, 0});
num_scanned = 0;
num_loss = 0;
ret = scan_and_handle_kv(begin, end, [&](std::string_view k, std::string_view v) {
num_scanned++;
int ret = check_meta_tmp_rowset_key(k, v);
if (ret == 1) {
num_loss++;
}
return ret;
});
if (ret == -1) {
LOG(WARNING) << "failed to check tmp meta rowset key";
return -1;
} else if (ret == 1) {
LOG(WARNING) << "meta tmp rowset key may be loss, num_scanned=" << num_scanned
<< ", num_loss=" << num_loss;
}
LOG(INFO) << "meta tmp rowset key check finish, num_scanned=" << num_scanned
<< ", num_loss=" << num_loss;
return ret;
}
StorageVaultAccessor* InstanceChecker::get_accessor(const std::string& id) {
auto it = accessor_map_.find(id);
if (it == accessor_map_.end()) {
return nullptr;
}
return it->second.get();
}
void InstanceChecker::get_all_accessor(std::vector<StorageVaultAccessor*>* accessors) {
for (const auto& [_, accessor] : accessor_map_) {
accessors->push_back(accessor.get());
}
}
int InstanceChecker::do_packed_file_check() {
LOG(INFO) << "begin to check packed files, instance_id=" << instance_id_;
int check_ret = 0;
long num_scanned_rowsets = 0;
long num_scanned_packed_files = 0;
long num_packed_file_loss = 0;
long num_packed_file_leak = 0;
long num_ref_count_mismatch = 0;
long num_small_file_ref_mismatch = 0;
using namespace std::chrono;
auto start_time = steady_clock::now();
DORIS_CLOUD_DEFER {
auto cost = duration<float>(steady_clock::now() - start_time).count();
LOG(INFO) << "check packed files finished, cost=" << cost
<< "s. instance_id=" << instance_id_
<< " num_scanned_rowsets=" << num_scanned_rowsets
<< " num_scanned_packed_files=" << num_scanned_packed_files
<< " num_packed_file_loss=" << num_packed_file_loss
<< " num_packed_file_leak=" << num_packed_file_leak
<< " num_ref_count_mismatch=" << num_ref_count_mismatch
<< " num_small_file_ref_mismatch=" << num_small_file_ref_mismatch;
};
// Map to track expected reference count for each packed file
// packed_file_path -> expected_ref_count (from rowset metas)
std::unordered_map<std::string, int64_t> expected_ref_counts;
// Map to track small files referenced in packed files
// packed_file_path -> set of small_file_paths
std::unordered_map<std::string, std::unordered_set<std::string>> packed_file_small_files;
// Step 1: Scan all rowset metas to collect packed_slice_locations references
// Use efficient range scan instead of iterating through each tablet_id
auto collect_packed_refs = [&](const doris::RowsetMetaCloudPB& rs_meta) {
const auto& index_map = rs_meta.packed_slice_locations();
for (const auto& [small_file_path, index_pb] : index_map) {
if (!index_pb.has_packed_file_path() || index_pb.packed_file_path().empty()) {
continue;
}
const std::string& packed_file_path = index_pb.packed_file_path();
expected_ref_counts[packed_file_path]++;
packed_file_small_files[packed_file_path].insert(small_file_path);
}
};
{
std::string start_key = meta_rowset_key({instance_id_, 0, 0});
std::string end_key = meta_rowset_key({instance_id_, INT64_MAX, 0});
std::unique_ptr<RangeGetIterator> it;
do {
if (stopped()) {
return -1;
}
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn for packed file check";
return -1;
}
err = txn->get(start_key, end_key, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to scan rowset metas, err=" << err;
check_ret = -1;
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
if (!it->has_next()) {
start_key = k;
}
doris::RowsetMetaCloudPB rs_meta;
if (!rs_meta.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed rowset meta, key=" << hex(k);
check_ret = -1;
continue;
}
num_scanned_rowsets++;
collect_packed_refs(rs_meta);
}
start_key.push_back('\x00'); // Update to next smallest key for iteration
} while (it->more() && !stopped());
}
// Rowsets in recycle keys may still hold packed file references while ref count
// updates are pending, so include them when calculating expected references.
{
std::string start_key = recycle_rowset_key({instance_id_, 0, ""});
std::string end_key = recycle_rowset_key({instance_id_, INT64_MAX, "\xff"});
std::unique_ptr<RangeGetIterator> it;
do {
if (stopped()) {
return -1;
}
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn for recycle rowset scan in packed file check";
return -1;
}
err = txn->get(start_key, end_key, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to scan recycle rowset metas, err=" << err;
check_ret = -1;
break;
}
while (it->has_next() && !stopped()) {
auto [k, v] = it->next();
if (!it->has_next()) {
start_key = k;
}
RecycleRowsetPB recycle_rowset;
if (!recycle_rowset.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "malformed recycle rowset, key=" << hex(k);
check_ret = -1;
continue;
}
if (!recycle_rowset.has_rowset_meta()) {
continue;
}
num_scanned_rowsets++;
collect_packed_refs(recycle_rowset.rowset_meta());
}
start_key.push_back('\x00'); // Update to next smallest key for iteration
} while (it->more() && !stopped());
}
// Step 2: Scan all packed file metadata and verify
// Also collect all packed file paths from metadata for Step 3
// Map: resource_id -> set of packed_file_paths
std::unordered_map<std::string, std::unordered_set<std::string>> packed_files_in_metadata;
std::string begin = packed_file_key({instance_id_, ""});
std::string end = packed_file_key({instance_id_, "\xff"});
std::string scan_begin = begin;
while (true) {
if (stopped()) {
return -1;
}
std::unique_ptr<Transaction> txn;
TxnErrorCode err = txn_kv_->create_txn(&txn);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to create txn for scanning packed files";
return -1;
}
std::unique_ptr<RangeGetIterator> it;
err = txn->get(scan_begin, end, &it);
if (err != TxnErrorCode::TXN_OK) {
LOG(WARNING) << "failed to scan packed file keys, err=" << err;
return -1;
}
if (!it->has_next()) {
break;
}
std::string last_key;
while (it->has_next()) {
auto [k, v] = it->next();
last_key.assign(k.data(), k.size());
num_scanned_packed_files++;
std::string packed_file_path;
if (!InstanceRecycler::decode_packed_file_key(k, &packed_file_path)) {
LOG(WARNING) << "failed to decode packed file key, key=" << hex(k);
check_ret = -1;
continue;
}
cloud::PackedFileInfoPB packed_info;
if (!packed_info.ParseFromArray(v.data(), v.size())) {
LOG(WARNING) << "failed to parse packed file info, packed_file_path="
<< packed_file_path;
check_ret = -1;
continue;
}
// Step 2.1: Verify packed file exists in storage
if (!packed_info.resource_id().empty()) {
// Collect packed file path for Step 3
packed_files_in_metadata[packed_info.resource_id()].insert(packed_file_path);
auto* accessor = get_accessor(packed_info.resource_id());
if (accessor == nullptr) {
LOG(WARNING) << "accessor not found for packed file, resource_id="
<< packed_info.resource_id()
<< ", packed_file_path=" << packed_file_path;
check_ret = -1;
continue;
}
int ret = accessor->exists(packed_file_path);
if (ret < 0) {
LOG(WARNING) << "failed to check packed file existence, packed_file_path="
<< packed_file_path << ", ret=" << ret;
check_ret = -1;
continue;
}
if (ret != 0) {
// ret == 1 means file not found, ret > 1 means other error
// When packed file doesn't exist in storage, ref_cnt must be 0 and state must be RECYCLING
bool ref_cnt_valid = (packed_info.ref_cnt() == 0);
bool state_valid = (packed_info.state() == cloud::PackedFileInfoPB::RECYCLING);
if (!ref_cnt_valid || !state_valid) {
LOG(WARNING) << "packed file not found in storage but metadata is invalid, "
"packed_file_path="
<< packed_file_path << ", ref_cnt=" << packed_info.ref_cnt()
<< " (expected=0), state=" << packed_info.state()
<< " (expected=RECYCLING), ret=" << ret;
num_packed_file_loss++;
check_ret = 1; // Data inconsistency identified
}
// If ref_cnt == 0 and state == RECYCLING, this is expected (file is being recycled)
}
// ret == 0 means file exists, which is expected
}
// Step 2.2: Verify reference count matches expected count
int64_t expected_ref = expected_ref_counts[packed_file_path];
if (packed_info.ref_cnt() != expected_ref) {
LOG(WARNING) << "packed file ref count mismatch, packed_file_path="
<< packed_file_path << ", expected=" << expected_ref
<< ", actual=" << packed_info.ref_cnt();
num_ref_count_mismatch++;
check_ret = 1; // Data inconsistency identified
}
// Step 2.3: Verify small files in packed_info match rowset references
std::unordered_set<std::string> small_files_in_meta;
for (const auto& small_file : packed_info.slices()) {
if (!small_file.deleted()) {
small_files_in_meta.insert(small_file.path());
}
}
const auto& expected_small_files = packed_file_small_files[packed_file_path];
if (small_files_in_meta != expected_small_files) {
// Check for missing small files
for (const auto& expected_path : expected_small_files) {
if (small_files_in_meta.find(expected_path) == small_files_in_meta.end()) {
LOG(WARNING) << "small file missing in packed file info, packed_file_path="
<< packed_file_path << ", small_file_path=" << expected_path;
num_small_file_ref_mismatch++;
check_ret = 1;
}
}
// Check for extra small files (may be deleted, so less critical)
for (const auto& meta_path : small_files_in_meta) {
if (expected_small_files.find(meta_path) == expected_small_files.end()) {
LOG(INFO) << "small file in packed file info not found in rowset metas, "
"may be deleted, packed_file_path="
<< packed_file_path << ", small_file_path=" << meta_path;
}
}
}
}
if (!it->more()) {
break;
}
scan_begin = last_key;
scan_begin.push_back('\x00');
}
// Step 3: Check for leaked packed files (exist in storage but not in metadata)
// Scan all storage vaults to find packed files and verify they are in metadata
{
std::vector<StorageVaultAccessor*> accessors;
get_all_accessor(&accessors);
for (StorageVaultAccessor* accessor : accessors) {
if (stopped()) {
return -1;
}
// Find resource_id for this accessor
std::string resource_id;
for (const auto& [id, acc] : accessor_map_) {
if (acc.get() == accessor) {
resource_id = id;
break;
}
}
if (resource_id.empty()) {
continue;
}
// List all files under data/packed_file/ directory
std::unique_ptr<ListIterator> list_it;
int ret = accessor->list_directory("data/packed_file", &list_it);
if (ret != 0) {
// Directory may not exist, which is fine
if (ret < 0) {
LOG(WARNING) << "failed to list packed_file directory, resource_id="
<< resource_id << ", ret=" << ret;
check_ret = -1;
}
continue;
}
const auto& expected_packed_files = packed_files_in_metadata[resource_id];
while (list_it->has_next()) {
if (stopped()) {
return -1;
}
auto file_meta = list_it->next();
if (!file_meta.has_value()) {
break;
}
const std::string& file_path = file_meta->path;
// Only check files (not directories), and ensure it's a packed file
// Skip directories (paths ending with '/') and non-packed-file paths
if (file_path.empty() || file_path.back() == '/' ||
!file_path.starts_with("data/packed_file/")) {
continue;
}
// Check if this packed file is in metadata
if (expected_packed_files.find(file_path) == expected_packed_files.end()) {
LOG(WARNING) << "packed file found in storage but not in metadata, "
"resource_id="
<< resource_id << ", packed_file_path=" << file_path;
num_packed_file_leak++;
check_ret = 1; // Data leak identified
}
}
}
}
if (num_packed_file_loss > 0 || num_packed_file_leak > 0 || num_ref_count_mismatch > 0 ||
num_small_file_ref_mismatch > 0) {
return 1; // Data loss or inconsistency identified
}
if (check_ret < 0) {
return check_ret; // Temporary error
}
return 0; // Success
}
} // namespace doris::cloud