| /* |
| * Licensed to the Apache Software Foundation (ASF) under one |
| * or more contributor license agreements. See the NOTICE file |
| * distributed with this work for additional information |
| * regarding copyright ownership. The ASF licenses this file |
| * to you under the Apache License, Version 2.0 (the |
| * "License"); you may not use this file except in compliance |
| * with the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, |
| * software distributed under the License is distributed on an |
| * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| * KIND, either express or implied. See the License for the |
| * specific language governing permissions and limitations |
| * under the License. |
| */ |
| |
| #ifndef REQ_SKETCH_IMPL_HPP_ |
| #define REQ_SKETCH_IMPL_HPP_ |
| |
| #include <sstream> |
| #include <stdexcept> |
| |
| namespace datasketches { |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>::req_sketch(uint16_t k, bool hra, const A& allocator): |
| allocator_(allocator), |
| k_(std::max(static_cast<int>(k) & -2, static_cast<int>(req_constants::MIN_K))), //rounds down one if odd |
| hra_(hra), |
| max_nom_size_(0), |
| num_retained_(0), |
| n_(0), |
| compactors_(allocator), |
| min_value_(nullptr), |
| max_value_(nullptr) |
| { |
| grow(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>::~req_sketch() { |
| if (min_value_ != nullptr) { |
| min_value_->~T(); |
| allocator_.deallocate(min_value_, 1); |
| } |
| if (max_value_ != nullptr) { |
| max_value_->~T(); |
| allocator_.deallocate(max_value_, 1); |
| } |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>::req_sketch(const req_sketch& other): |
| allocator_(other.allocator_), |
| k_(other.k_), |
| hra_(other.hra_), |
| max_nom_size_(other.max_nom_size_), |
| num_retained_(other.num_retained_), |
| n_(other.n_), |
| compactors_(other.compactors_), |
| min_value_(nullptr), |
| max_value_(nullptr) |
| { |
| if (other.min_value_ != nullptr) min_value_ = new (A().allocate(1)) T(*other.min_value_); |
| if (other.max_value_ != nullptr) max_value_ = new (A().allocate(1)) T(*other.max_value_); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>::req_sketch(req_sketch&& other) noexcept : |
| allocator_(std::move(other.allocator_)), |
| k_(other.k_), |
| hra_(other.hra_), |
| max_nom_size_(other.max_nom_size_), |
| num_retained_(other.num_retained_), |
| n_(other.n_), |
| compactors_(std::move(other.compactors_)), |
| min_value_(other.min_value_), |
| max_value_(other.max_value_) |
| { |
| other.min_value_ = nullptr; |
| other.max_value_ = nullptr; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>& req_sketch<T, C, S, A>::operator=(const req_sketch& other) { |
| req_sketch copy(other); |
| std::swap(allocator_, copy.allocator_); |
| std::swap(k_, copy.k_); |
| std::swap(hra_, copy.hra_); |
| std::swap(max_nom_size_, copy.max_nom_size_); |
| std::swap(num_retained_, copy.num_retained_); |
| std::swap(n_, copy.n_); |
| std::swap(compactors_, copy.compactors_); |
| std::swap(min_value_, copy.min_value_); |
| std::swap(max_value_, copy.max_value_); |
| return *this; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>& req_sketch<T, C, S, A>::operator=(req_sketch&& other) { |
| std::swap(allocator_, other.allocator_); |
| std::swap(k_, other.k_); |
| std::swap(hra_, other.hra_); |
| std::swap(max_nom_size_, other.max_nom_size_); |
| std::swap(num_retained_, other.num_retained_); |
| std::swap(n_, other.n_); |
| std::swap(compactors_, other.compactors_); |
| std::swap(min_value_, other.min_value_); |
| std::swap(max_value_, other.max_value_); |
| return *this; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| uint16_t req_sketch<T, C, S, A>::get_k() const { |
| return k_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| bool req_sketch<T, C, S, A>::is_HRA() const { |
| return hra_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| bool req_sketch<T, C, S, A>::is_empty() const { |
| return n_ == 0; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| uint64_t req_sketch<T, C, S, A>::get_n() const { |
| return n_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| uint32_t req_sketch<T, C, S, A>::get_num_retained() const { |
| return num_retained_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| bool req_sketch<T, C, S, A>::is_estimation_mode() const { |
| return compactors_.size() > 1; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<typename FwdT> |
| void req_sketch<T, C, S, A>::update(FwdT&& item) { |
| if (!check_update_value(item)) { return; } |
| if (is_empty()) { |
| min_value_ = new (allocator_.allocate(1)) T(item); |
| max_value_ = new (allocator_.allocate(1)) T(item); |
| } else { |
| if (C()(item, *min_value_)) *min_value_ = item; |
| if (C()(*max_value_, item)) *max_value_ = item; |
| } |
| compactors_[0].append(std::forward<FwdT>(item)); |
| ++num_retained_; |
| ++n_; |
| if (num_retained_ == max_nom_size_) compress(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<typename FwdSk> |
| void req_sketch<T, C, S, A>::merge(FwdSk&& other) { |
| if (is_HRA() != other.is_HRA()) throw std::invalid_argument("merging HRA and LRA is not valid"); |
| if (other.is_empty()) return; |
| if (is_empty()) { |
| min_value_ = new (allocator_.allocate(1)) T(conditional_forward<FwdSk>(*other.min_value_)); |
| max_value_ = new (allocator_.allocate(1)) T(conditional_forward<FwdSk>(*other.max_value_)); |
| } else { |
| if (C()(*other.min_value_, *min_value_)) *min_value_ = conditional_forward<FwdSk>(*other.min_value_); |
| if (C()(*max_value_, *other.max_value_)) *max_value_ = conditional_forward<FwdSk>(*other.max_value_); |
| } |
| // grow until this has at least as many compactors as other |
| while (get_num_levels() < other.get_num_levels()) grow(); |
| // merge the items in all height compactors |
| for (size_t i = 0; i < other.get_num_levels(); ++i) { |
| compactors_[i].merge(conditional_forward<FwdSk>(other.compactors_[i])); |
| } |
| n_ += other.n_; |
| update_max_nom_size(); |
| update_num_retained(); |
| if (num_retained_ >= max_nom_size_) compress(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| const T& req_sketch<T, C, S, A>::get_min_value() const { |
| if (is_empty()) return get_invalid_value(); |
| return *min_value_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| const T& req_sketch<T, C, S, A>::get_max_value() const { |
| if (is_empty()) return get_invalid_value(); |
| return *max_value_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<bool inclusive> |
| double req_sketch<T, C, S, A>::get_rank(const T& item) const { |
| uint64_t weight = 0; |
| for (const auto& compactor: compactors_) { |
| weight += compactor.template compute_weight<inclusive>(item); |
| } |
| return static_cast<double>(weight) / n_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<bool inclusive> |
| auto req_sketch<T, C, S, A>::get_PMF(const T* split_points, uint32_t size) const -> vector_double { |
| auto buckets = get_CDF<inclusive>(split_points, size); |
| for (uint32_t i = size; i > 0; --i) { |
| buckets[i] -= buckets[i - 1]; |
| } |
| return buckets; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<bool inclusive> |
| auto req_sketch<T, C, S, A>::get_CDF(const T* split_points, uint32_t size) const -> vector_double { |
| vector_double buckets(allocator_); |
| if (is_empty()) return buckets; |
| check_split_points(split_points, size); |
| buckets.reserve(size + 1); |
| for (uint32_t i = 0; i < size; ++i) buckets.push_back(get_rank<inclusive>(split_points[i])); |
| buckets.push_back(1); |
| return buckets; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<bool inclusive> |
| const T& req_sketch<T, C, S, A>::get_quantile(double rank) const { |
| if (is_empty()) return get_invalid_value(); |
| if (rank == 0.0) return *min_value_; |
| if (rank == 1.0) return *max_value_; |
| if ((rank < 0.0) || (rank > 1.0)) { |
| throw std::invalid_argument("Rank cannot be less than zero or greater than 1.0"); |
| } |
| return *(get_quantile_calculator<inclusive>()->get_quantile(rank)); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| template<bool inclusive> |
| std::vector<T, A> req_sketch<T, C, S, A>::get_quantiles(const double* ranks, uint32_t size) const { |
| std::vector<T, A> quantiles(allocator_); |
| if (is_empty()) return quantiles; |
| QuantileCalculatorPtr quantile_calculator(nullptr, calculator_deleter(allocator_)); |
| quantiles.reserve(size); |
| for (uint32_t i = 0; i < size; ++i) { |
| const double rank = ranks[i]; |
| if ((rank < 0.0) || (rank > 1.0)) { |
| throw std::invalid_argument("rank cannot be less than zero or greater than 1.0"); |
| } |
| if (rank == 0.0) quantiles.push_back(*min_value_); |
| else if (rank == 1.0) quantiles.push_back(*max_value_); |
| else { |
| if (!quantile_calculator) { |
| // has side effect of sorting level zero if needed |
| quantile_calculator = const_cast<req_sketch*>(this)->get_quantile_calculator<inclusive>(); |
| } |
| quantiles.push_back(*(quantile_calculator->get_quantile(rank))); |
| } |
| } |
| return quantiles; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| class req_sketch<T, C, S, A>::calculator_deleter { |
| public: |
| calculator_deleter(const AllocCalc& allocator): allocator_(allocator) {} |
| void operator() (QuantileCalculator* ptr) { |
| if (ptr != nullptr) { |
| ptr->~QuantileCalculator(); |
| allocator_.deallocate(ptr, 1); |
| } |
| } |
| private: |
| AllocCalc allocator_; |
| }; |
| |
| template<typename T, typename C, typename S, typename A> |
| template<bool inclusive> |
| auto req_sketch<T, C, S, A>::get_quantile_calculator() const -> QuantileCalculatorPtr { |
| if (!compactors_[0].is_sorted()) { |
| const_cast<Compactor&>(compactors_[0]).sort(); // allow this side effect |
| } |
| AllocCalc ac(allocator_); |
| QuantileCalculatorPtr quantile_calculator( |
| new (ac.allocate(1)) req_quantile_calculator<T, C, A>(n_, ac), |
| calculator_deleter(ac) |
| ); |
| |
| for (auto& compactor: compactors_) { |
| quantile_calculator->add(compactor.begin(), compactor.end(), compactor.get_lg_weight()); |
| } |
| quantile_calculator->template convert_to_cummulative<inclusive>(); |
| return quantile_calculator; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| double req_sketch<T, C, S, A>::get_rank_lower_bound(double rank, uint8_t num_std_dev) const { |
| return get_rank_lb(get_k(), get_num_levels(), rank, num_std_dev, get_n(), hra_); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| double req_sketch<T, C, S, A>::get_rank_upper_bound(double rank, uint8_t num_std_dev) const { |
| return get_rank_ub(get_k(), get_num_levels(), rank, num_std_dev, get_n(), hra_); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| double req_sketch<T, C, S, A>::get_RSE(uint16_t k, double rank, bool hra, uint64_t n) { |
| return get_rank_lb(k, 2, rank, 1, n, hra); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| double req_sketch<T, C, S, A>::get_rank_lb(uint16_t k, uint8_t num_levels, double rank, uint8_t num_std_dev, uint64_t n, bool hra) { |
| if (is_exact_rank(k, num_levels, rank, n, hra)) return rank; |
| const double relative = relative_rse_factor() / k * (hra ? 1.0 - rank : rank); |
| const double fixed = FIXED_RSE_FACTOR / k; |
| const double lb_rel = rank - num_std_dev * relative; |
| const double lb_fix = rank - num_std_dev * fixed; |
| return std::max(lb_rel, lb_fix); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| double req_sketch<T, C, S, A>::get_rank_ub(uint16_t k, uint8_t num_levels, double rank, uint8_t num_std_dev, uint64_t n, bool hra) { |
| if (is_exact_rank(k, num_levels, rank, n, hra)) return rank; |
| const double relative = relative_rse_factor() / k * (hra ? 1.0 - rank : rank); |
| const double fixed = FIXED_RSE_FACTOR / k; |
| const double ub_rel = rank + num_std_dev * relative; |
| const double ub_fix = rank + num_std_dev * fixed; |
| return std::min(ub_rel, ub_fix); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| bool req_sketch<T, C, S, A>::is_exact_rank(uint16_t k, uint8_t num_levels, double rank, uint64_t n, bool hra) { |
| const unsigned base_cap = k * req_constants::INIT_NUM_SECTIONS; |
| if (num_levels == 1 || n <= base_cap) return true; |
| const double exact_rank_thresh = static_cast<double>(base_cap) / n; |
| return (hra && rank >= 1.0 - exact_rank_thresh) || (!hra && rank <= exact_rank_thresh); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| double req_sketch<T, C, S, A>::relative_rse_factor() { |
| return sqrt(0.0512 / req_constants::INIT_NUM_SECTIONS); |
| } |
| |
| // implementation for fixed-size arithmetic types (integral and floating point) |
| template<typename T, typename C, typename S, typename A> |
| template<typename TT, typename std::enable_if<std::is_arithmetic<TT>::value, int>::type> |
| size_t req_sketch<T, C, S, A>::get_serialized_size_bytes() const { |
| size_t size = PREAMBLE_SIZE_BYTES; |
| if (is_empty()) return size; |
| if (is_estimation_mode()) { |
| size += sizeof(n_) + sizeof(TT) * 2; // min and max |
| } |
| if (n_ == 1) { |
| size += sizeof(TT); |
| } else { |
| for (const auto& compactor: compactors_) size += compactor.get_serialized_size_bytes(S()); |
| } |
| return size; |
| } |
| |
| // implementation for all other types |
| template<typename T, typename C, typename S, typename A> |
| template<typename TT, typename std::enable_if<!std::is_arithmetic<TT>::value, int>::type> |
| size_t req_sketch<T, C, S, A>::get_serialized_size_bytes() const { |
| size_t size = PREAMBLE_SIZE_BYTES; |
| if (is_empty()) return size; |
| if (is_estimation_mode()) { |
| size += sizeof(n_); |
| size += S().size_of_item(*min_value_); |
| size += S().size_of_item(*max_value_); |
| } |
| if (n_ == 1) { |
| size += S().size_of_item(*compactors_[0].begin()); |
| } else { |
| for (const auto& compactor: compactors_) size += compactor.get_serialized_size_bytes(S()); |
| } |
| return size; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::serialize(std::ostream& os) const { |
| const uint8_t preamble_ints = is_estimation_mode() ? 4 : 2; |
| write(os, preamble_ints); |
| const uint8_t serial_version = SERIAL_VERSION; |
| write(os, serial_version); |
| const uint8_t family = FAMILY; |
| write(os, family); |
| const bool raw_items = n_ <= req_constants::MIN_K; |
| const uint8_t flags_byte( |
| (is_empty() ? 1 << flags::IS_EMPTY : 0) |
| | (hra_ ? 1 << flags::IS_HIGH_RANK : 0) |
| | (raw_items ? 1 << flags::RAW_ITEMS : 0) |
| | (compactors_[0].is_sorted() ? 1 << flags::IS_LEVEL_ZERO_SORTED : 0) |
| ); |
| write(os, flags_byte); |
| write(os, k_); |
| const uint8_t num_levels = is_empty() ? 0 : get_num_levels(); |
| write(os, num_levels); |
| const uint8_t num_raw_items = raw_items ? n_ : 0; |
| write(os, num_raw_items); |
| if (is_empty()) return; |
| if (is_estimation_mode()) { |
| write(os, n_); |
| S().serialize(os, min_value_, 1); |
| S().serialize(os, max_value_, 1); |
| } |
| if (raw_items) { |
| S().serialize(os, compactors_[0].begin(), num_raw_items); |
| } else { |
| for (const auto& compactor: compactors_) compactor.serialize(os, S()); |
| } |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| auto req_sketch<T, C, S, A>::serialize(unsigned header_size_bytes) const -> vector_bytes { |
| const size_t size = header_size_bytes + get_serialized_size_bytes(); |
| vector_bytes bytes(size, 0, allocator_); |
| uint8_t* ptr = bytes.data() + header_size_bytes; |
| const uint8_t* end_ptr = ptr + size; |
| |
| const uint8_t preamble_ints = is_estimation_mode() ? 4 : 2; |
| ptr += copy_to_mem(preamble_ints, ptr); |
| const uint8_t serial_version = SERIAL_VERSION; |
| ptr += copy_to_mem(serial_version, ptr); |
| const uint8_t family = FAMILY; |
| ptr += copy_to_mem(family, ptr); |
| const bool raw_items = n_ <= req_constants::MIN_K; |
| const uint8_t flags_byte( |
| (is_empty() ? 1 << flags::IS_EMPTY : 0) |
| | (hra_ ? 1 << flags::IS_HIGH_RANK : 0) |
| | (raw_items ? 1 << flags::RAW_ITEMS : 0) |
| | (compactors_[0].is_sorted() ? 1 << flags::IS_LEVEL_ZERO_SORTED : 0) |
| ); |
| ptr += copy_to_mem(flags_byte, ptr); |
| ptr += copy_to_mem(k_, ptr); |
| const uint8_t num_levels = is_empty() ? 0 : get_num_levels(); |
| ptr += copy_to_mem(num_levels, ptr); |
| const uint8_t num_raw_items = raw_items ? n_ : 0; |
| ptr += copy_to_mem(num_raw_items, ptr); |
| if (!is_empty()) { |
| if (is_estimation_mode()) { |
| ptr += copy_to_mem(n_, ptr); |
| ptr += S().serialize(ptr, end_ptr - ptr, min_value_, 1); |
| ptr += S().serialize(ptr, end_ptr - ptr, max_value_, 1); |
| } |
| if (raw_items) { |
| ptr += S().serialize(ptr, end_ptr - ptr, compactors_[0].begin(), num_raw_items); |
| } else { |
| for (const auto& compactor: compactors_) ptr += compactor.serialize(ptr, end_ptr - ptr, S()); |
| } |
| } |
| return bytes; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A> req_sketch<T, C, S, A>::deserialize(std::istream& is, const A& allocator) { |
| const auto preamble_ints = read<uint8_t>(is); |
| const auto serial_version = read<uint8_t>(is); |
| const auto family_id = read<uint8_t>(is); |
| const auto flags_byte = read<uint8_t>(is); |
| const auto k = read<uint16_t>(is); |
| const auto num_levels = read<uint8_t>(is); |
| const auto num_raw_items = read<uint8_t>(is); |
| |
| check_preamble_ints(preamble_ints, num_levels); |
| check_serial_version(serial_version); |
| check_family_id(family_id); |
| |
| if (!is.good()) throw std::runtime_error("error reading from std::istream"); |
| const bool is_empty = flags_byte & (1 << flags::IS_EMPTY); |
| const bool hra = flags_byte & (1 << flags::IS_HIGH_RANK); |
| if (is_empty) return req_sketch(k, hra, allocator); |
| |
| A alloc(allocator); |
| auto item_buffer_deleter = [&alloc](T* ptr) { alloc.deallocate(ptr, 1); }; |
| std::unique_ptr<T, decltype(item_buffer_deleter)> min_value_buffer(alloc.allocate(1), item_buffer_deleter); |
| std::unique_ptr<T, decltype(item_buffer_deleter)> max_value_buffer(alloc.allocate(1), item_buffer_deleter); |
| std::unique_ptr<T, item_deleter> min_value(nullptr, item_deleter(allocator)); |
| std::unique_ptr<T, item_deleter> max_value(nullptr, item_deleter(allocator)); |
| |
| const bool raw_items = flags_byte & (1 << flags::RAW_ITEMS); |
| const bool is_level_0_sorted = flags_byte & (1 << flags::IS_LEVEL_ZERO_SORTED); |
| std::vector<Compactor, AllocCompactor> compactors(allocator); |
| |
| uint64_t n = 1; |
| if (num_levels > 1) { |
| n = read<uint64_t>(is); |
| S().deserialize(is, min_value_buffer.get(), 1); |
| // serde call did not throw, repackage with destrtuctor |
| min_value = std::unique_ptr<T, item_deleter>(min_value_buffer.release(), item_deleter(allocator)); |
| S().deserialize(is, max_value_buffer.get(), 1); |
| // serde call did not throw, repackage with destrtuctor |
| max_value = std::unique_ptr<T, item_deleter>(max_value_buffer.release(), item_deleter(allocator)); |
| } |
| |
| if (raw_items) { |
| compactors.push_back(Compactor::deserialize(is, S(), allocator, is_level_0_sorted, k, num_raw_items, hra)); |
| } else { |
| for (size_t i = 0; i < num_levels; ++i) { |
| compactors.push_back(Compactor::deserialize(is, S(), allocator, i == 0 ? is_level_0_sorted : true, hra)); |
| } |
| } |
| if (num_levels == 1) { |
| const auto begin = compactors[0].begin(); |
| const auto end = compactors[0].end(); |
| n = compactors[0].get_num_items(); |
| auto min_it = begin; |
| auto max_it = begin; |
| for (auto it = begin; it != end; ++it) { |
| if (C()(*it, *min_it)) min_it = it; |
| if (C()(*max_it, *it)) max_it = it; |
| } |
| new (min_value_buffer.get()) T(*min_it); |
| // copy did not throw, repackage with destrtuctor |
| min_value = std::unique_ptr<T, item_deleter>(min_value_buffer.release(), item_deleter(allocator)); |
| new (max_value_buffer.get()) T(*max_it); |
| // copy did not throw, repackage with destrtuctor |
| max_value = std::unique_ptr<T, item_deleter>(max_value_buffer.release(), item_deleter(allocator)); |
| } |
| |
| if (!is.good()) throw std::runtime_error("error reading from std::istream"); |
| return req_sketch(k, hra, n, std::move(min_value), std::move(max_value), std::move(compactors)); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A> req_sketch<T, C, S, A>::deserialize(const void* bytes, size_t size, const A& allocator) { |
| ensure_minimum_memory(size, 8); |
| const char* ptr = static_cast<const char*>(bytes); |
| const char* end_ptr = static_cast<const char*>(bytes) + size; |
| |
| uint8_t preamble_ints; |
| ptr += copy_from_mem(ptr, preamble_ints); |
| uint8_t serial_version; |
| ptr += copy_from_mem(ptr, serial_version); |
| uint8_t family_id; |
| ptr += copy_from_mem(ptr, family_id); |
| uint8_t flags_byte; |
| ptr += copy_from_mem(ptr, flags_byte); |
| uint16_t k; |
| ptr += copy_from_mem(ptr, k); |
| uint8_t num_levels; |
| ptr += copy_from_mem(ptr, num_levels); |
| uint8_t num_raw_items; |
| ptr += copy_from_mem(ptr, num_raw_items); |
| |
| check_preamble_ints(preamble_ints, num_levels); |
| check_serial_version(serial_version); |
| check_family_id(family_id); |
| |
| const bool is_empty = flags_byte & (1 << flags::IS_EMPTY); |
| const bool hra = flags_byte & (1 << flags::IS_HIGH_RANK); |
| if (is_empty) return req_sketch(k, hra, allocator); |
| |
| A alloc(allocator); |
| auto item_buffer_deleter = [&alloc](T* ptr) { alloc.deallocate(ptr, 1); }; |
| std::unique_ptr<T, decltype(item_buffer_deleter)> min_value_buffer(alloc.allocate(1), item_buffer_deleter); |
| std::unique_ptr<T, decltype(item_buffer_deleter)> max_value_buffer(alloc.allocate(1), item_buffer_deleter); |
| std::unique_ptr<T, item_deleter> min_value(nullptr, item_deleter(allocator)); |
| std::unique_ptr<T, item_deleter> max_value(nullptr, item_deleter(allocator)); |
| |
| const bool raw_items = flags_byte & (1 << flags::RAW_ITEMS); |
| const bool is_level_0_sorted = flags_byte & (1 << flags::IS_LEVEL_ZERO_SORTED); |
| std::vector<Compactor, AllocCompactor> compactors(allocator); |
| |
| uint64_t n = 1; |
| if (num_levels > 1) { |
| ensure_minimum_memory(end_ptr - ptr, sizeof(n)); |
| ptr += copy_from_mem(ptr, n); |
| ptr += S().deserialize(ptr, end_ptr - ptr, min_value_buffer.get(), 1); |
| // serde call did not throw, repackage with destrtuctor |
| min_value = std::unique_ptr<T, item_deleter>(min_value_buffer.release(), item_deleter(allocator)); |
| ptr += S().deserialize(ptr, end_ptr - ptr, max_value_buffer.get(), 1); |
| // serde call did not throw, repackage with destrtuctor |
| max_value = std::unique_ptr<T, item_deleter>(max_value_buffer.release(), item_deleter(allocator)); |
| } |
| |
| if (raw_items) { |
| auto pair = Compactor::deserialize(ptr, end_ptr - ptr, S(), allocator, is_level_0_sorted, k, num_raw_items, hra); |
| compactors.push_back(std::move(pair.first)); |
| ptr += pair.second; |
| } else { |
| for (size_t i = 0; i < num_levels; ++i) { |
| auto pair = Compactor::deserialize(ptr, end_ptr - ptr, S(), allocator, i == 0 ? is_level_0_sorted : true, hra); |
| compactors.push_back(std::move(pair.first)); |
| ptr += pair.second; |
| } |
| } |
| if (num_levels == 1) { |
| const auto begin = compactors[0].begin(); |
| const auto end = compactors[0].end(); |
| n = compactors[0].get_num_items(); |
| auto min_it = begin; |
| auto max_it = begin; |
| for (auto it = begin; it != end; ++it) { |
| if (C()(*it, *min_it)) min_it = it; |
| if (C()(*max_it, *it)) max_it = it; |
| } |
| new (min_value_buffer.get()) T(*min_it); |
| // copy did not throw, repackage with destrtuctor |
| min_value = std::unique_ptr<T, item_deleter>(min_value_buffer.release(), item_deleter(allocator)); |
| new (max_value_buffer.get()) T(*max_it); |
| // copy did not throw, repackage with destrtuctor |
| max_value = std::unique_ptr<T, item_deleter>(max_value_buffer.release(), item_deleter(allocator)); |
| } |
| |
| return req_sketch(k, hra, n, std::move(min_value), std::move(max_value), std::move(compactors)); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::grow() { |
| const uint8_t lg_weight = get_num_levels(); |
| compactors_.push_back(Compactor(hra_, lg_weight, k_, allocator_)); |
| update_max_nom_size(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| uint8_t req_sketch<T, C, S, A>::get_num_levels() const { |
| return compactors_.size(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::update_max_nom_size() { |
| max_nom_size_ = 0; |
| for (const auto& compactor: compactors_) max_nom_size_ += compactor.get_nom_capacity(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::update_num_retained() { |
| num_retained_ = 0; |
| for (const auto& compactor: compactors_) num_retained_ += compactor.get_num_items(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::compress() { |
| for (size_t h = 0; h < compactors_.size(); ++h) { |
| if (compactors_[h].get_num_items() >= compactors_[h].get_nom_capacity()) { |
| if (h == 0) compactors_[0].sort(); |
| if (h + 1 >= get_num_levels()) { // at the top? |
| grow(); // add a level, increases max_nom_size |
| } |
| auto pair = compactors_[h].compact(compactors_[h + 1]); |
| num_retained_ -= pair.first; |
| max_nom_size_ += pair.second; |
| if (LAZY_COMPRESSION && num_retained_ < max_nom_size_) break; |
| } |
| } |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| string<A> req_sketch<T, C, S, A>::to_string(bool print_levels, bool print_items) const { |
| std::basic_ostringstream<char, std::char_traits<char>, AllocChar<A>> os; |
| os << "### REQ sketch summary:" << std::endl; |
| os << " K : " << k_ << std::endl; |
| os << " High Rank Acc : " << (hra_ ? "true" : "false") << std::endl; |
| os << " Empty : " << (is_empty() ? "true" : "false") << std::endl; |
| os << " Estimation mode: " << (is_estimation_mode() ? "true" : "false") << std::endl; |
| os << " Sorted : " << (compactors_[0].is_sorted() ? "true" : "false") << std::endl; |
| os << " N : " << n_ << std::endl; |
| os << " Levels : " << compactors_.size() << std::endl; |
| os << " Retained items : " << num_retained_ << std::endl; |
| os << " Capacity items : " << max_nom_size_ << std::endl; |
| if (!is_empty()) { |
| os << " Min value : " << *min_value_ << std::endl; |
| os << " Max value : " << *max_value_ << std::endl; |
| } |
| os << "### End sketch summary" << std::endl; |
| |
| if (print_levels) { |
| os << "### REQ sketch levels:" << std::endl; |
| os << " index: nominal capacity, actual size" << std::endl; |
| for (uint8_t i = 0; i < compactors_.size(); i++) { |
| os << " " << (unsigned int) i << ": " |
| << compactors_[i].get_nom_capacity() << ", " |
| << compactors_[i].get_num_items() << std::endl; |
| } |
| os << "### End sketch levels" << std::endl; |
| } |
| |
| if (print_items) { |
| os << "### REQ sketch data:" << std::endl; |
| unsigned level = 0; |
| for (const auto& compactor: compactors_) { |
| os << " level " << level << ": " << std::endl; |
| for (auto it = compactor.begin(); it != compactor.end(); ++it) { |
| os << " " << *it << std::endl; |
| } |
| ++level; |
| } |
| os << "### End sketch data" << std::endl; |
| } |
| return os.str(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| class req_sketch<T, C, S, A>::item_deleter { |
| public: |
| item_deleter(const A& allocator): allocator_(allocator) {} |
| void operator() (T* ptr) { |
| if (ptr != nullptr) { |
| ptr->~T(); |
| allocator_.deallocate(ptr, 1); |
| } |
| } |
| private: |
| A allocator_; |
| }; |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>::req_sketch(uint32_t k, bool hra, uint64_t n, std::unique_ptr<T, item_deleter> min_value, std::unique_ptr<T, item_deleter> max_value, std::vector<Compactor, AllocCompactor>&& compactors): |
| allocator_(compactors.get_allocator()), |
| k_(k), |
| hra_(hra), |
| max_nom_size_(0), |
| num_retained_(0), |
| n_(n), |
| compactors_(std::move(compactors)), |
| min_value_(min_value.release()), |
| max_value_(max_value.release()) |
| { |
| update_max_nom_size(); |
| update_num_retained(); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::check_preamble_ints(uint8_t preamble_ints, uint8_t num_levels) { |
| const uint8_t expected_preamble_ints = num_levels > 1 ? 4 : 2; |
| if (preamble_ints != expected_preamble_ints) { |
| throw std::invalid_argument("Possible corruption: preamble ints must be " |
| + std::to_string(expected_preamble_ints) + ", got " + std::to_string(preamble_ints)); |
| } |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::check_serial_version(uint8_t serial_version) { |
| if (serial_version != SERIAL_VERSION) { |
| throw std::invalid_argument("Possible corruption: serial version mismatch: expected " |
| + std::to_string(SERIAL_VERSION) |
| + ", got " + std::to_string(serial_version)); |
| } |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| void req_sketch<T, C, S, A>::check_family_id(uint8_t family_id) { |
| if (family_id != FAMILY) { |
| throw std::invalid_argument("Possible corruption: family mismatch: expected " |
| + std::to_string(FAMILY) + ", got " + std::to_string(family_id)); |
| } |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| auto req_sketch<T, C, S, A>::begin() const -> const_iterator { |
| return const_iterator(compactors_.begin(), compactors_.end()); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| auto req_sketch<T, C, S, A>::end() const -> const_iterator { |
| return const_iterator(compactors_.end(), compactors_.end()); |
| } |
| |
| // iterator |
| |
| template<typename T, typename C, typename S, typename A> |
| req_sketch<T, C, S, A>::const_iterator::const_iterator(LevelsIterator begin, LevelsIterator end): |
| levels_it_(begin), |
| levels_end_(end), |
| compactor_it_((*levels_it_).begin()) |
| {} |
| |
| template<typename T, typename C, typename S, typename A> |
| auto req_sketch<T, C, S, A>::const_iterator::operator++() -> const_iterator& { |
| ++compactor_it_; |
| if (compactor_it_ == (*levels_it_).end()) { |
| ++levels_it_; |
| if (levels_it_ != levels_end_) compactor_it_ = (*levels_it_).begin(); |
| } |
| return *this; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| auto req_sketch<T, C, S, A>::const_iterator::operator++(int) -> const_iterator& { |
| const_iterator tmp(*this); |
| operator++(); |
| return tmp; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| bool req_sketch<T, C, S, A>::const_iterator::operator==(const const_iterator& other) const { |
| if (levels_it_ != other.levels_it_) return false; |
| if (levels_it_ == levels_end_) return true; |
| return compactor_it_ == other.compactor_it_; |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| bool req_sketch<T, C, S, A>::const_iterator::operator!=(const const_iterator& other) const { |
| return !operator==(other); |
| } |
| |
| template<typename T, typename C, typename S, typename A> |
| std::pair<const T&, const uint64_t> req_sketch<T, C, S, A>::const_iterator::operator*() const { |
| return std::pair<const T&, const uint64_t>(*compactor_it_, 1 << (*levels_it_).get_lg_weight()); |
| } |
| |
| } /* namespace datasketches */ |
| |
| #endif |