blob: 558c13c554688e82d46513417806a5fe05eead26 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing,
* software distributed under the License is distributed on an
* "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
* KIND, either express or implied. See the License for the
* specific language governing permissions and limitations
* under the License.
*/
#ifndef _QUANTILES_SKETCH_IMPL_HPP_
#define _QUANTILES_SKETCH_IMPL_HPP_
#include <cmath>
#include <algorithm>
#include <stdexcept>
#include <iomanip>
#include <sstream>
#include "count_zeros.hpp"
#include "conditional_forward.hpp"
namespace datasketches {
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>::quantiles_sketch(uint16_t k, const C& comparator, const A& allocator):
comparator_(comparator),
allocator_(allocator),
is_base_buffer_sorted_(true),
k_(k),
n_(0),
bit_pattern_(0),
base_buffer_(allocator_),
levels_(allocator_),
min_item_(),
max_item_(),
sorted_view_(nullptr)
{
check_k(k_);
base_buffer_.reserve(2 * std::min(quantiles_constants::MIN_K, k));
}
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>::quantiles_sketch(const quantiles_sketch& other):
comparator_(other.comparator_),
allocator_(other.allocator_),
is_base_buffer_sorted_(other.is_base_buffer_sorted_),
k_(other.k_),
n_(other.n_),
bit_pattern_(other.bit_pattern_),
base_buffer_(other.base_buffer_),
levels_(other.levels_),
min_item_(other.min_item_),
max_item_(other.max_item_),
sorted_view_(nullptr)
{
for (size_t i = 0; i < levels_.size(); ++i) {
if (levels_[i].capacity() != other.levels_[i].capacity()) {
levels_[i].reserve(other.levels_[i].capacity());
}
}
}
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>::quantiles_sketch(quantiles_sketch&& other) noexcept:
comparator_(other.comparator_),
allocator_(other.allocator_),
is_base_buffer_sorted_(other.is_base_buffer_sorted_),
k_(other.k_),
n_(other.n_),
bit_pattern_(other.bit_pattern_),
base_buffer_(std::move(other.base_buffer_)),
levels_(std::move(other.levels_)),
min_item_(std::move(other.min_item_)),
max_item_(std::move(other.max_item_)),
sorted_view_(nullptr)
{}
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>& quantiles_sketch<T, C, A>::operator=(const quantiles_sketch& other) {
quantiles_sketch<T, C, A> copy(other);
std::swap(comparator_, copy.comparator_);
std::swap(allocator_, copy.allocator_);
std::swap(is_base_buffer_sorted_, copy.is_base_buffer_sorted_);
std::swap(k_, copy.k_);
std::swap(n_, copy.n_);
std::swap(bit_pattern_, copy.bit_pattern_);
std::swap(base_buffer_, copy.base_buffer_);
std::swap(levels_, copy.levels_);
std::swap(min_item_, copy.min_item_);
std::swap(max_item_, copy.max_item_);
reset_sorted_view();
return *this;
}
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>& quantiles_sketch<T, C, A>::operator=(quantiles_sketch&& other) noexcept {
std::swap(comparator_, other.comparator_);
std::swap(allocator_, other.allocator_);
std::swap(is_base_buffer_sorted_, other.is_base_buffer_sorted_);
std::swap(k_, other.k_);
std::swap(n_, other.n_);
std::swap(bit_pattern_, other.bit_pattern_);
std::swap(base_buffer_, other.base_buffer_);
std::swap(levels_, other.levels_);
std::swap(min_item_, other.min_item_);
std::swap(max_item_, other.max_item_);
reset_sorted_view();
return *this;
}
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>::quantiles_sketch(uint16_t k, uint64_t n, uint64_t bit_pattern,
Level&& base_buffer, VectorLevels&& levels,
optional<T>&& min_item, optional<T>&& max_item,
bool is_sorted, const C& comparator, const A& allocator):
comparator_(comparator),
allocator_(allocator),
is_base_buffer_sorted_(is_sorted),
k_(k),
n_(n),
bit_pattern_(bit_pattern),
base_buffer_(std::move(base_buffer)),
levels_(std::move(levels)),
min_item_(std::move(min_item)),
max_item_(std::move(max_item)),
sorted_view_(nullptr)
{
uint32_t item_count = static_cast<uint32_t>(base_buffer_.size());
for (Level& lvl : levels_) {
item_count += static_cast<uint32_t>(lvl.size());
}
if (item_count != compute_retained_items(k_, n_))
throw std::logic_error("Item count does not match value computed from k, n");
}
template<typename T, typename C, typename A>
template<typename From, typename FC, typename FA>
quantiles_sketch<T, C, A>::quantiles_sketch(const quantiles_sketch<From, FC, FA>& other,
const C& comparator, const A& allocator):
comparator_(comparator),
allocator_(allocator),
is_base_buffer_sorted_(false),
k_(other.get_k()),
n_(other.get_n()),
bit_pattern_(compute_bit_pattern(other.get_k(), other.get_n())),
base_buffer_(allocator),
levels_(allocator),
min_item_(other.min_item_),
max_item_(other.max_item_),
sorted_view_(nullptr)
{
static_assert(std::is_constructible<T, From>::value,
"Type converting constructor requires new type to be constructible from existing type");
base_buffer_.reserve(2 * std::min(quantiles_constants::MIN_K, k_));
if (!other.is_empty()) {
// reserve space in levels
const uint8_t num_levels = compute_levels_needed(k_, n_);
levels_.reserve(num_levels);
for (int i = 0; i < num_levels; ++i) {
Level level(allocator);
level.reserve(k_);
levels_.push_back(std::move(level));
}
// iterate through points, assigning to the correct level as needed
for (auto pair : other) {
const uint64_t wt = pair.second;
if (wt == 1) {
base_buffer_.push_back(T(pair.first));
// resize where needed as if adding points via update()
if (base_buffer_.size() + 1 > base_buffer_.capacity()) {
const size_t new_size = std::max(std::min(static_cast<size_t>(2 * k_), 2 * base_buffer_.size()), static_cast<size_t>(1));
base_buffer_.reserve(new_size);
}
}
else {
const uint8_t idx = count_trailing_zeros_in_u64(pair.second) - 1;
levels_[idx].push_back(T(pair.first));
}
}
// validate that ordering within each level is preserved
// base_buffer_ can be considered unsorted for this purpose
for (int i = 0; i < num_levels; ++i) {
if (!std::is_sorted(levels_[i].begin(), levels_[i].end(), comparator_)) {
throw std::logic_error("Copy construction across types produces invalid sorting");
}
}
}
}
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>::~quantiles_sketch() {
reset_sorted_view();
}
template<typename T, typename C, typename A>
template<typename FwdT>
void quantiles_sketch<T, C, A>::update(FwdT&& item) {
if (!check_update_item(item)) { return; }
if (is_empty()) {
min_item_.emplace(item);
max_item_.emplace(item);
} else {
if (comparator_(item, *min_item_)) *min_item_ = item;
if (comparator_(*max_item_, item)) *max_item_ = item;
}
// if exceed capacity, grow until size 2k -- assumes eager processing
if (base_buffer_.size() + 1 > base_buffer_.capacity()) grow_base_buffer();
base_buffer_.push_back(std::forward<FwdT>(item));
++n_;
if (base_buffer_.size() > 1) is_base_buffer_sorted_ = false;
if (base_buffer_.size() == 2 * k_) process_full_base_buffer();
reset_sorted_view();
}
template<typename T, typename C, typename A>
template<typename FwdSk>
void quantiles_sketch<T, C, A>::merge(FwdSk&& other) {
if (other.is_empty()) {
return; // nothing to do
} else if (!other.is_estimation_mode()) {
// other is exact, stream in regardless of k
for (auto item : other.base_buffer_) {
update(conditional_forward<FwdSk>(item));
}
reset_sorted_view();
return;
}
// other has data and is in estimation mode
if (is_estimation_mode()) {
if (k_ == other.get_k()) {
standard_merge(*this, std::forward<FwdSk>(other));
} else if (k_ > other.get_k()) {
quantiles_sketch sk_copy(std::forward<FwdSk>(other));
downsampling_merge(sk_copy, std::move(*this));
*this = std::move(sk_copy);
} else { // k_ < other.get_k()
downsampling_merge(*this, std::forward<FwdSk>(other));
}
} else {
// exact or empty
quantiles_sketch sk_copy(std::forward<FwdSk>(other));
if (k_ <= other.get_k()) {
if (!is_empty()) {
for (uint16_t i = 0; i < base_buffer_.size(); ++i) {
sk_copy.update(std::move(base_buffer_[i]));
}
}
} else { // k_ > other.get_k()
downsampling_merge(sk_copy, std::move(*this));
}
*this = std::move(sk_copy);
}
reset_sorted_view();
}
template<typename T, typename C, typename A>
template<typename SerDe>
void quantiles_sketch<T, C, A>::serialize(std::ostream& os, const SerDe& serde) const {
const uint8_t preamble_longs = is_empty() ? PREAMBLE_LONGS_SHORT : PREAMBLE_LONGS_FULL;
write(os, preamble_longs);
const uint8_t ser_ver = SERIAL_VERSION;
write(os, ser_ver);
const uint8_t family = FAMILY;
write(os, family);
// side-effect: sort base buffer since always compact
std::sort(const_cast<Level&>(base_buffer_).begin(), const_cast<Level&>(base_buffer_).end(), comparator_);
const_cast<quantiles_sketch*>(this)->is_base_buffer_sorted_ = true;
// empty, ordered, compact are valid flags
const uint8_t flags_byte(
(is_empty() ? 1 << flags::IS_EMPTY : 0)
| (1 << flags::IS_SORTED) // always sorted as side effect noted above
| (1 << flags::IS_COMPACT) // always compact -- could be optional for numeric types?
);
write(os, flags_byte);
write(os, k_);
const uint16_t unused = 0;
write(os, unused);
if (!is_empty()) {
write(os, n_);
// min and max
serde.serialize(os, &*min_item_, 1);
serde.serialize(os, &*max_item_, 1);
// base buffer items
serde.serialize(os, base_buffer_.data(), static_cast<unsigned>(base_buffer_.size()));
// levels, only when data is present
for (Level lvl : levels_) {
if (lvl.size() > 0)
serde.serialize(os, lvl.data(), static_cast<unsigned>(lvl.size()));
}
}
}
template<typename T, typename C, typename A>
template<typename SerDe>
auto quantiles_sketch<T, C, A>::serialize(unsigned header_size_bytes, const SerDe& serde) const -> vector_bytes {
const size_t size = get_serialized_size_bytes(serde) + header_size_bytes;
vector_bytes bytes(size, 0, allocator_);
uint8_t* ptr = bytes.data() + header_size_bytes;
const uint8_t* end_ptr = ptr + size;
const uint8_t preamble_longs = is_empty() ? PREAMBLE_LONGS_SHORT : PREAMBLE_LONGS_FULL;
ptr += copy_to_mem(preamble_longs, ptr);
const uint8_t ser_ver = SERIAL_VERSION;
ptr += copy_to_mem(ser_ver, ptr);
const uint8_t family = FAMILY;
ptr += copy_to_mem(family, ptr);
// side-effect: sort base buffer since always compact
std::sort(const_cast<Level&>(base_buffer_).begin(), const_cast<Level&>(base_buffer_).end(), comparator_);
const_cast<quantiles_sketch*>(this)->is_base_buffer_sorted_ = true;
// empty, ordered, compact are valid flags
const uint8_t flags_byte(
(is_empty() ? 1 << flags::IS_EMPTY : 0)
| (1 << flags::IS_SORTED) // always sorted as side effect noted above
| (1 << flags::IS_COMPACT) // always compact
);
ptr += copy_to_mem(flags_byte, ptr);
ptr += copy_to_mem(k_, ptr);
ptr += sizeof(uint16_t); // 2 unused bytes
if (!is_empty()) {
ptr += copy_to_mem(n_, ptr);
// min and max
ptr += serde.serialize(ptr, end_ptr - ptr, &*min_item_, 1);
ptr += serde.serialize(ptr, end_ptr - ptr, &*max_item_, 1);
// base buffer items
if (base_buffer_.size() > 0)
ptr += serde.serialize(ptr, end_ptr - ptr, base_buffer_.data(), static_cast<unsigned>(base_buffer_.size()));
// levels, only when data is present
for (Level lvl : levels_) {
if (lvl.size() > 0)
ptr += serde.serialize(ptr, end_ptr - ptr, lvl.data(), static_cast<unsigned>(lvl.size()));
}
}
return bytes;
}
template<typename T, typename C, typename A>
template<typename SerDe>
auto quantiles_sketch<T, C, A>::deserialize(std::istream &is, const SerDe& serde,
const C& comparator, const A &allocator) -> quantiles_sketch {
const auto preamble_longs = read<uint8_t>(is);
const auto serial_version = read<uint8_t>(is);
const auto family_id = read<uint8_t>(is);
const auto flags_byte = read<uint8_t>(is);
const auto k = read<uint16_t>(is);
read<uint16_t>(is); // unused
check_k(k);
check_serial_version(serial_version); // a little redundant with the header check
check_family_id(family_id);
check_header_validity(preamble_longs, flags_byte, serial_version);
if (!is.good()) throw std::runtime_error("error reading from std::istream");
const bool is_empty = (flags_byte & (1 << flags::IS_EMPTY)) > 0;
if (is_empty) {
return quantiles_sketch(k, comparator, allocator);
}
const auto items_seen = read<uint64_t>(is);
const bool is_compact = (serial_version == 2) | ((flags_byte & (1 << flags::IS_COMPACT)) > 0);
const bool is_sorted = (flags_byte & (1 << flags::IS_SORTED)) > 0;
optional<T> tmp; // space to deserialize min and max
optional<T> min_item;
optional<T> max_item;
serde.deserialize(is, &*tmp, 1);
// serde call did not throw, repackage and cleanup
min_item.emplace(*tmp);
(*tmp).~T();
serde.deserialize(is, &*tmp, 1);
// serde call did not throw, repackage and cleanup
max_item.emplace(*tmp);
(*tmp).~T();
if (serial_version == 1) {
read<uint64_t>(is); // no longer used
}
// allocate buffers as needed
const uint8_t levels_needed = compute_levels_needed(k, items_seen);
const uint64_t bit_pattern = compute_bit_pattern(k, items_seen);
// Java provides a compact storage layout for a sketch of primitive doubles. The C++ version
// does not currently operate sketches in compact mode, but will only serialize as compact
// to avoid complications around serialization of empty values for generic type T. We also need
// to be able to ingest either serialized format from Java.
// load base buffer
const uint32_t bb_items = compute_base_buffer_items(k, items_seen);
uint32_t items_to_read = (levels_needed == 0 || is_compact) ? bb_items : 2 * k;
Level base_buffer = deserialize_array(is, bb_items, 2 * k, serde, allocator);
if (items_to_read > bb_items) { // either equal or greater, never read fewer items
// read remaining items, but don't store them
deserialize_array(is, items_to_read - bb_items, items_to_read - bb_items, serde, allocator);
}
// populate vector of Levels directly
VectorLevels levels(allocator);
levels.reserve(levels_needed);
if (levels_needed > 0) {
uint64_t working_pattern = bit_pattern;
for (size_t i = 0; i < levels_needed; ++i, working_pattern >>= 1) {
if ((working_pattern & 0x01) == 1) {
Level level = deserialize_array(is, k, k, serde, allocator);
levels.push_back(std::move(level));
} else {
Level level(allocator);
level.reserve(k);
levels.push_back(std::move(level));
}
}
}
return quantiles_sketch(k, items_seen, bit_pattern,
std::move(base_buffer), std::move(levels), std::move(min_item), std::move(max_item), is_sorted,
comparator, allocator);
}
template<typename T, typename C, typename A>
template<typename SerDe>
auto quantiles_sketch<T, C, A>::deserialize_array(std::istream& is, uint32_t num_items, uint32_t capacity, const SerDe& serde, const A& allocator) -> Level {
A alloc(allocator);
std::unique_ptr<T, items_deleter> items(alloc.allocate(num_items), items_deleter(allocator, false, num_items));
serde.deserialize(is, items.get(), num_items);
// serde did not throw, enable destructors
items.get_deleter().set_destroy(true);
if (!is.good()) throw std::runtime_error("error reading from std::istream");
// successfully read, now put into a Level
Level level(allocator);
level.reserve(capacity);
level.insert(level.begin(),
std::make_move_iterator(items.get()),
std::make_move_iterator(items.get() + num_items));
return level;
}
template<typename T, typename C, typename A>
template<typename SerDe>
auto quantiles_sketch<T, C, A>::deserialize(const void* bytes, size_t size, const SerDe& serde,
const C& comparator, const A &allocator) -> quantiles_sketch {
ensure_minimum_memory(size, 8);
const char* ptr = static_cast<const char*>(bytes);
const char* end_ptr = static_cast<const char*>(bytes) + size;
uint8_t preamble_longs;
ptr += copy_from_mem(ptr, preamble_longs);
uint8_t serial_version;
ptr += copy_from_mem(ptr, serial_version);
uint8_t family_id;
ptr += copy_from_mem(ptr, family_id);
uint8_t flags_byte;
ptr += copy_from_mem(ptr, flags_byte);
uint16_t k;
ptr += copy_from_mem(ptr, k);
uint16_t unused;
ptr += copy_from_mem(ptr, unused);
check_k(k);
check_serial_version(serial_version); // a little redundant with the header check
check_family_id(family_id);
check_header_validity(preamble_longs, flags_byte, serial_version);
const bool is_empty = (flags_byte & (1 << flags::IS_EMPTY)) > 0;
if (is_empty) {
return quantiles_sketch(k, comparator, allocator);
}
ensure_minimum_memory(size, 16);
uint64_t items_seen;
ptr += copy_from_mem(ptr, items_seen);
const bool is_compact = (serial_version == 2) | ((flags_byte & (1 << flags::IS_COMPACT)) > 0);
const bool is_sorted = (flags_byte & (1 << flags::IS_SORTED)) > 0;
optional<T> tmp; // space to deserialize min and max
optional<T> min_item;
optional<T> max_item;
ptr += serde.deserialize(ptr, end_ptr - ptr, &*tmp, 1);
// serde call did not throw, repackage and cleanup
min_item.emplace(*tmp);
(*tmp).~T();
ptr += serde.deserialize(ptr, end_ptr - ptr, &*tmp, 1);
// serde call did not throw, repackage and cleanup
max_item.emplace(*tmp);
(*tmp).~T();
if (serial_version == 1) {
uint64_t unused_long;
ptr += copy_from_mem(ptr, unused_long); // no longer used
}
// allocate buffers as needed
const uint8_t levels_needed = compute_levels_needed(k, items_seen);
const uint64_t bit_pattern = compute_bit_pattern(k, items_seen);
// Java provides a compact storage layout for a sketch of primitive doubles. The C++ version
// does not currently operate sketches in compact mode, but will only serialize as compact
// to avoid complications around serialization of empty values for generic type T. We also need
// to be able to ingest either serialized format from Java.
// load base buffer
const uint32_t bb_items = compute_base_buffer_items(k, items_seen);
uint32_t items_to_read = (levels_needed == 0 || is_compact) ? bb_items : 2 * k;
auto base_buffer_pair = deserialize_array(ptr, end_ptr - ptr, bb_items, 2 * k, serde, allocator);
ptr += base_buffer_pair.second;
if (items_to_read > bb_items) { // either equal or greater, never read fewer items
// read remaining items, only use to advance the pointer
auto extras = deserialize_array(ptr, end_ptr - ptr, items_to_read - bb_items, items_to_read - bb_items, serde, allocator);
ptr += extras.second;
}
// populate vector of Levels directly
VectorLevels levels(allocator);
levels.reserve(levels_needed);
if (levels_needed > 0) {
uint64_t working_pattern = bit_pattern;
for (size_t i = 0; i < levels_needed; ++i, working_pattern >>= 1) {
if ((working_pattern & 0x01) == 1) {
auto pair = deserialize_array(ptr, end_ptr - ptr, k, k, serde, allocator);
ptr += pair.second;
levels.push_back(std::move(pair.first));
} else {
Level level(allocator);
level.reserve(k);
levels.push_back(std::move(level));
}
}
}
return quantiles_sketch(k, items_seen, bit_pattern,
std::move(base_buffer_pair.first), std::move(levels), std::move(min_item), std::move(max_item), is_sorted,
comparator, allocator);
}
template<typename T, typename C, typename A>
template<typename SerDe>
auto quantiles_sketch<T, C, A>::deserialize_array(const void* bytes, size_t size, uint32_t num_items, uint32_t capacity, const SerDe& serde, const A& allocator)
-> std::pair<Level, size_t> {
const char* ptr = static_cast<const char*>(bytes);
const char* end_ptr = static_cast<const char*>(bytes) + size;
A alloc(allocator);
std::unique_ptr<T, items_deleter> items(alloc.allocate(num_items), items_deleter(allocator, false, num_items));
ptr += serde.deserialize(ptr, end_ptr - ptr, items.get(), num_items);
// serde did not throw, enable destructors
items.get_deleter().set_destroy(true);
// succesfully read, now put into a Level
Level level(allocator);
level.reserve(capacity);
level.insert(level.begin(),
std::make_move_iterator(items.get()),
std::make_move_iterator(items.get() + num_items));
return std::pair<Level, size_t>(std::move(level), ptr - static_cast<const char*>(bytes));
}
template<typename T, typename C, typename A>
string<A> quantiles_sketch<T, C, A>::to_string(bool print_levels, bool print_items) const {
// Using a temporary stream for implementation here does not comply with AllocatorAwareContainer requirements.
// The stream does not support passing an allocator instance, and alternatives are complicated.
std::ostringstream os;
os << "### Quantiles Sketch summary:" << std::endl;
os << " K : " << k_ << std::endl;
os << " N : " << n_ << std::endl;
os << " Epsilon : " << std::setprecision(3) << get_normalized_rank_error(false) * 100 << "%" << std::endl;
os << " Epsilon PMF : " << get_normalized_rank_error(true) * 100 << "%" << std::endl;
os << " Empty : " << (is_empty() ? "true" : "false") << std::endl;
os << " Estimation mode: " << (is_estimation_mode() ? "true" : "false") << std::endl;
os << " Levels (w/o BB): " << levels_.size() << std::endl;
os << " Used Levels : " << count_valid_levels(bit_pattern_) << std::endl;
os << " Retained items : " << get_num_retained() << std::endl;
if (!is_empty()) {
os << " Min item : " << *min_item_ << std::endl;
os << " Max item : " << *max_item_ << std::endl;
}
os << "### End sketch summary" << std::endl;
if (print_levels) {
os << "### Quantiles Sketch levels:" << std::endl;
os << " index: items in use" << std::endl;
os << " BB: " << base_buffer_.size() << std::endl;
for (uint8_t i = 0; i < levels_.size(); i++) {
os << " " << static_cast<unsigned int>(i) << ": " << levels_[i].size() << std::endl;
}
os << "### End sketch levels" << std::endl;
}
if (print_items) {
os << "### Quantiles Sketch data:" << std::endl;
uint8_t level = 0;
os << " BB:" << std::endl;
for (const T& item : base_buffer_) {
os << " " << item << std::endl;
}
for (uint8_t i = 0; i < levels_.size(); ++i) {
os << " level " << static_cast<unsigned int>(level) << ":" << std::endl;
for (const T& item : levels_[i]) {
os << " " << item << std::endl;
}
}
os << "### End sketch data" << std::endl;
}
return string<A>(os.str().c_str(), allocator_);
}
template<typename T, typename C, typename A>
uint16_t quantiles_sketch<T, C, A>::get_k() const {
return k_;
}
template<typename T, typename C, typename A>
uint64_t quantiles_sketch<T, C, A>::get_n() const {
return n_;
}
template<typename T, typename C, typename A>
bool quantiles_sketch<T, C, A>::is_empty() const {
return n_ == 0;
}
template<typename T, typename C, typename A>
bool quantiles_sketch<T, C, A>::is_estimation_mode() const {
return bit_pattern_ != 0;
}
template<typename T, typename C, typename A>
uint32_t quantiles_sketch<T, C, A>::get_num_retained() const {
return compute_retained_items(k_, n_);
}
template<typename T, typename C, typename A>
const T& quantiles_sketch<T, C, A>::get_min_item() const {
if (is_empty()) throw std::runtime_error("operation is undefined for an empty sketch");
return *min_item_;
}
template<typename T, typename C, typename A>
const T& quantiles_sketch<T, C, A>::get_max_item() const {
if (is_empty()) throw std::runtime_error("operation is undefined for an empty sketch");
return *max_item_;
}
template<typename T, typename C, typename A>
C quantiles_sketch<T, C, A>::get_comparator() const {
return comparator_;
}
template<typename T, typename C, typename A>
A quantiles_sketch<T, C, A>::get_allocator() const {
return allocator_;
}
// implementation for fixed-size arithmetic types (integral and floating point)
template<typename T, typename C, typename A>
template<typename SerDe, typename TT, typename std::enable_if<std::is_arithmetic<TT>::value, int>::type>
size_t quantiles_sketch<T, C, A>::get_serialized_size_bytes(const SerDe&) const {
if (is_empty()) { return EMPTY_SIZE_BYTES; }
return DATA_START + ((get_num_retained() + 2) * sizeof(TT));
}
// implementation for all other types
template<typename T, typename C, typename A>
template<typename SerDe, typename TT, typename std::enable_if<!std::is_arithmetic<TT>::value, int>::type>
size_t quantiles_sketch<T, C, A>::get_serialized_size_bytes(const SerDe& serde) const {
if (is_empty()) { return EMPTY_SIZE_BYTES; }
size_t size = DATA_START;
size += serde.size_of_item(*min_item_);
size += serde.size_of_item(*max_item_);
for (auto it: *this) size += serde.size_of_item(it.first);
return size;
}
template<typename T, typename C, typename A>
double quantiles_sketch<T, C, A>::get_normalized_rank_error(bool is_pmf) const {
return get_normalized_rank_error(k_, is_pmf);
}
template<typename T, typename C, typename A>
double quantiles_sketch<T, C, A>::get_normalized_rank_error(uint16_t k, bool is_pmf) {
return is_pmf
? 1.854 / std::pow(k, 0.9657)
: 1.576 / std::pow(k, 0.9726);
}
template<typename T, typename C, typename A>
quantiles_sorted_view<T, C, A> quantiles_sketch<T, C, A>::get_sorted_view() const {
// allow side-effect of sorting the base buffer
if (!is_base_buffer_sorted_) {
std::sort(const_cast<Level&>(base_buffer_).begin(), const_cast<Level&>(base_buffer_).end(), comparator_);
const_cast<quantiles_sketch*>(this)->is_base_buffer_sorted_ = true;
}
quantiles_sorted_view<T, C, A> view(get_num_retained(), comparator_, allocator_);
uint64_t weight = 1;
view.add(base_buffer_.begin(), base_buffer_.end(), weight);
for (const auto& level: levels_) {
weight <<= 1;
if (level.empty()) { continue; }
view.add(level.begin(), level.end(), weight);
}
view.convert_to_cummulative();
return view;
}
template<typename T, typename C, typename A>
auto quantiles_sketch<T, C, A>::get_quantile(double rank, bool inclusive) const -> quantile_return_type {
if (is_empty()) throw std::runtime_error("operation is undefined for an empty sketch");
if ((rank < 0.0) || (rank > 1.0)) {
throw std::invalid_argument("Normalized rank cannot be less than 0 or greater than 1");
}
// possible side-effect: sorting base buffer
setup_sorted_view();
return sorted_view_->get_quantile(rank, inclusive);
}
template<typename T, typename C, typename A>
double quantiles_sketch<T, C, A>::get_rank(const T& item, bool inclusive) const {
if (is_empty()) throw std::runtime_error("operation is undefined for an empty sketch");
setup_sorted_view();
return sorted_view_->get_rank(item, inclusive);
}
template<typename T, typename C, typename A>
auto quantiles_sketch<T, C, A>::get_PMF(const T* split_points, uint32_t size, bool inclusive) const -> vector_double {
if (is_empty()) throw std::runtime_error("operation is undefined for an empty sketch");
setup_sorted_view();
return sorted_view_->get_PMF(split_points, size, inclusive);
}
template<typename T, typename C, typename A>
auto quantiles_sketch<T, C, A>::get_CDF(const T* split_points, uint32_t size, bool inclusive) const -> vector_double {
if (is_empty()) throw std::runtime_error("operation is undefined for an empty sketch");
setup_sorted_view();
return sorted_view_->get_CDF(split_points, size, inclusive);
}
template<typename T, typename C, typename A>
uint32_t quantiles_sketch<T, C, A>::compute_retained_items(uint16_t k, uint64_t n) {
const uint32_t bb_count = compute_base_buffer_items(k, n);
const uint64_t bit_pattern = compute_bit_pattern(k, n);
const uint32_t valid_levels = count_valid_levels(bit_pattern);
return bb_count + (k * valid_levels);
}
template<typename T, typename C, typename A>
uint32_t quantiles_sketch<T, C, A>::compute_base_buffer_items(uint16_t k, uint64_t n) {
return n % (static_cast<uint64_t>(2) * k);
}
template<typename T, typename C, typename A>
uint64_t quantiles_sketch<T, C, A>::compute_bit_pattern(uint16_t k, uint64_t n) {
return n / (static_cast<uint64_t>(2) * k);
}
template<typename T, typename C, typename A>
uint32_t quantiles_sketch<T, C, A>::count_valid_levels(uint64_t bit_pattern) {
uint32_t count = 0;
for (; bit_pattern > 0; ++count) bit_pattern &= bit_pattern - 1;
return count;
}
template<typename T, typename C, typename A>
uint8_t quantiles_sketch<T, C, A>::compute_levels_needed(uint16_t k, uint64_t n) {
return static_cast<uint8_t>(64U) - count_leading_zeros_in_u64(n / (2 * k));
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::check_k(uint16_t k) {
if (k < quantiles_constants::MIN_K || k > quantiles_constants::MAX_K || (k & (k - 1)) != 0) {
throw std::invalid_argument("k must be a power of 2 that is >= "
+ std::to_string(quantiles_constants::MIN_K) + " and <= "
+ std::to_string(quantiles_constants::MAX_K) + ". Found: " + std::to_string(k));
}
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::check_serial_version(uint8_t serial_version) {
if (serial_version == SERIAL_VERSION || serial_version == SERIAL_VERSION_1 || serial_version == SERIAL_VERSION_2)
return;
else
throw std::invalid_argument("Possible corruption. Unrecognized serialization version: " + std::to_string(serial_version));
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::check_family_id(uint8_t family_id) {
if (family_id == FAMILY)
return;
else
throw std::invalid_argument("Possible corruption. Family id does not indicate quantiles sketch: " + std::to_string(family_id));
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::check_header_validity(uint8_t preamble_longs, uint8_t flags_byte, uint8_t serial_version) {
const bool empty = (flags_byte & (1 << flags::IS_EMPTY)) > 0;
const bool compact = (flags_byte & (1 << flags::IS_COMPACT)) > 0;
const uint8_t sw = (compact ? 1 : 0) + (2 * (empty ? 1 : 0))
+ (4 * (serial_version & 0xF)) + (32 * (preamble_longs & 0x3F));
bool valid = true;
switch (sw) { // exhaustive list and description of all valid cases
case 38 : break; //!compact, empty, serVer = 1, preLongs = 1; always stored as not compact
case 164 : break; //!compact, !empty, serVer = 1, preLongs = 5; always stored as not compact
case 42 : break; //!compact, empty, serVer = 2, preLongs = 1; always stored as compact
case 72 : break; //!compact, !empty, serVer = 2, preLongs = 2; always stored as compact
case 47 : break; // compact, empty, serVer = 3, preLongs = 1;
case 46 : break; //!compact, empty, serVer = 3, preLongs = 1;
case 79 : break; // compact, empty, serVer = 3, preLongs = 2;
case 78 : break; //!compact, empty, serVer = 3, preLongs = 2;
case 77 : break; // compact, !empty, serVer = 3, preLongs = 2;
case 76 : break; //!compact, !empty, serVer = 3, preLongs = 2;
default : //all other case values are invalid
valid = false;
}
if (!valid) {
std::ostringstream os;
os << "Possible sketch corruption. Inconsistent state: "
<< "preamble_longs = " << preamble_longs
<< ", empty = " << (empty ? "true" : "false")
<< ", serialization_version = " << serial_version
<< ", compact = " << (compact ? "true" : "false");
throw std::invalid_argument(os.str());
}
}
template <typename T, typename C, typename A>
typename quantiles_sketch<T, C, A>::const_iterator quantiles_sketch<T, C, A>::begin() const {
return quantiles_sketch<T, C, A>::const_iterator(base_buffer_, levels_, k_, n_, false);
}
template <typename T, typename C, typename A>
typename quantiles_sketch<T, C, A>::const_iterator quantiles_sketch<T, C, A>::end() const {
return quantiles_sketch<T, C, A>::const_iterator(base_buffer_, levels_, k_, n_, true);
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::grow_base_buffer() {
const size_t new_size = std::max(std::min(static_cast<size_t>(2 * k_), 2 * base_buffer_.size()), static_cast<size_t>(1));
base_buffer_.reserve(new_size);
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::process_full_base_buffer() {
// make sure there will be enough levels for the propagation
grow_levels_if_needed(); // note: n_ was already incremented by update() before this
std::sort(base_buffer_.begin(), base_buffer_.end(), comparator_);
in_place_propagate_carry(0,
levels_[0], // unused here, but 0 is guaranteed to exist
base_buffer_,
true, *this);
base_buffer_.clear();
is_base_buffer_sorted_ = true;
if (n_ / (2 * k_) != bit_pattern_) {
throw std::logic_error("Internal error: n / 2k (" + std::to_string(n_ / 2 * k_)
+ " != bit_pattern " + std::to_string(bit_pattern_));
}
}
template<typename T, typename C, typename A>
bool quantiles_sketch<T, C, A>::grow_levels_if_needed() {
const uint8_t levels_needed = compute_levels_needed(k_, n_);
if (levels_needed == 0)
return false; // don't need levels and might have small base buffer. Possible during merges.
// from here on, assume full size base buffer (2k) and at least one additional level
if (levels_needed <= levels_.size())
return false;
Level empty_level(allocator_);
empty_level.reserve(k_);
levels_.push_back(std::move(empty_level));
return true;
}
template<typename T, typename C, typename A>
template<typename FwdV>
void quantiles_sketch<T, C, A>::in_place_propagate_carry(uint8_t starting_level,
FwdV&& buf_size_k, Level& buf_size_2k,
bool apply_as_update,
quantiles_sketch& sketch) {
const uint64_t bit_pattern = sketch.bit_pattern_;
const int k = sketch.k_;
uint8_t ending_level = lowest_zero_bit_starting_at(bit_pattern, starting_level);
if (apply_as_update) {
// update version of computation
// its is okay for buf_size_k to be null in this case
zip_buffer(buf_size_2k, sketch.levels_[ending_level]);
} else {
// merge_into version of computation
for (uint16_t i = 0; i < k; ++i) {
sketch.levels_[ending_level].push_back(conditional_forward<FwdV>(buf_size_k[i]));
}
}
for (uint64_t lvl = starting_level; lvl < ending_level; lvl++) {
if ((bit_pattern & (static_cast<uint64_t>(1) << lvl)) == 0) {
throw std::logic_error("unexpected empty level in bit_pattern");
}
merge_two_size_k_buffers(
sketch.levels_[lvl],
sketch.levels_[ending_level],
buf_size_2k, sketch.get_comparator());
sketch.levels_[lvl].clear();
sketch.levels_[ending_level].clear();
zip_buffer(buf_size_2k, sketch.levels_[ending_level]);
} // end of loop over lower levels
// update bit pattern with binary-arithmetic ripple carry
sketch.bit_pattern_ = bit_pattern + (static_cast<uint64_t>(1) << starting_level);
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::zip_buffer(Level& buf_in, Level& buf_out) {
#ifdef QUANTILES_VALIDATION
static uint32_t next_offset = 0;
uint32_t rand_offset = next_offset;
next_offset = 1 - next_offset;
#else
uint32_t rand_offset = random_utils::random_bit();
#endif
if ((buf_in.size() != 2 * buf_out.capacity())
|| (buf_out.size() > 0)) {
throw std::logic_error("zip_buffer requires buf_in.size() == "
"2*buf_out.capacity() and empty buf_out");
}
size_t k = buf_out.capacity();
for (uint32_t i = rand_offset, o = 0; o < k; i += 2, ++o) {
buf_out.push_back(std::move(buf_in[i]));
}
buf_in.clear();
}
template<typename T, typename C, typename A>
template<typename FwdV>
void quantiles_sketch<T, C, A>::zip_buffer_with_stride(FwdV&& buf_in, Level& buf_out, uint16_t stride) {
// Random offset in range [0, stride)
std::uniform_int_distribution<uint16_t> dist(0, stride - 1);
const uint16_t rand_offset = dist(random_utils::rand);
if ((buf_in.size() != stride * buf_out.capacity())
|| (buf_out.size() > 0)) {
throw std::logic_error("zip_buffer_with_stride requires buf_in.size() == "
"stride*buf_out.capacity() and empty buf_out");
}
const size_t k = buf_out.capacity();
for (uint16_t i = rand_offset, o = 0; o < k; i += stride, ++o) {
buf_out.push_back(conditional_forward<FwdV>(buf_in[i]));
}
// do not clear input buffer
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::merge_two_size_k_buffers(Level& src_1, Level& src_2,
Level& dst, const C& comparator) {
if (src_1.size() != src_2.size()
|| src_1.size() * 2 != dst.capacity()
|| dst.size() != 0) {
throw std::logic_error("Input invariants violated in merge_two_size_k_buffers()");
}
auto end1 = src_1.end(), end2 = src_2.end();
auto it1 = src_1.begin(), it2 = src_2.begin();
// TODO: probably actually doing copies given Level&?
while (it1 != end1 && it2 != end2) {
if (comparator(*it1, *it2)) {
dst.push_back(std::move(*it1++));
} else {
dst.push_back(std::move(*it2++));
}
}
if (it1 != end1) {
dst.insert(dst.end(), it1, end1);
} else {
if (it2 == end2) { throw std::logic_error("it2 unexpectedly already at end of range"); }
dst.insert(dst.end(), it2, end2);
}
}
template<typename T, typename C, typename A>
template<typename FwdSk>
void quantiles_sketch<T, C, A>::standard_merge(quantiles_sketch& tgt, FwdSk&& src) {
if (src.get_k() != tgt.get_k()) {
throw std::invalid_argument("src.get_k() != tgt.get_k()");
}
if (src.is_empty()) {
return;
}
uint64_t new_n = src.get_n() + tgt.get_n();
// move items from src's base buffer
for (uint16_t i = 0; i < src.base_buffer_.size(); ++i) {
tgt.update(conditional_forward<FwdSk>(src.base_buffer_[i]));
}
// check (after moving raw items) if we need to extend levels array
uint8_t levels_needed = compute_levels_needed(tgt.get_k(), new_n);
if (levels_needed > tgt.levels_.size()) {
tgt.levels_.reserve(levels_needed);
while (tgt.levels_.size() < levels_needed) {
Level empty_level(tgt.allocator_);
empty_level.reserve(tgt.get_k());
tgt.levels_.push_back(std::move(empty_level));
}
}
Level scratch_buf(tgt.allocator_);
scratch_buf.reserve(2 * tgt.get_k());
uint64_t src_pattern = src.bit_pattern_;
for (uint8_t src_lvl = 0; src_pattern != 0; ++src_lvl, src_pattern >>= 1) {
if ((src_pattern & 1) > 0) {
scratch_buf.clear();
// propagate-carry
in_place_propagate_carry(src_lvl,
src.levels_[src_lvl], scratch_buf,
false, tgt);
// update n_ at the end
}
}
tgt.n_ = new_n;
if ((tgt.get_n() / (2 * tgt.get_k())) != tgt.bit_pattern_) {
throw std::logic_error("Failed internal consistency check after standard_merge()");
}
// update min and max items
// can't just check is_empty() since min and max might not have been set if
// there were no base buffer items added via update()
if (!tgt.min_item_) {
tgt.min_item_.emplace(conditional_forward<FwdSk>(*src.min_item_));
} else {
if (tgt.comparator_(*src.min_item_, *tgt.min_item_))
*tgt.min_item_ = conditional_forward<FwdSk>(*src.min_item_);
}
if (!tgt.max_item_) {
tgt.max_item_.emplace(conditional_forward<FwdSk>(*src.max_item_));
} else {
if (tgt.comparator_(*tgt.max_item_, *src.max_item_))
*tgt.max_item_ = conditional_forward<FwdSk>(*src.max_item_);
}
}
template<typename T, typename C, typename A>
template<typename FwdSk>
void quantiles_sketch<T, C, A>::downsampling_merge(quantiles_sketch& tgt, FwdSk&& src) {
if (src.get_k() % tgt.get_k() != 0) {
throw std::invalid_argument("src.get_k() is not a multiple of tgt.get_k()");
}
if (src.is_empty()) {
return;
}
const uint16_t downsample_factor = src.get_k() / tgt.get_k();
const uint8_t lg_sample_factor = count_trailing_zeros_in_u32(downsample_factor);
const uint64_t new_n = src.get_n() + tgt.get_n();
// move items from src's base buffer
for (uint16_t i = 0; i < src.base_buffer_.size(); ++i) {
tgt.update(conditional_forward<FwdSk>(src.base_buffer_[i]));
}
// check (after moving raw items) if we need to extend levels array
const uint8_t levels_needed = compute_levels_needed(tgt.get_k(), new_n);
if (levels_needed > tgt.levels_.size()) {
tgt.levels_.reserve(levels_needed);
while (tgt.levels_.size() < levels_needed) {
Level empty_level(tgt.allocator_);
empty_level.reserve(tgt.get_k());
tgt.levels_.push_back(std::move(empty_level));
}
}
Level down_buf(tgt.allocator_);
down_buf.reserve(tgt.get_k());
Level scratch_buf(tgt.allocator_);
scratch_buf.reserve(2 * tgt.get_k());
uint64_t src_pattern = src.bit_pattern_;
for (uint8_t src_lvl = 0; src_pattern != 0; ++src_lvl, src_pattern >>= 1) {
if ((src_pattern & 1) > 0) {
down_buf.clear();
scratch_buf.clear();
// zip with stride, leaving input buffer intact
zip_buffer_with_stride(src.levels_[src_lvl], down_buf, downsample_factor);
// propagate-carry
in_place_propagate_carry(src_lvl + lg_sample_factor,
down_buf, scratch_buf,
false, tgt);
// update n_ at the end
}
}
tgt.n_ = new_n;
if ((tgt.get_n() / (2 * tgt.get_k())) != tgt.bit_pattern_) {
throw std::logic_error("Failed internal consistency check after downsampling_merge()");
}
// update min and max items
// can't just check is_empty() since min and max might not have been set if
// there were no base buffer items added via update()
if (!tgt.min_item_) {
tgt.min_item_.emplace(conditional_forward<FwdSk>(*src.min_item_));
} else {
if (tgt.comparator_(*src.min_item_, *tgt.min_item_))
*tgt.min_item_ = conditional_forward<FwdSk>(*src.min_item_);
}
if (!tgt.max_item_) {
tgt.max_item_.emplace(conditional_forward<FwdSk>(*src.max_item_));
} else {
if (tgt.comparator_(*tgt.max_item_, *src.max_item_))
*tgt.max_item_ = conditional_forward<FwdSk>(*src.max_item_);
}
}
template<typename T, typename C, typename A>
uint8_t quantiles_sketch<T, C, A>::lowest_zero_bit_starting_at(uint64_t bits, uint8_t starting_bit) {
uint8_t pos = starting_bit & 0X3F;
uint64_t my_bits = bits >> pos;
while ((my_bits & static_cast<uint64_t>(1)) != 0) {
my_bits >>= 1;
pos++;
}
return pos;
}
template<typename T, typename C, typename A>
class quantiles_sketch<T, C, A>::items_deleter {
public:
items_deleter(const A& allocator, bool destroy, size_t num): allocator_(allocator), destroy_(destroy), num_(num) {}
void operator() (T* ptr) {
if (ptr != nullptr) {
if (destroy_) {
for (size_t i = 0; i < num_; ++i) {
ptr[i].~T();
}
}
allocator_.deallocate(ptr, num_);
}
}
void set_destroy(bool destroy) { destroy_ = destroy; }
private:
A allocator_;
bool destroy_;
size_t num_;
};
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::setup_sorted_view() const {
if (sorted_view_ == nullptr) {
using AllocSortedView = typename std::allocator_traits<A>::template rebind_alloc<quantiles_sorted_view<T, C, A>>;
sorted_view_ = new (AllocSortedView(allocator_).allocate(1)) quantiles_sorted_view<T, C, A>(get_sorted_view());
}
}
template<typename T, typename C, typename A>
void quantiles_sketch<T, C, A>::reset_sorted_view() {
if (sorted_view_ != nullptr) {
sorted_view_->~quantiles_sorted_view();
using AllocSortedView = typename std::allocator_traits<A>::template rebind_alloc<quantiles_sorted_view<T, C, A>>;
AllocSortedView(allocator_).deallocate(sorted_view_, 1);
sorted_view_ = nullptr;
}
}
// quantiles_sketch::const_iterator implementation
template<typename T, typename C, typename A>
quantiles_sketch<T, C, A>::const_iterator::const_iterator(const Level& base_buffer,
const std::vector<Level, AllocLevel>& levels,
uint16_t k,
uint64_t n,
bool is_end):
base_buffer_(base_buffer),
levels_(levels),
level_(-1),
index_(0),
bb_count_(compute_base_buffer_items(k, n)),
bit_pattern_(compute_bit_pattern(k, n)),
weight_(1),
k_(k)
{
if (is_end) {
// if exact mode: index_ = n is end
// if sampling, level_ = max_level + 1 and index_ = 0 is end
if (bit_pattern_ == 0) // only a valid check for exact mode in constructor
index_ = static_cast<uint32_t>(n);
else
level_ = static_cast<int>(levels_.size());
} else { // find first non-empty item
if (bb_count_ == 0 && bit_pattern_ > 0) {
level_ = 0;
weight_ = 2;
while ((bit_pattern_ & 0x01) == 0) {
weight_ *= 2;
++level_;
bit_pattern_ >>= 1;
}
}
}
}
template<typename T, typename C, typename A>
typename quantiles_sketch<T, C, A>::const_iterator& quantiles_sketch<T, C, A>::const_iterator::operator++() {
++index_;
if ((level_ == -1 && index_ == base_buffer_.size() && levels_.size() > 0) || (level_ >= 0 && index_ == k_)) { // go to the next non-empty level
index_ = 0;
do {
++level_;
if (level_ > 0) bit_pattern_ = bit_pattern_ >> 1;
if (bit_pattern_ == 0) return *this;
weight_ *= 2;
} while ((bit_pattern_ & static_cast<uint64_t>(1)) == 0);
}
return *this;
}
template<typename T, typename C, typename A>
typename quantiles_sketch<T, C, A>::const_iterator& quantiles_sketch<T, C, A>::const_iterator::operator++(int) {
const_iterator tmp(*this);
operator++();
return tmp;
}
template<typename T, typename C, typename A>
bool quantiles_sketch<T, C, A>::const_iterator::operator==(const const_iterator& other) const {
return level_ == other.level_ && index_ == other.index_;
}
template<typename T, typename C, typename A>
bool quantiles_sketch<T, C, A>::const_iterator::operator!=(const const_iterator& other) const {
return !operator==(other);
}
template<typename T, typename C, typename A>
auto quantiles_sketch<T, C, A>::const_iterator::operator*() const -> reference {
return value_type(level_ == -1 ? base_buffer_[index_] : levels_[level_][index_], weight_);
}
template<typename T, typename C, typename A>
auto quantiles_sketch<T, C, A>::const_iterator::operator->() const -> pointer {
return **this;
}
} /* namespace datasketches */
#endif // _QUANTILES_SKETCH_IMPL_HPP_