| // Licensed to the Apache Software Foundation (ASF) under one |
| // or more contributor license agreements. See the NOTICE file |
| // distributed with this work for additional information |
| // regarding copyright ownership. The ASF licenses this file |
| // to you under the Apache License, Version 2.0 (the |
| // "License"); you may not use this file except in compliance |
| // with the License. You may obtain a copy of the License at |
| // |
| // http://www.apache.org/licenses/LICENSE-2.0 |
| // |
| // Unless required by applicable law or agreed to in writing, |
| // software distributed under the License is distributed on an |
| // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY |
| // KIND, either express or implied. See the License for the |
| // specific language governing permissions and limitations |
| // under the License. |
| |
| #include "olap/lru_cache.h" |
| |
| #include <gen_cpp/Metrics_types.h> |
| #include <gtest/gtest-message.h> |
| #include <gtest/gtest-test-part.h> |
| |
| #include <iosfwd> |
| #include <vector> |
| |
| #include "gtest/gtest_pred_impl.h" |
| #include "runtime/memory/lru_cache_policy.h" |
| #include "runtime/memory/mem_tracker_limiter.h" |
| #include "testutil/test_util.h" |
| |
| using namespace doris; |
| using namespace std; |
| |
| namespace doris { |
| |
| void PutFixed32(std::string* dst, uint32_t value) { |
| char buf[sizeof(value)]; |
| memcpy(buf, &value, sizeof(value)); |
| dst->append(buf, sizeof(buf)); |
| } |
| |
| uint32_t DecodeFixed32(const char* ptr) { |
| // Load the raw bytes |
| uint32_t result; |
| memcpy(&result, ptr, sizeof(result)); // gcc optimizes this to a plain load |
| return result; |
| } |
| |
| // Conversions between numeric keys/values and the types expected by Cache. |
| const CacheKey EncodeKey(std::string* result, int k) { |
| PutFixed32(result, k); |
| return CacheKey(result->c_str(), result->size()); |
| } |
| |
| static int DecodeKey(const CacheKey& k) { |
| EXPECT_EQ(k.size(), 4); |
| return DecodeFixed32(k.data()); |
| } |
| static void* EncodeValue(uintptr_t v) { |
| return reinterpret_cast<void*>(v); |
| } |
| static int DecodeValue(void* v) { |
| return reinterpret_cast<uintptr_t>(v); |
| } |
| |
| class CacheTest : public testing::Test { |
| public: |
| static CacheTest* _s_current; |
| |
| static void Deleter(const CacheKey& key, void* v) { |
| _s_current->_deleted_keys.push_back(DecodeKey(key)); |
| _s_current->_deleted_values.push_back(DecodeValue(v)); |
| } |
| |
| class CacheTestPolicy : public LRUCachePolicy { |
| public: |
| CacheTestPolicy(size_t capacity) |
| : LRUCachePolicy(CachePolicy::CacheType::FOR_UT, capacity, LRUCacheType::SIZE, -1) { |
| } |
| }; |
| |
| // there is 16 shards in ShardedLRUCache |
| // And the LRUHandle size is about 100B. So the cache size should big enough |
| // to run the UT. |
| static const int kCacheSize = 1000 * 16; |
| std::vector<int> _deleted_keys; |
| std::vector<int> _deleted_values; |
| CacheTestPolicy* _cache; |
| |
| CacheTest() : _cache(new CacheTestPolicy(kCacheSize)) { _s_current = this; } |
| |
| ~CacheTest() { delete _cache; } |
| |
| Cache* cache() { return _cache->cache(); } |
| |
| int Lookup(int key) { |
| std::string result; |
| Cache::Handle* handle = cache()->lookup(EncodeKey(&result, key)); |
| const int r = (handle == nullptr) ? -1 : DecodeValue(cache()->value(handle)); |
| |
| if (handle != nullptr) { |
| cache()->release(handle); |
| } |
| |
| return r; |
| } |
| |
| void Insert(int key, int value, int charge) { |
| std::string result; |
| cache()->release(cache()->insert(EncodeKey(&result, key), EncodeValue(value), charge, |
| &CacheTest::Deleter)); |
| } |
| |
| void InsertDurable(int key, int value, int charge) { |
| std::string result; |
| cache()->release(cache()->insert(EncodeKey(&result, key), EncodeValue(value), charge, |
| &CacheTest::Deleter, CachePriority::DURABLE)); |
| } |
| |
| void Erase(int key) { |
| std::string result; |
| cache()->erase(EncodeKey(&result, key)); |
| } |
| |
| void SetUp() {} |
| |
| void TearDown() {} |
| }; |
| CacheTest* CacheTest::_s_current; |
| |
| TEST_F(CacheTest, HitAndMiss) { |
| EXPECT_EQ(-1, Lookup(100)); |
| |
| Insert(100, 101, 1); |
| EXPECT_EQ(101, Lookup(100)); |
| EXPECT_EQ(-1, Lookup(200)); |
| EXPECT_EQ(-1, Lookup(300)); |
| |
| Insert(200, 201, 1); |
| EXPECT_EQ(101, Lookup(100)); |
| EXPECT_EQ(201, Lookup(200)); |
| EXPECT_EQ(-1, Lookup(300)); |
| |
| Insert(100, 102, 1); |
| EXPECT_EQ(102, Lookup(100)); |
| EXPECT_EQ(201, Lookup(200)); |
| EXPECT_EQ(-1, Lookup(300)); |
| |
| EXPECT_EQ(1, _deleted_keys.size()); |
| EXPECT_EQ(100, _deleted_keys[0]); |
| EXPECT_EQ(101, _deleted_values[0]); |
| } |
| |
| TEST_F(CacheTest, Erase) { |
| Erase(200); |
| EXPECT_EQ(0, _deleted_keys.size()); |
| |
| Insert(100, 101, 1); |
| Insert(200, 201, 1); |
| Erase(100); |
| EXPECT_EQ(-1, Lookup(100)); |
| EXPECT_EQ(201, Lookup(200)); |
| EXPECT_EQ(1, _deleted_keys.size()); |
| EXPECT_EQ(100, _deleted_keys[0]); |
| EXPECT_EQ(101, _deleted_values[0]); |
| |
| Erase(100); |
| EXPECT_EQ(-1, Lookup(100)); |
| EXPECT_EQ(201, Lookup(200)); |
| EXPECT_EQ(1, _deleted_keys.size()); |
| } |
| |
| TEST_F(CacheTest, EntriesArePinned) { |
| Insert(100, 101, 1); |
| std::string result1; |
| Cache::Handle* h1 = cache()->lookup(EncodeKey(&result1, 100)); |
| EXPECT_EQ(101, DecodeValue(cache()->value(h1))); |
| |
| Insert(100, 102, 1); |
| std::string result2; |
| Cache::Handle* h2 = cache()->lookup(EncodeKey(&result2, 100)); |
| EXPECT_EQ(102, DecodeValue(cache()->value(h2))); |
| EXPECT_EQ(0, _deleted_keys.size()); |
| |
| cache()->release(h1); |
| EXPECT_EQ(1, _deleted_keys.size()); |
| EXPECT_EQ(100, _deleted_keys[0]); |
| EXPECT_EQ(101, _deleted_values[0]); |
| |
| Erase(100); |
| EXPECT_EQ(-1, Lookup(100)); |
| EXPECT_EQ(1, _deleted_keys.size()); |
| |
| cache()->release(h2); |
| EXPECT_EQ(2, _deleted_keys.size()); |
| EXPECT_EQ(100, _deleted_keys[1]); |
| EXPECT_EQ(102, _deleted_values[1]); |
| } |
| |
| TEST_F(CacheTest, EvictionPolicy) { |
| Insert(100, 101, 1); |
| Insert(200, 201, 1); |
| |
| // Frequently used entry must be kept around |
| for (int i = 0; i < kCacheSize + 100; i++) { |
| Insert(1000 + i, 2000 + i, 1); |
| EXPECT_EQ(2000 + i, Lookup(1000 + i)); |
| EXPECT_EQ(101, Lookup(100)); |
| } |
| |
| EXPECT_EQ(101, Lookup(100)); |
| EXPECT_EQ(-1, Lookup(200)); |
| } |
| |
| TEST_F(CacheTest, EvictionPolicyWithDurable) { |
| Insert(100, 101, 1); |
| InsertDurable(200, 201, 1); |
| Insert(300, 101, 1); |
| |
| // Frequently used entry must be kept around |
| for (int i = 0; i < kCacheSize + 100; i++) { |
| Insert(1000 + i, 2000 + i, 1); |
| EXPECT_EQ(2000 + i, Lookup(1000 + i)); |
| EXPECT_EQ(101, Lookup(100)); |
| } |
| |
| EXPECT_EQ(-1, Lookup(300)); |
| EXPECT_EQ(101, Lookup(100)); |
| EXPECT_EQ(201, Lookup(200)); |
| } |
| |
| static void deleter(const CacheKey& key, void* v) {} |
| |
| static void insert_LRUCache(LRUCache& cache, const CacheKey& key, int value, |
| CachePriority priority) { |
| uint32_t hash = key.hash(key.data(), key.size(), 0); |
| static std::unique_ptr<MemTrackerLimiter> lru_cache_tracker = |
| std::make_unique<MemTrackerLimiter>(MemTrackerLimiter::Type::GLOBAL, |
| "TestSizeLruCache"); |
| cache.release(cache.insert(key, hash, EncodeValue(value), value, &deleter, |
| lru_cache_tracker.get(), priority, value)); |
| } |
| |
| static void insert_number_LRUCache(LRUCache& cache, const CacheKey& key, int value, int charge, |
| CachePriority priority) { |
| uint32_t hash = key.hash(key.data(), key.size(), 0); |
| static std::unique_ptr<MemTrackerLimiter> lru_cache_tracker = |
| std::make_unique<MemTrackerLimiter>(MemTrackerLimiter::Type::GLOBAL, |
| "TestNumberLruCache"); |
| cache.release(cache.insert(key, hash, EncodeValue(value), charge, &deleter, |
| lru_cache_tracker.get(), priority, value)); |
| } |
| |
| TEST_F(CacheTest, Usage) { |
| LRUCache cache(LRUCacheType::SIZE); |
| cache.set_capacity(1040); |
| |
| // The lru usage is handle_size + charge. |
| // handle_size = sizeof(handle) - 1 + key size = 104 - 1 + 3 = 106 |
| CacheKey key1("100"); |
| insert_LRUCache(cache, key1, 100, CachePriority::NORMAL); |
| ASSERT_EQ(206, cache.get_usage()); // 100 + 106 |
| |
| CacheKey key2("200"); |
| insert_LRUCache(cache, key2, 200, CachePriority::DURABLE); |
| ASSERT_EQ(512, cache.get_usage()); // 206 + 306(d), d = DURABLE |
| |
| CacheKey key3("300"); |
| insert_LRUCache(cache, key3, 300, CachePriority::NORMAL); |
| ASSERT_EQ(918, cache.get_usage()); // 206 + 306(d) + 406 |
| |
| CacheKey key4("400"); |
| insert_LRUCache(cache, key4, 400, CachePriority::NORMAL); |
| ASSERT_EQ(812, cache.get_usage()); // 306(d) + 506, evict 206 406 |
| |
| CacheKey key5("500"); |
| insert_LRUCache(cache, key5, 500, CachePriority::NORMAL); |
| ASSERT_EQ(912, cache.get_usage()); // 306(d) + 606, evict 506 |
| |
| CacheKey key6("600"); |
| insert_LRUCache(cache, key6, 600, CachePriority::NORMAL); |
| ASSERT_EQ(1012, cache.get_usage()); // 306(d) + 706, evict 506 |
| |
| CacheKey key7("950"); |
| insert_LRUCache(cache, key7, 950, CachePriority::DURABLE); |
| ASSERT_EQ(0, cache.get_usage()); // evict 306 706, because 950 + 106 > 1040, so insert failed |
| } |
| |
| TEST_F(CacheTest, Prune) { |
| LRUCache cache(LRUCacheType::NUMBER); |
| cache.set_capacity(5); |
| |
| // The lru usage is 1, add one entry |
| CacheKey key1("100"); |
| insert_number_LRUCache(cache, key1, 100, 1, CachePriority::NORMAL); |
| EXPECT_EQ(1, cache.get_usage()); |
| |
| CacheKey key2("200"); |
| insert_number_LRUCache(cache, key2, 200, 1, CachePriority::DURABLE); |
| EXPECT_EQ(2, cache.get_usage()); |
| |
| CacheKey key3("300"); |
| insert_number_LRUCache(cache, key3, 300, 1, CachePriority::NORMAL); |
| EXPECT_EQ(3, cache.get_usage()); |
| |
| CacheKey key4("400"); |
| insert_number_LRUCache(cache, key4, 400, 1, CachePriority::NORMAL); |
| EXPECT_EQ(4, cache.get_usage()); |
| |
| CacheKey key5("500"); |
| insert_number_LRUCache(cache, key5, 500, 1, CachePriority::NORMAL); |
| EXPECT_EQ(5, cache.get_usage()); |
| |
| CacheKey key6("600"); |
| insert_number_LRUCache(cache, key6, 600, 1, CachePriority::NORMAL); |
| EXPECT_EQ(5, cache.get_usage()); |
| |
| CacheKey key7("700"); |
| insert_number_LRUCache(cache, key7, 700, 1, CachePriority::DURABLE); |
| EXPECT_EQ(5, cache.get_usage()); |
| |
| auto pred = [](const void* value) -> bool { return false; }; |
| cache.prune_if(pred); |
| EXPECT_EQ(5, cache.get_usage()); |
| |
| auto pred2 = [](const void* value) -> bool { return DecodeValue((void*)value) > 400; }; |
| cache.prune_if(pred2); |
| EXPECT_EQ(2, cache.get_usage()); |
| |
| cache.prune(); |
| EXPECT_EQ(0, cache.get_usage()); |
| |
| for (int i = 1; i <= 5; ++i) { |
| insert_number_LRUCache(cache, CacheKey {std::to_string(i)}, i, 1, CachePriority::NORMAL); |
| EXPECT_EQ(i, cache.get_usage()); |
| } |
| cache.prune_if([](const void*) { return true; }); |
| EXPECT_EQ(0, cache.get_usage()); |
| } |
| |
| TEST_F(CacheTest, PruneIfLazyMode) { |
| LRUCache cache(LRUCacheType::NUMBER); |
| cache.set_capacity(10); |
| |
| // The lru usage is 1, add one entry |
| CacheKey key1("100"); |
| insert_number_LRUCache(cache, key1, 100, 1, CachePriority::NORMAL); |
| EXPECT_EQ(1, cache.get_usage()); |
| |
| CacheKey key2("200"); |
| insert_number_LRUCache(cache, key2, 200, 1, CachePriority::DURABLE); |
| EXPECT_EQ(2, cache.get_usage()); |
| |
| CacheKey key3("300"); |
| insert_number_LRUCache(cache, key3, 300, 1, CachePriority::NORMAL); |
| EXPECT_EQ(3, cache.get_usage()); |
| |
| CacheKey key4("666"); |
| insert_number_LRUCache(cache, key4, 666, 1, CachePriority::NORMAL); |
| EXPECT_EQ(4, cache.get_usage()); |
| |
| CacheKey key5("500"); |
| insert_number_LRUCache(cache, key5, 500, 1, CachePriority::NORMAL); |
| EXPECT_EQ(5, cache.get_usage()); |
| |
| CacheKey key6("600"); |
| insert_number_LRUCache(cache, key6, 600, 1, CachePriority::NORMAL); |
| EXPECT_EQ(6, cache.get_usage()); |
| |
| CacheKey key7("700"); |
| insert_number_LRUCache(cache, key7, 700, 1, CachePriority::DURABLE); |
| EXPECT_EQ(7, cache.get_usage()); |
| |
| auto pred = [](const void* value) -> bool { return false; }; |
| cache.prune_if(pred, true); |
| EXPECT_EQ(7, cache.get_usage()); |
| |
| // in lazy mode, the first item not satisfied the pred2, `prune_if` then stopped |
| // and no item's removed. |
| auto pred2 = [](const void* value) -> bool { return DecodeValue((void*)value) > 400; }; |
| cache.prune_if(pred2, true); |
| EXPECT_EQ(7, cache.get_usage()); |
| |
| // in normal priority, 100, 300 are removed |
| // in durable priority, 200 is removed |
| auto pred3 = [](const void* value) -> bool { return DecodeValue((void*)value) <= 600; }; |
| EXPECT_EQ(3, cache.prune_if(pred3, true)); |
| EXPECT_EQ(4, cache.get_usage()); |
| } |
| |
| TEST_F(CacheTest, Number) { |
| LRUCache cache(LRUCacheType::NUMBER); |
| cache.set_capacity(5); |
| |
| // The lru usage is 1, add one entry |
| CacheKey key1("1"); |
| insert_number_LRUCache(cache, key1, 0, 1, CachePriority::NORMAL); |
| EXPECT_EQ(1, cache.get_usage()); |
| |
| CacheKey key2("2"); |
| insert_number_LRUCache(cache, key2, 0, 2, CachePriority::DURABLE); |
| EXPECT_EQ(3, cache.get_usage()); |
| |
| CacheKey key3("3"); |
| insert_number_LRUCache(cache, key3, 0, 3, CachePriority::NORMAL); |
| EXPECT_EQ(5, cache.get_usage()); |
| |
| CacheKey key4("4"); |
| insert_number_LRUCache(cache, key4, 0, 3, CachePriority::NORMAL); |
| EXPECT_EQ(5, cache.get_usage()); // evict key3, because key3 is NORMAL, key2 is DURABLE. |
| |
| CacheKey key5("5"); |
| insert_number_LRUCache(cache, key5, 0, 5, CachePriority::NORMAL); |
| EXPECT_EQ(5, cache.get_usage()); // evict key2, key4 |
| |
| CacheKey key6("6"); |
| insert_number_LRUCache(cache, key6, 0, 6, CachePriority::NORMAL); |
| EXPECT_EQ(0, cache.get_usage()); // evict key5, evict key6 at the same time as release. |
| |
| CacheKey key7("7"); |
| insert_number_LRUCache(cache, key7, 200, 1, CachePriority::NORMAL); |
| EXPECT_EQ(1, cache.get_usage()); |
| |
| CacheKey key8("8"); |
| insert_number_LRUCache(cache, key8, 100, 1, CachePriority::DURABLE); |
| EXPECT_EQ(2, cache.get_usage()); |
| |
| insert_number_LRUCache(cache, key8, 100, 1, CachePriority::DURABLE); |
| EXPECT_EQ(2, cache.get_usage()); |
| |
| auto pred = [](const void* value) -> bool { return false; }; |
| cache.prune_if(pred); |
| EXPECT_EQ(2, cache.get_usage()); |
| |
| auto pred2 = [](const void* value) -> bool { return DecodeValue((void*)value) > 100; }; |
| cache.prune_if(pred2); |
| EXPECT_EQ(1, cache.get_usage()); |
| |
| cache.prune(); |
| EXPECT_EQ(0, cache.get_usage()); |
| } |
| |
| TEST_F(CacheTest, HeavyEntries) { |
| // Add a bunch of light and heavy entries and then count the combined |
| // size of items still in the cache, which must be approximately the |
| // same as the total capacity. |
| const int kLight = 1; |
| const int kHeavy = 10; |
| int added = 0; |
| int index = 0; |
| |
| while (added < 2 * kCacheSize) { |
| const int weight = (index & 1) ? kLight : kHeavy; |
| Insert(index, 1000 + index, weight); |
| added += weight; |
| index++; |
| } |
| |
| int cached_weight = 0; |
| |
| for (int i = 0; i < index; i++) { |
| const int weight = (i & 1 ? kLight : kHeavy); |
| int r = Lookup(i); |
| |
| if (r >= 0) { |
| cached_weight += weight; |
| EXPECT_EQ(1000 + i, r); |
| } |
| } |
| |
| EXPECT_LE(cached_weight, kCacheSize + kCacheSize / 10); |
| } |
| |
| TEST_F(CacheTest, NewId) { |
| uint64_t a = cache()->new_id(); |
| uint64_t b = cache()->new_id(); |
| EXPECT_NE(a, b); |
| } |
| |
| TEST_F(CacheTest, SimpleBenchmark) { |
| for (int i = 0; i < kCacheSize * LOOP_LESS_OR_MORE(10, 10000); i++) { |
| Insert(1000 + i, 2000 + i, 1); |
| EXPECT_EQ(2000 + i, Lookup(1000 + i)); |
| } |
| } |
| |
| TEST(CacheHandleTest, HandleTableTest) { |
| HandleTable ht; |
| |
| for (uint32_t i = 0; i < ht._length; ++i) { |
| EXPECT_EQ(ht._list[i], nullptr); |
| } |
| |
| const int count = 10; |
| CacheKey keys[count] = {"0", "1", "2", "3", "4", "5", "6", "7", "8", "9"}; |
| EXPECT_NE(keys[0], keys[1]); |
| LRUHandle* hs[count]; |
| for (int i = 0; i < count; ++i) { |
| CacheKey* key = &keys[i]; |
| LRUHandle* h = reinterpret_cast<LRUHandle*>(malloc(sizeof(LRUHandle) - 1 + key->size())); |
| h->value = nullptr; |
| h->deleter = nullptr; |
| h->charge = 1; |
| h->total_size = sizeof(LRUHandle) - 1 + key->size() + 1; |
| h->key_length = key->size(); |
| h->hash = 1; // make them in a same hash table linked-list |
| h->refs = 0; |
| h->next = h->prev = nullptr; |
| h->next_hash = nullptr; |
| h->in_cache = false; |
| h->priority = CachePriority::NORMAL; |
| memcpy(h->key_data, key->data(), key->size()); |
| |
| LRUHandle* old = ht.insert(h); |
| EXPECT_EQ(ht._elems, i + 1); |
| EXPECT_EQ(old, nullptr); // there is no entry with the same key and hash |
| hs[i] = h; |
| } |
| EXPECT_EQ(ht._elems, count); |
| LRUHandle* h = ht.lookup(keys[0], 1); |
| LRUHandle** head_ptr = &(ht._list[1 & (ht._length - 1)]); |
| LRUHandle* head = *head_ptr; |
| ASSERT_EQ(head, h); |
| int index = 0; |
| while (h != nullptr) { |
| EXPECT_EQ(hs[index], h) << index; |
| h = h->next_hash; |
| ++index; |
| } |
| |
| for (int i = 0; i < count; ++i) { |
| CacheKey* key = &keys[i]; |
| LRUHandle* h = reinterpret_cast<LRUHandle*>(malloc(sizeof(LRUHandle) - 1 + key->size())); |
| h->value = nullptr; |
| h->deleter = nullptr; |
| h->charge = 1; |
| h->total_size = sizeof(LRUHandle) - 1 + key->size() + 1; |
| h->key_length = key->size(); |
| h->hash = 1; // make them in a same hash table linked-list |
| h->refs = 0; |
| h->next = h->prev = nullptr; |
| h->next_hash = nullptr; |
| h->in_cache = false; |
| h->priority = CachePriority::NORMAL; |
| memcpy(h->key_data, key->data(), key->size()); |
| |
| EXPECT_EQ(ht.insert(h), hs[i]); // there is an entry with the same key and hash |
| EXPECT_EQ(ht._elems, count); |
| free(hs[i]); |
| hs[i] = h; |
| } |
| EXPECT_EQ(ht._elems, count); |
| |
| for (int i = 0; i < count; ++i) { |
| EXPECT_EQ(ht.lookup(keys[i], 1), hs[i]); |
| } |
| |
| LRUHandle* old = ht.remove(CacheKey("0"), 1); // first in hash table linked-list |
| ASSERT_EQ(old, hs[0]); |
| ASSERT_EQ(old->next_hash, hs[1]); // hs[1] is the new first node |
| ASSERT_EQ(*head_ptr, hs[1]); |
| |
| old = ht.remove(CacheKey("9"), 1); // last in hash table linked-list |
| ASSERT_EQ(old, hs[9]); |
| |
| old = ht.remove(CacheKey("5"), 1); // middle in hash table linked-list |
| ASSERT_EQ(old, hs[5]); |
| ASSERT_EQ(old->next_hash, hs[6]); |
| ASSERT_EQ(hs[4]->next_hash, hs[6]); |
| |
| ht.remove(hs[4]); // middle in hash table linked-list |
| ASSERT_EQ(hs[3]->next_hash, hs[6]); |
| |
| EXPECT_EQ(ht._elems, count - 4); |
| |
| for (int i = 0; i < count; ++i) { |
| free(hs[i]); |
| } |
| } |
| |
| } // namespace doris |