diff options
Diffstat (limited to 'unittests/DebugInfo/PDB/HashTableTest.cpp')
-rw-r--r-- | unittests/DebugInfo/PDB/HashTableTest.cpp | 168 |
1 files changed, 168 insertions, 0 deletions
diff --git a/unittests/DebugInfo/PDB/HashTableTest.cpp b/unittests/DebugInfo/PDB/HashTableTest.cpp new file mode 100644 index 000000000000..94c9ee86c4a6 --- /dev/null +++ b/unittests/DebugInfo/PDB/HashTableTest.cpp @@ -0,0 +1,168 @@ +//===- llvm/unittest/DebugInfo/PDB/HashTableTest.cpp ----------------------===// +// +// The LLVM Compiler Infrastructure +// +// This file is distributed under the University of Illinois Open Source +// License. See LICENSE.TXT for details. +// +//===----------------------------------------------------------------------===// + +#include "ErrorChecking.h" +#include "gtest/gtest.h" + +#include "llvm/DebugInfo/PDB/Native/HashTable.h" +#include "llvm/Support/BinaryByteStream.h" +#include "llvm/Support/BinaryStreamReader.h" +#include "llvm/Support/BinaryStreamWriter.h" + +#include <vector> + +using namespace llvm; +using namespace llvm::pdb; +using namespace llvm::support; + +namespace { +class HashTableInternals : public HashTable { +public: + using HashTable::Buckets; + using HashTable::Present; + using HashTable::Deleted; +}; +} + +TEST(HashTableTest, TestSimple) { + HashTable Table; + EXPECT_EQ(0u, Table.size()); + EXPECT_GT(Table.capacity(), 0u); + + Table.set(3, 7); + EXPECT_EQ(1u, Table.size()); + ASSERT_NE(Table.end(), Table.find(3)); + EXPECT_EQ(7u, Table.get(3)); +} + +TEST(HashTableTest, TestCollision) { + HashTable Table; + EXPECT_EQ(0u, Table.size()); + EXPECT_GT(Table.capacity(), 0u); + + // We use knowledge of the hash table's implementation details to make sure + // to add another value that is the equivalent to the first value modulo the + // hash table's capacity. + uint32_t N1 = Table.capacity() + 1; + uint32_t N2 = 2 * N1; + + Table.set(N1, 7); + Table.set(N2, 12); + EXPECT_EQ(2u, Table.size()); + ASSERT_NE(Table.end(), Table.find(N1)); + ASSERT_NE(Table.end(), Table.find(N2)); + + EXPECT_EQ(7u, Table.get(N1)); + EXPECT_EQ(12u, Table.get(N2)); +} + +TEST(HashTableTest, TestRemove) { + HashTable Table; + EXPECT_EQ(0u, Table.size()); + EXPECT_GT(Table.capacity(), 0u); + + Table.set(1, 2); + Table.set(3, 4); + EXPECT_EQ(2u, Table.size()); + ASSERT_NE(Table.end(), Table.find(1)); + ASSERT_NE(Table.end(), Table.find(3)); + + EXPECT_EQ(2u, Table.get(1)); + EXPECT_EQ(4u, Table.get(3)); + + Table.remove(1u); + EXPECT_EQ(1u, Table.size()); + EXPECT_EQ(Table.end(), Table.find(1)); + ASSERT_NE(Table.end(), Table.find(3)); + EXPECT_EQ(4u, Table.get(3)); +} + +TEST(HashTableTest, TestCollisionAfterMultipleProbes) { + HashTable Table; + EXPECT_EQ(0u, Table.size()); + EXPECT_GT(Table.capacity(), 0u); + + // Probing looks for the first available slot. A slot may already be filled + // as a result of an item with a *different* hash value already being there. + // Test that when this happens, the probe still finds the value. + uint32_t N1 = Table.capacity() + 1; + uint32_t N2 = N1 + 1; + uint32_t N3 = 2 * N1; + + Table.set(N1, 7); + Table.set(N2, 11); + Table.set(N3, 13); + EXPECT_EQ(3u, Table.size()); + ASSERT_NE(Table.end(), Table.find(N1)); + ASSERT_NE(Table.end(), Table.find(N2)); + ASSERT_NE(Table.end(), Table.find(N3)); + + EXPECT_EQ(7u, Table.get(N1)); + EXPECT_EQ(11u, Table.get(N2)); + EXPECT_EQ(13u, Table.get(N3)); + + // Remove the one that had been filled in the middle, then insert another one + // with a collision. It should fill the newly emptied slot. + Table.remove(N2); + uint32_t N4 = N1 * 3; + Table.set(N4, 17); + EXPECT_EQ(3u, Table.size()); + ASSERT_NE(Table.end(), Table.find(N1)); + ASSERT_NE(Table.end(), Table.find(N3)); + ASSERT_NE(Table.end(), Table.find(N4)); + + EXPECT_EQ(7u, Table.get(N1)); + EXPECT_EQ(13u, Table.get(N3)); + EXPECT_EQ(17u, Table.get(N4)); +} + +TEST(HashTableTest, Grow) { + // So that we are independent of the load factor, `capacity` items, which is + // guaranteed to trigger a grow. Then verify that the size is the same, the + // capacity is larger, and all the original items are still in the table. + + HashTable Table; + uint32_t OldCapacity = Table.capacity(); + for (uint32_t I = 0; I < OldCapacity; ++I) { + Table.set(OldCapacity + I * 2 + 1, I * 2 + 3); + } + EXPECT_EQ(OldCapacity, Table.size()); + EXPECT_GT(Table.capacity(), OldCapacity); + for (uint32_t I = 0; I < OldCapacity; ++I) { + ASSERT_NE(Table.end(), Table.find(OldCapacity + I * 2 + 1)); + EXPECT_EQ(I * 2 + 3, Table.get(OldCapacity + I * 2 + 1)); + } +} + +TEST(HashTableTest, Serialization) { + HashTableInternals Table; + uint32_t Cap = Table.capacity(); + for (uint32_t I = 0; I < Cap; ++I) { + Table.set(Cap + I * 2 + 1, I * 2 + 3); + } + + std::vector<uint8_t> Buffer(Table.calculateSerializedLength()); + MutableBinaryByteStream Stream(Buffer, little); + BinaryStreamWriter Writer(Stream); + EXPECT_NO_ERROR(Table.commit(Writer)); + // We should have written precisely the number of bytes we calculated earlier. + EXPECT_EQ(Buffer.size(), Writer.getOffset()); + + HashTableInternals Table2; + BinaryStreamReader Reader(Stream); + EXPECT_NO_ERROR(Table2.load(Reader)); + // We should have read precisely the number of bytes we calculated earlier. + EXPECT_EQ(Buffer.size(), Reader.getOffset()); + + EXPECT_EQ(Table.size(), Table2.size()); + EXPECT_EQ(Table.capacity(), Table2.capacity()); + EXPECT_EQ(Table.Buckets, Table2.Buckets); + EXPECT_EQ(Table.Present, Table2.Present); + EXPECT_EQ(Table.Deleted, Table2.Deleted); +} |