blob: 5e7d1ee701c6090e2d2aeb7df4caba03b5f6ea49 [file] [log] [blame]
//===-- MappedHash.h --------------------------------------------*- C++ -*-===//
//
// The LLVM Compiler Infrastructure
//
// This file is distributed under the University of Illinois Open Source
// License. See LICENSE.TXT for details.
//
//===----------------------------------------------------------------------===//
#ifndef liblldb_MappedHash_h_
#define liblldb_MappedHash_h_
// C Includes
#include <assert.h>
#include <stdint.h>
// C++ Includes
#include <algorithm>
#include <functional>
#include <map>
#include <vector>
// Other libraries and framework includes
// Project includes
#include "lldb/Core/DataExtractor.h"
#include "lldb/Core/Stream.h"
class MappedHash {
public:
enum HashFunctionType {
eHashFunctionDJB = 0u // Daniel J Bernstein hash function that is also used
// by the ELF GNU_HASH sections
};
static uint32_t HashStringUsingDJB(const char *s) {
uint32_t h = 5381;
for (unsigned char c = *s; c; c = *++s)
h = ((h << 5) + h) + c;
return h;
}
static uint32_t HashString(uint32_t hash_function, const char *s) {
if (!s)
return 0;
switch (hash_function) {
case MappedHash::eHashFunctionDJB:
return HashStringUsingDJB(s);
default:
break;
}
llvm_unreachable("Invalid hash function index");
}
static const uint32_t HASH_MAGIC = 0x48415348u;
static const uint32_t HASH_CIGAM = 0x48534148u;
template <typename T> struct Header {
typedef T HeaderData;
uint32_t
magic; // HASH_MAGIC or HASH_CIGAM magic value to allow endian detection
uint16_t version; // Version number
uint16_t hash_function; // The hash function enumeration that was used
uint32_t bucket_count; // The number of buckets in this hash table
uint32_t hashes_count; // The total number of unique hash values and hash
// data offsets in this table
uint32_t header_data_len; // The size in bytes of the "header_data" template
// member below
HeaderData header_data; //
Header()
: magic(HASH_MAGIC), version(1), hash_function(eHashFunctionDJB),
bucket_count(0), hashes_count(0), header_data_len(sizeof(T)),
header_data() {}
virtual ~Header() = default;
size_t GetByteSize() const {
return sizeof(magic) + sizeof(version) + sizeof(hash_function) +
sizeof(bucket_count) + sizeof(hashes_count) +
sizeof(header_data_len) + header_data_len;
}
virtual size_t GetByteSize(const HeaderData &header_data) = 0;
void SetHeaderDataByteSize(uint32_t header_data_byte_size) {
header_data_len = header_data_byte_size;
}
void Dump(lldb_private::Stream &s) {
s.Printf("header.magic = 0x%8.8x\n", magic);
s.Printf("header.version = 0x%4.4x\n", version);
s.Printf("header.hash_function = 0x%4.4x\n", hash_function);
s.Printf("header.bucket_count = 0x%8.8x %u\n", bucket_count,
bucket_count);
s.Printf("header.hashes_count = 0x%8.8x %u\n", hashes_count,
hashes_count);
s.Printf("header.header_data_len = 0x%8.8x %u\n", header_data_len,
header_data_len);
}
virtual lldb::offset_t Read(lldb_private::DataExtractor &data,
lldb::offset_t offset) {
if (data.ValidOffsetForDataOfSize(
offset, sizeof(magic) + sizeof(version) + sizeof(hash_function) +
sizeof(bucket_count) + sizeof(hashes_count) +
sizeof(header_data_len))) {
magic = data.GetU32(&offset);
if (magic != HASH_MAGIC) {
if (magic == HASH_CIGAM) {
switch (data.GetByteOrder()) {
case lldb::eByteOrderBig:
data.SetByteOrder(lldb::eByteOrderLittle);
break;
case lldb::eByteOrderLittle:
data.SetByteOrder(lldb::eByteOrderBig);
break;
default:
return LLDB_INVALID_OFFSET;
}
} else {
// Magic bytes didn't match
version = 0;
return LLDB_INVALID_OFFSET;
}
}
version = data.GetU16(&offset);
if (version != 1) {
// Unsupported version
return LLDB_INVALID_OFFSET;
}
hash_function = data.GetU16(&offset);
if (hash_function == 4)
hash_function = 0; // Deal with pre-release version of this table...
bucket_count = data.GetU32(&offset);
hashes_count = data.GetU32(&offset);
header_data_len = data.GetU32(&offset);
return offset;
}
return LLDB_INVALID_OFFSET;
}
//
// // Returns a buffer that contains a serialized version of this
// table
// // that must be freed with free().
// virtual void *
// Write (int fd);
};
template <typename __KeyType, class __HeaderDataType, class __ValueType>
class ExportTable {
public:
typedef __HeaderDataType HeaderDataType;
typedef Header<HeaderDataType> HeaderType;
typedef __KeyType KeyType;
typedef __ValueType ValueType;
struct Entry {
uint32_t hash;
KeyType key;
ValueType value;
};
typedef std::vector<ValueType> ValueArrayType;
typedef std::map<KeyType, ValueArrayType> HashData;
// Map a name hash to one or more name infos
typedef std::map<uint32_t, HashData> HashToHashData;
virtual KeyType GetKeyForStringType(const char *cstr) const = 0;
virtual size_t GetByteSize(const HashData &key_to_key_values) = 0;
virtual bool WriteHashData(const HashData &hash_data,
lldb_private::Stream &ostrm) = 0;
//
void AddEntry(const char *cstr, const ValueType &value) {
Entry entry;
entry.hash = MappedHash::HashString(eHashFunctionDJB, cstr);
entry.key = GetKeyForStringType(cstr);
entry.value = value;
m_entries.push_back(entry);
}
void Save(const HeaderDataType &header_data, lldb_private::Stream &ostrm) {
if (m_entries.empty())
return;
const uint32_t num_entries = m_entries.size();
uint32_t i = 0;
HeaderType header;
header.magic = HASH_MAGIC;
header.version = 1;
header.hash_function = eHashFunctionDJB;
header.bucket_count = 0;
header.hashes_count = 0;
header.prologue_length = header_data.GetByteSize();
// We need to figure out the number of unique hashes first before we can
// calculate the number of buckets we want to use.
typedef std::vector<uint32_t> hash_coll;
hash_coll unique_hashes;
unique_hashes.resize(num_entries);
for (i = 0; i < num_entries; ++i)
unique_hashes[i] = m_entries[i].hash;
std::sort(unique_hashes.begin(), unique_hashes.end());
hash_coll::iterator pos =
std::unique(unique_hashes.begin(), unique_hashes.end());
const size_t num_unique_hashes =
std::distance(unique_hashes.begin(), pos);
if (num_unique_hashes > 1024)
header.bucket_count = num_unique_hashes / 4;
else if (num_unique_hashes > 16)
header.bucket_count = num_unique_hashes / 2;
else
header.bucket_count = num_unique_hashes;
if (header.bucket_count == 0)
header.bucket_count = 1;
std::vector<HashToHashData> hash_buckets;
std::vector<uint32_t> hash_indexes(header.bucket_count, 0);
std::vector<uint32_t> hash_values;
std::vector<uint32_t> hash_offsets;
hash_buckets.resize(header.bucket_count);
uint32_t bucket_entry_empties = 0;
// StreamString hash_file_data(Stream::eBinary,
// dwarf->GetObjectFile()->GetAddressByteSize(),
// dwarf->GetObjectFile()->GetByteSize());
// Push all of the hashes into their buckets and create all bucket
// entries all populated with data.
for (i = 0; i < num_entries; ++i) {
const uint32_t hash = m_entries[i].hash;
const uint32_t bucket_idx = hash % header.bucket_count;
const uint32_t strp_offset = m_entries[i].str_offset;
const uint32_t die_offset = m_entries[i].die_offset;
hash_buckets[bucket_idx][hash][strp_offset].push_back(die_offset);
}
// Now for each bucket we write the bucket value which is the
// number of hashes and the hash index encoded into a single
// 32 bit unsigned integer.
for (i = 0; i < header.bucket_count; ++i) {
HashToHashData &bucket_entry = hash_buckets[i];
if (bucket_entry.empty()) {
// Empty bucket
++bucket_entry_empties;
hash_indexes[i] = UINT32_MAX;
} else {
const uint32_t hash_value_index = hash_values.size();
uint32_t hash_count = 0;
typename HashToHashData::const_iterator pos, end = bucket_entry.end();
for (pos = bucket_entry.begin(); pos != end; ++pos) {
hash_values.push_back(pos->first);
hash_offsets.push_back(GetByteSize(pos->second));
++hash_count;
}
hash_indexes[i] = hash_value_index;
}
}
header.hashes_count = hash_values.size();
// Write the header out now that we have the hash_count
header.Write(ostrm);
// Now for each bucket we write the start index of the hashes
// for the current bucket, or UINT32_MAX if the bucket is empty
for (i = 0; i < header.bucket_count; ++i) {
ostrm.PutHex32(hash_indexes[i]);
}
// Now we need to write out all of the hash values
for (i = 0; i < header.hashes_count; ++i) {
ostrm.PutHex32(hash_values[i]);
}
// Now we need to write out all of the hash data offsets,
// there is an offset for each hash in the hashes array
// that was written out above
for (i = 0; i < header.hashes_count; ++i) {
ostrm.PutHex32(hash_offsets[i]);
}
// Now we write the data for each hash and verify we got the offset
// correct above...
for (i = 0; i < header.bucket_count; ++i) {
HashToHashData &bucket_entry = hash_buckets[i];
typename HashToHashData::const_iterator pos, end = bucket_entry.end();
for (pos = bucket_entry.begin(); pos != end; ++pos) {
if (!bucket_entry.empty()) {
WriteHashData(pos->second);
}
}
}
}
protected:
typedef std::vector<Entry> collection;
collection m_entries;
};
// A class for reading and using a saved hash table from a block of data
// in memory
template <typename __KeyType, class __HeaderType, class __HashData>
class MemoryTable {
public:
typedef __HeaderType HeaderType;
typedef __KeyType KeyType;
typedef __HashData HashData;
enum Result {
eResultKeyMatch = 0u, // The entry was found, key matched and "pair" was
// filled in successfully
eResultKeyMismatch =
1u, // Bucket hash data collision, but key didn't match
eResultEndOfHashData = 2u, // The chain of items for this hash data in
// this bucket is terminated, search no more
eResultError = 3u // Error parsing the hash data, abort
};
struct Pair {
KeyType key;
HashData value;
};
MemoryTable(lldb_private::DataExtractor &data)
: m_header(), m_hash_indexes(nullptr), m_hash_values(nullptr),
m_hash_offsets(nullptr) {
lldb::offset_t offset = m_header.Read(data, 0);
if (offset != LLDB_INVALID_OFFSET && IsValid()) {
m_hash_indexes = (const uint32_t *)data.GetData(
&offset, m_header.bucket_count * sizeof(uint32_t));
m_hash_values = (const uint32_t *)data.GetData(
&offset, m_header.hashes_count * sizeof(uint32_t));
m_hash_offsets = (const uint32_t *)data.GetData(
&offset, m_header.hashes_count * sizeof(uint32_t));
}
}
virtual ~MemoryTable() = default;
bool IsValid() const {
return m_header.version == 1 &&
m_header.hash_function == eHashFunctionDJB &&
m_header.bucket_count > 0;
}
uint32_t GetHashIndex(uint32_t bucket_idx) const {
if (m_hash_indexes && bucket_idx < m_header.bucket_count)
return m_hash_indexes[bucket_idx];
return UINT32_MAX;
}
uint32_t GetHashValue(uint32_t hash_idx) const {
if (m_hash_values && hash_idx < m_header.hashes_count)
return m_hash_values[hash_idx];
return UINT32_MAX;
}
uint32_t GetHashDataOffset(uint32_t hash_idx) const {
if (m_hash_offsets && hash_idx < m_header.hashes_count)
return m_hash_offsets[hash_idx];
return UINT32_MAX;
}
bool Find(const char *name, Pair &pair) const {
if (!name || !name[0])
return false;
if (IsValid()) {
const uint32_t bucket_count = m_header.bucket_count;
const uint32_t hash_count = m_header.hashes_count;
const uint32_t hash_value =
MappedHash::HashString(m_header.hash_function, name);
const uint32_t bucket_idx = hash_value % bucket_count;
uint32_t hash_idx = GetHashIndex(bucket_idx);
if (hash_idx < hash_count) {
for (; hash_idx < hash_count; ++hash_idx) {
const uint32_t curr_hash_value = GetHashValue(hash_idx);
if (curr_hash_value == hash_value) {
lldb::offset_t hash_data_offset = GetHashDataOffset(hash_idx);
while (hash_data_offset != UINT32_MAX) {
const lldb::offset_t prev_hash_data_offset = hash_data_offset;
Result hash_result =
GetHashDataForName(name, &hash_data_offset, pair);
// Check the result of getting our hash data
switch (hash_result) {
case eResultKeyMatch:
return true;
case eResultKeyMismatch:
if (prev_hash_data_offset == hash_data_offset)
return false;
break;
case eResultEndOfHashData:
// The last HashData for this key has been reached, stop
// searching
return false;
case eResultError:
// Error parsing the hash data, abort
return false;
}
}
}
if ((curr_hash_value % bucket_count) != bucket_idx)
break;
}
}
}
return false;
}
// This method must be implemented in any subclasses.
// The KeyType is user specified and must somehow result in a string
// value. For example, the KeyType might be a string offset in a string
// table and subclasses can store their string table as a member of the
// subclass and return a valie "const char *" given a "key". The value
// could also be a C string pointer, in which case just returning "key"
// will suffice.
virtual const char *GetStringForKeyType(KeyType key) const = 0;
virtual bool ReadHashData(uint32_t hash_data_offset,
HashData &hash_data) const = 0;
// This method must be implemented in any subclasses and it must try to
// read one "Pair" at the offset pointed to by the "hash_data_offset_ptr"
// parameter. This offset should be updated as bytes are consumed and
// a value "Result" enum should be returned. If the "name" matches the
// full name for the "pair.key" (which must be filled in by this call),
// then the HashData in the pair ("pair.value") should be extracted and
// filled in and "eResultKeyMatch" should be returned. If "name" doesn't
// match this string for the key, then "eResultKeyMismatch" should be
// returned and all data for the current HashData must be consumed or
// skipped and the "hash_data_offset_ptr" offset needs to be updated to
// point to the next HashData. If the end of the HashData objects for
// a given hash value have been reached, then "eResultEndOfHashData"
// should be returned. If anything else goes wrong during parsing,
// return "eResultError" and the corresponding "Find()" function will
// be canceled and return false.
virtual Result GetHashDataForName(const char *name,
lldb::offset_t *hash_data_offset_ptr,
Pair &pair) const = 0;
const HeaderType &GetHeader() { return m_header; }
void ForEach(
std::function<bool(const HashData &hash_data)> const &callback) const {
const size_t num_hash_offsets = m_header.hashes_count;
for (size_t i = 0; i < num_hash_offsets; ++i) {
uint32_t hash_data_offset = GetHashDataOffset(i);
if (hash_data_offset != UINT32_MAX) {
HashData hash_data;
if (ReadHashData(hash_data_offset, hash_data)) {
// If the callback returns false, then we are done and should stop
if (callback(hash_data) == false)
return;
}
}
}
}
protected:
// Implementation agnostic information
HeaderType m_header;
const uint32_t *m_hash_indexes;
const uint32_t *m_hash_values;
const uint32_t *m_hash_offsets;
};
};
#endif // liblldb_MappedHash_h_