| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #ifndef ANDROID_BASIC_HASHTABLE_H |
| #define ANDROID_BASIC_HASHTABLE_H |
| |
| #include <stdint.h> |
| #include <sys/types.h> |
| #include <utils/SharedBuffer.h> |
| #include <utils/TypeHelpers.h> |
| |
| namespace android { |
| |
| /* Implementation type. Nothing to see here. */ |
| class BasicHashtableImpl { |
| protected: |
| struct Bucket { |
| // The collision flag indicates that the bucket is part of a collision chain |
| // such that at least two entries both hash to this bucket. When true, we |
| // may need to seek further along the chain to find the entry. |
| static const uint32_t COLLISION = 0x80000000UL; |
| |
| // The present flag indicates that the bucket contains an initialized entry value. |
| static const uint32_t PRESENT = 0x40000000UL; |
| |
| // Mask for 30 bits worth of the hash code that are stored within the bucket to |
| // speed up lookups and rehashing by eliminating the need to recalculate the |
| // hash code of the entry's key. |
| static const uint32_t HASH_MASK = 0x3fffffffUL; |
| |
| // Combined value that stores the collision and present flags as well as |
| // a 30 bit hash code. |
| uint32_t cookie; |
| |
| // Storage for the entry begins here. |
| char entry[0]; |
| }; |
| |
| BasicHashtableImpl(size_t entrySize, bool hasTrivialDestructor, |
| size_t minimumInitialCapacity, float loadFactor); |
| BasicHashtableImpl(const BasicHashtableImpl& other); |
| virtual ~BasicHashtableImpl(); |
| |
| void dispose(); |
| |
| inline void edit() { |
| if (mBuckets && !SharedBuffer::bufferFromData(mBuckets)->onlyOwner()) { |
| clone(); |
| } |
| } |
| |
| void setTo(const BasicHashtableImpl& other); |
| void clear(); |
| |
| ssize_t next(ssize_t index) const; |
| ssize_t find(ssize_t index, hash_t hash, const void* __restrict__ key) const; |
| size_t add(hash_t hash, const void* __restrict__ entry); |
| void removeAt(size_t index); |
| void rehash(size_t minimumCapacity, float loadFactor); |
| |
| const size_t mBucketSize; // number of bytes per bucket including the entry |
| const bool mHasTrivialDestructor; // true if the entry type does not require destruction |
| size_t mCapacity; // number of buckets that can be filled before exceeding load factor |
| float mLoadFactor; // load factor |
| size_t mSize; // number of elements actually in the table |
| size_t mFilledBuckets; // number of buckets for which collision or present is true |
| size_t mBucketCount; // number of slots in the mBuckets array |
| void* mBuckets; // array of buckets, as a SharedBuffer |
| |
| inline const Bucket& bucketAt(const void* __restrict__ buckets, size_t index) const { |
| return *reinterpret_cast<const Bucket*>( |
| static_cast<const uint8_t*>(buckets) + index * mBucketSize); |
| } |
| |
| inline Bucket& bucketAt(void* __restrict__ buckets, size_t index) const { |
| return *reinterpret_cast<Bucket*>(static_cast<uint8_t*>(buckets) + index * mBucketSize); |
| } |
| |
| virtual bool compareBucketKey(const Bucket& bucket, const void* __restrict__ key) const = 0; |
| virtual void initializeBucketEntry(Bucket& bucket, const void* __restrict__ entry) const = 0; |
| virtual void destroyBucketEntry(Bucket& bucket) const = 0; |
| |
| private: |
| void clone(); |
| |
| // Allocates a bucket array as a SharedBuffer. |
| void* allocateBuckets(size_t count) const; |
| |
| // Releases a bucket array's associated SharedBuffer. |
| void releaseBuckets(void* __restrict__ buckets, size_t count) const; |
| |
| // Destroys the contents of buckets (invokes destroyBucketEntry for each |
| // populated bucket if needed). |
| void destroyBuckets(void* __restrict__ buckets, size_t count) const; |
| |
| // Copies the content of buckets (copies the cookie and invokes copyBucketEntry |
| // for each populated bucket if needed). |
| void copyBuckets(const void* __restrict__ fromBuckets, |
| void* __restrict__ toBuckets, size_t count) const; |
| |
| // Determines the appropriate size of a bucket array to store a certain minimum |
| // number of entries and returns its effective capacity. |
| static void determineCapacity(size_t minimumCapacity, float loadFactor, |
| size_t* __restrict__ outBucketCount, size_t* __restrict__ outCapacity); |
| |
| // Trim a hash code to 30 bits to match what we store in the bucket's cookie. |
| inline static hash_t trimHash(hash_t hash) { |
| return (hash & Bucket::HASH_MASK) ^ (hash >> 30); |
| } |
| |
| // Returns the index of the first bucket that is in the collision chain |
| // for the specified hash code, given the total number of buckets. |
| // (Primary hash) |
| inline static size_t chainStart(hash_t hash, size_t count) { |
| return hash % count; |
| } |
| |
| // Returns the increment to add to a bucket index to seek to the next bucket |
| // in the collision chain for the specified hash code, given the total number of buckets. |
| // (Secondary hash) |
| inline static size_t chainIncrement(hash_t hash, size_t count) { |
| return ((hash >> 7) | (hash << 25)) % (count - 1) + 1; |
| } |
| |
| // Returns the index of the next bucket that is in the collision chain |
| // that is defined by the specified increment, given the total number of buckets. |
| inline static size_t chainSeek(size_t index, size_t increment, size_t count) { |
| return (index + increment) % count; |
| } |
| }; |
| |
| /* |
| * A BasicHashtable stores entries that are indexed by hash code in place |
| * within an array. The basic operations are finding entries by key, |
| * adding new entries and removing existing entries. |
| * |
| * This class provides a very limited set of operations with simple semantics. |
| * It is intended to be used as a building block to construct more complex |
| * and interesting data structures such as HashMap. Think very hard before |
| * adding anything extra to BasicHashtable, it probably belongs at a |
| * higher level of abstraction. |
| * |
| * TKey: The key type. |
| * TEntry: The entry type which is what is actually stored in the array. |
| * |
| * TKey must support the following contract: |
| * bool operator==(const TKey& other) const; // return true if equal |
| * bool operator!=(const TKey& other) const; // return true if unequal |
| * |
| * TEntry must support the following contract: |
| * const TKey& getKey() const; // get the key from the entry |
| * |
| * This class supports storing entries with duplicate keys. Of course, it can't |
| * tell them apart during removal so only the first entry will be removed. |
| * We do this because it means that operations like add() can't fail. |
| */ |
| template <typename TKey, typename TEntry> |
| class BasicHashtable : private BasicHashtableImpl { |
| public: |
| /* Creates a hashtable with the specified minimum initial capacity. |
| * The underlying array will be created when the first entry is added. |
| * |
| * minimumInitialCapacity: The minimum initial capacity for the hashtable. |
| * Default is 0. |
| * loadFactor: The desired load factor for the hashtable, between 0 and 1. |
| * Default is 0.75. |
| */ |
| BasicHashtable(size_t minimumInitialCapacity = 0, float loadFactor = 0.75f); |
| |
| /* Copies a hashtable. |
| * The underlying storage is shared copy-on-write. |
| */ |
| BasicHashtable(const BasicHashtable& other); |
| |
| /* Clears and destroys the hashtable. |
| */ |
| virtual ~BasicHashtable(); |
| |
| /* Making this hashtable a copy of the other hashtable. |
| * The underlying storage is shared copy-on-write. |
| * |
| * other: The hashtable to copy. |
| */ |
| inline BasicHashtable<TKey, TEntry>& operator =(const BasicHashtable<TKey, TEntry> & other) { |
| setTo(other); |
| return *this; |
| } |
| |
| /* Returns the number of entries in the hashtable. |
| */ |
| inline size_t size() const { |
| return mSize; |
| } |
| |
| /* Returns the capacity of the hashtable, which is the number of elements that can |
| * added to the hashtable without requiring it to be grown. |
| */ |
| inline size_t capacity() const { |
| return mCapacity; |
| } |
| |
| /* Returns the number of buckets that the hashtable has, which is the size of its |
| * underlying array. |
| */ |
| inline size_t bucketCount() const { |
| return mBucketCount; |
| } |
| |
| /* Returns the load factor of the hashtable. */ |
| inline float loadFactor() const { |
| return mLoadFactor; |
| }; |
| |
| /* Returns a const reference to the entry at the specified index. |
| * |
| * index: The index of the entry to retrieve. Must be a valid index within |
| * the bounds of the hashtable. |
| */ |
| inline const TEntry& entryAt(size_t index) const { |
| return entryFor(bucketAt(mBuckets, index)); |
| } |
| |
| /* Returns a non-const reference to the entry at the specified index. |
| * |
| * index: The index of the entry to edit. Must be a valid index within |
| * the bounds of the hashtable. |
| */ |
| inline TEntry& editEntryAt(size_t index) { |
| edit(); |
| return entryFor(bucketAt(mBuckets, index)); |
| } |
| |
| /* Clears the hashtable. |
| * All entries in the hashtable are destroyed immediately. |
| * If you need to do something special with the entries in the hashtable then iterate |
| * over them and do what you need before clearing the hashtable. |
| */ |
| inline void clear() { |
| BasicHashtableImpl::clear(); |
| } |
| |
| /* Returns the index of the next entry in the hashtable given the index of a previous entry. |
| * If the given index is -1, then returns the index of the first entry in the hashtable, |
| * if there is one, or -1 otherwise. |
| * If the given index is not -1, then returns the index of the next entry in the hashtable, |
| * in strictly increasing order, or -1 if there are none left. |
| * |
| * index: The index of the previous entry that was iterated, or -1 to begin |
| * iteration at the beginning of the hashtable. |
| */ |
| inline ssize_t next(ssize_t index) const { |
| return BasicHashtableImpl::next(index); |
| } |
| |
| /* Finds the index of an entry with the specified key. |
| * If the given index is -1, then returns the index of the first matching entry, |
| * otherwise returns the index of the next matching entry. |
| * If the hashtable contains multiple entries with keys that match the requested |
| * key, then the sequence of entries returned is arbitrary. |
| * Returns -1 if no entry was found. |
| * |
| * index: The index of the previous entry with the specified key, or -1 to |
| * find the first matching entry. |
| * hash: The hashcode of the key. |
| * key: The key. |
| */ |
| inline ssize_t find(ssize_t index, hash_t hash, const TKey& key) const { |
| return BasicHashtableImpl::find(index, hash, &key); |
| } |
| |
| /* Adds the entry to the hashtable. |
| * Returns the index of the newly added entry. |
| * If an entry with the same key already exists, then a duplicate entry is added. |
| * If the entry will not fit, then the hashtable's capacity is increased and |
| * its contents are rehashed. See rehash(). |
| * |
| * hash: The hashcode of the key. |
| * entry: The entry to add. |
| */ |
| inline size_t add(hash_t hash, const TEntry& entry) { |
| return BasicHashtableImpl::add(hash, &entry); |
| } |
| |
| /* Removes the entry with the specified index from the hashtable. |
| * The entry is destroyed immediately. |
| * The index must be valid. |
| * |
| * The hashtable is not compacted after an item is removed, so it is legal |
| * to continue iterating over the hashtable using next() or find(). |
| * |
| * index: The index of the entry to remove. Must be a valid index within the |
| * bounds of the hashtable, and it must refer to an existing entry. |
| */ |
| inline void removeAt(size_t index) { |
| BasicHashtableImpl::removeAt(index); |
| } |
| |
| /* Rehashes the contents of the hashtable. |
| * Grows the hashtable to at least the specified minimum capacity or the |
| * current number of elements, whichever is larger. |
| * |
| * Rehashing causes all entries to be copied and the entry indices may change. |
| * Although the hash codes are cached by the hashtable, rehashing can be an |
| * expensive operation and should be avoided unless the hashtable's size |
| * needs to be changed. |
| * |
| * Rehashing is the only way to change the capacity or load factor of the |
| * hashtable once it has been created. It can be used to compact the |
| * hashtable by choosing a minimum capacity that is smaller than the current |
| * capacity (such as 0). |
| * |
| * minimumCapacity: The desired minimum capacity after rehashing. |
| * loadFactor: The desired load factor after rehashing. |
| */ |
| inline void rehash(size_t minimumCapacity, float loadFactor) { |
| BasicHashtableImpl::rehash(minimumCapacity, loadFactor); |
| } |
| |
| /* Determines whether there is room to add another entry without rehashing. |
| * When this returns true, a subsequent add() operation is guaranteed to |
| * complete without performing a rehash. |
| */ |
| inline bool hasMoreRoom() const { |
| return mCapacity > mFilledBuckets; |
| } |
| |
| protected: |
| static inline const TEntry& entryFor(const Bucket& bucket) { |
| return reinterpret_cast<const TEntry&>(bucket.entry); |
| } |
| |
| static inline TEntry& entryFor(Bucket& bucket) { |
| return reinterpret_cast<TEntry&>(bucket.entry); |
| } |
| |
| virtual bool compareBucketKey(const Bucket& bucket, const void* __restrict__ key) const; |
| virtual void initializeBucketEntry(Bucket& bucket, const void* __restrict__ entry) const; |
| virtual void destroyBucketEntry(Bucket& bucket) const; |
| |
| private: |
| // For dumping the raw contents of a hashtable during testing. |
| friend class BasicHashtableTest; |
| inline uint32_t cookieAt(size_t index) const { |
| return bucketAt(mBuckets, index).cookie; |
| } |
| }; |
| |
| template <typename TKey, typename TEntry> |
| BasicHashtable<TKey, TEntry>::BasicHashtable(size_t minimumInitialCapacity, float loadFactor) : |
| BasicHashtableImpl(sizeof(TEntry), traits<TEntry>::has_trivial_dtor, |
| minimumInitialCapacity, loadFactor) { |
| } |
| |
| template <typename TKey, typename TEntry> |
| BasicHashtable<TKey, TEntry>::BasicHashtable(const BasicHashtable<TKey, TEntry>& other) : |
| BasicHashtableImpl(other) { |
| } |
| |
| template <typename TKey, typename TEntry> |
| BasicHashtable<TKey, TEntry>::~BasicHashtable() { |
| dispose(); |
| } |
| |
| template <typename TKey, typename TEntry> |
| bool BasicHashtable<TKey, TEntry>::compareBucketKey(const Bucket& bucket, |
| const void* __restrict__ key) const { |
| return entryFor(bucket).getKey() == *static_cast<const TKey*>(key); |
| } |
| |
| template <typename TKey, typename TEntry> |
| void BasicHashtable<TKey, TEntry>::initializeBucketEntry(Bucket& bucket, |
| const void* __restrict__ entry) const { |
| if (!traits<TEntry>::has_trivial_copy) { |
| new (&entryFor(bucket)) TEntry(*(static_cast<const TEntry*>(entry))); |
| } else { |
| memcpy(&entryFor(bucket), entry, sizeof(TEntry)); |
| } |
| } |
| |
| template <typename TKey, typename TEntry> |
| void BasicHashtable<TKey, TEntry>::destroyBucketEntry(Bucket& bucket) const { |
| if (!traits<TEntry>::has_trivial_dtor) { |
| entryFor(bucket).~TEntry(); |
| } |
| } |
| |
| }; // namespace android |
| |
| #endif // ANDROID_BASIC_HASHTABLE_H |