Heaps are a data structure for supporting the priority queue operations insert and find-minimum. They work by maintaining a partial order on the set of elements that is weaker than the sorted order (so it can be efficient to maintain), yet stronger than random order (so the minimum element can be quickly identified).
Heap-Labeled Tree
A heap-labeled tree is defined to be a binary tree such that the key of each node dominates the keys of its children.
- In a min-heap, a node dominates its children by having a smaller key than they do
- In a max-heap, a node dominates its children by having a bigger key than they do
The most natural implementation of this binary tree would store each key in a node with pointers to its two children. However, similar to binary search trees, the memory used by the pointers can easily outweigh the size of keys, which is the data we’re mostly interested in.
Sometimes, “heap” is used to refer to a heap-labeled tree.
Heap Array
The heap is a slick data structure that enables us to represent binary trees without using any pointers. We store data as as an array of keys, and use the position of the keys to implicitly play the role of pointers. We assume that the array starts with index 1 for simplicity.
We store the root of the tree at the first position of the array, and its left and right children in the second and third positions. In general, we store the keys of the th level of a complete binary tree from left to right in positions to , as shown above.
- Level 1 (1492): key, stored at position
- Level 2 (1783, 1776): keys, stored between positions and
- Level 3 (1804, 1865, 1945, 1963): keys, stored between positions and .
What is especially nice about this representation is that the positions of the parent and children of the key at position are readily determined. Thus, we can move around the tree without any pointers.
- The left child of sits in position and the right child in
- The parent of k holds court in position .
This approach means that we can store any binary tree in an array without pointers. What is the catch? Suppose our height tree was sparse, meaning that the number of nodes . All of the missing internal nodes still take up space in our structure, since we must represent a full binary tree to maintain the positional mapping between parents and children.
To avoid these holes and ensure space efficiency, we make each level be packed as much as it can be, such that:
- Only the last level may be incomplete.
- The elements of the last level as far left as possible.
Thus, we can represent an -key tree using the first elements of the array. If we did not enforce these structural constraints, we might need an array of size to store elements.
With heaps, all but the last level are filled, so the height of an element heap is logarithmic because:
Benefits/Drawbacks
This implicit representation of binary trees saves memory, but is less flexible than using pointers.
- We cannot store arbitrary tree topologies without wasting large amounts of space.
- We cannot move subtrees around by just changing a single pointer, only by explicitly moving all the elements in the subtree.
This loss of flexibility explains why we cannot use this idea to represent binary search trees, but it works just fine for heaps.
How can we efficiently search for a particular key in a heap? We can’t. Binary search does not work because a heap is not a binary search tree, and the array form of the heap is not sorted. We know almost nothing about the relative order of the leaf elements in a heap that would let us avoid doing linear search through them.
Heap Construction
Heaps can be constructed incrementally, by inserting each new element into the left-most open spot in the array (st position of a previously -element heap). This ensures the desired balanced shape of the heap-labeled tree, but does not maintain the dominance ordering of the keys. The new key might be less than its parent in a min-heap, or greater than its parent in a max-heap, which we don’t want.
The solution to this is to swap any dissatisfied element with its parent:
- The old parent is now happy, as it is properly dominated
- The other child of the old parent is happy, because it is now dominated by an element even more extreme than before
- The new element is now happier, but may still dominate its new parent. So we recur at a higher level, bubbling up the new key to its proper position in the hierarchy.
- Since we replace the root of a subtree by a larger one at each step of bubbling up, we preserve the heap order elsewhere.
This swap process takes constant time at each level. Since the height of an -element heap is , each insertion takes at most time. A heap of elements can thus be constructed in time through such insertions:
Find and Delete Min/Max
The remaining priority queue operations are identifying and deleting the dominant element (min or max). Identification is easy, since the top of the heap sits in the first position of the array.
Removing the top elements leaves a hole in the array, which can be filled by moving the right-most leaf (sitting in the th position of the array) into the first position. This restores the shape of the tree, but the labeling of the root might not satisfy the heap property. The new root might be dominated by both of its children.
For example, the root of a min-heap should be the smallest of 3 elements: the current root and its 2 children. The cases are:
- If the current root is dominant, the heap order has been restored.
- If not, the dominant child should be swapped with the root, and the problem is pushed down to its next level.
The dissatisfied element bubbles down the heap until it dominates all its children. This operation is called heapify, because it merges two heaps (the subtrees below the original root) with a new key.
We will reach a leaf after steps of bubble_down
, each constant time. Thus, root deletion is completed in time.
Repeatedly exchanging the maximum element with the last element and caling heapify yields an sorting algorithm, heapsort.
Faster Heap Construction
As we saw previously, a heap can be constructed on elements by incremental insertion in time. We can actually construct them even faster by using our bubble_down
procedure above.
Suppose we pack keys into the first elements of our priority queue array. The shape of our heap will be right, but the dominance order will be wrong. How can we restore it?
Consider the array in reverse order, starting from the last (th) position. It represents a leaf of the tree and so dominates its non-existent children. The same is the case for the last positions in the array, because all are leaves. If we continue to walk backwards through the array we will eventually encounter an internal node with children. This element may not dominate its children, but its children represent well-formed (if small) heaps.
This is exactly the situation the bubble_down
procedure was designed to handle, restoring the heap order of an arbitrary root element sitting on top of two sub-heaps. Thus, we can create a heap by performing non-trivial calls to the bubble down procedure:
Multiplying the number of calls to bubble down () times an upper bound on the cost of each operation () gives us a running time analysis of . This would make it the same speed as the incremental insertion algorithm described above. But note that it is indeed an upper bound, because only the last insertion will actually take steps. Recall that bubble down takes time proportional to the height of the heaps it is merging. Most of these heaps are extremely small. In a full binary tree on nodes, there are nodes that are leaves (i.e. height 0), nodes that are height 1, nodes that are height 2, and so on. In general, there are at most nodes of height , so the cost of building a heap is:
Since this sum is not quite a geometric series, we can’t apply the usual identity to get the sum, but the puny contribution of the numerator () is crushed by the denominator (). The series quickly converges to linear.
Does it matter that we can construct heaps in linear time instead of ? Not really. The construction time did not dominate the complexity of heapsort, so improving the construction time does not improve its worst-case performance. Still, it is an impressive display of the power of careful analysis, and the free lunch that geometric series convergence can sometimes provide.