AVL tree

Source: Wikipedia, the free encyclopedia.

This is an old revision of this page, as edited by Ruud Koot (talk | contribs) at 14:07, 26 July 2016 (Reverted edits by NNcNannara (talk) to last version by 78.8.99.245). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

AVL tree
Typetree
Invented1962
Invented byGeorgy Adelson-Velsky and Evgenii Landis
Time complexity in big O notation
Operation Average Worst case
Search O(log n)[1] O(log n)[1]
Insert O(log n)[1] O(log n)[1]
Delete O(log n)[1] O(log n)[1]
Space complexity
Space O(n) O(n)
Fig. 1: AVL tree with balance factors (green)

In

child subtrees of any node differ by at most one; if at any time they differ by more than one, rebalancing is done to restore this property. Lookup, insertion, and deletion all take O(log n) time in both the average and worst cases, where n is the number of nodes in the tree prior to the operation. Insertions and deletions may require the tree to be rebalanced by one or more tree rotations
.

The AVL tree is named after its two Soviet inventors, Georgy Adelson-Velsky and Evgenii Landis, who published it in their 1962 paper "An algorithm for the organization of information".[3]

AVL trees are often compared with red–black trees because both support the same set of operations and take O(log n) time for the basic operations. For lookup-intensive applications, AVL trees are faster than red–black trees because they are more rigidly balanced.[4] Similar to red–black trees, AVL trees are height-balanced. Both are in general not weight-balanced nor μ-balanced for any μ≤12;[5] that is, sibling nodes can have hugely differing numbers of descendants.

Languages

  1. AVL Trees in Java
  2. AVL Trees in C#
  3. AVL Trees in C++

Definition

Balance factor

In a binary tree the balance factor of a node N is defined to be the height difference

BalanceFactor(N) := –Height(LeftSubtree(N)) + Height(RightSubtree(N)) [6]

of its two child subtrees. A binary tree is called AVL tree if the

invariant

BalanceFactor(N) ∈ {–1,0,+1}

holds for every node N in the tree.

A node N with BalanceFactor(N) < 0 is called "left-heavy", one with BalanceFactor(N) > 0 is called "right-heavy", and one with BalanceFactor(N) = 0 is sometimes simply called "balanced".

Remark

In the sequel, because there is a one-to-one correspondence between nodes and the subtrees rooted by them, we sometimes let it to the context whether the name of an object stands for the node or the subtree.

Properties

Balance factors can be kept up-to-date by knowing the previous balance factors and the change in height – it is not necessary to know the absolute height. For holding the AVL balance information, two bits per node are sufficient.[7]

The height h of an AVL tree with n nodes lies in the interval:[8]

log2(n+1) ≤ h < c log2(n+2)+b

with the golden ratio φ := (1+5) ⁄2 ≈ 1.618, c := 1⁄ log2 φ ≈ 1.44,  and  b := c2 log2 5 – 2 ≈ –0.328. This is because an AVL tree of height h contains at least Fh+2 – 1 nodes where {Fh} is the

Fibonacci sequence
with the seed values F1 = 1, F2 = 1.

Data structure

According to the original paper "An algorithm for the organization of information" AVL trees have been invented as binary search trees. In that sense they are a data structure together with its major associated operations, namely search, insert, delete, which rely on and maintain the AVL property. In that sense, the AVL tree is a "self-balancing binary search tree".

Operations

Read-only operations of an AVL tree involve carrying out the same actions as would be carried out on an unbalanced binary search tree, but modifications have to observe and restore the height balance of the subtrees.

Searching

Searching for a specific key in an AVL tree can be done the same way as that of a normal unbalanced binary search tree. In order for search to work effectively it has to employ a comparison function which establishes a total order (or at least a total preorder) on the set of keys. The number of comparisons required for successful search is limited by the height h and for unsuccessful search is very close to h, so both are in O(log n).

Traversal

Once a node has been found in an AVL tree, the next or previous node can be accessed in

amortized
constant time. Some instances of exploring these "nearby" nodes require traversing up to h ∝ log(n) links (particularly when navigating from the rightmost leaf of the root’s left subtree to the root or from the root to the leftmost leaf of the root’s right subtree; in the AVL tree of figure 1, moving from node P to the next but one node Q takes 3 steps). However, exploring all n nodes of the tree in this manner would visit each link exactly twice: one downward visit to enter the subtree rooted by that node, another visit upward to leave that node’s subtree after having explored it. And since there are n−1 links in any tree, the amortized cost is found to be 2×(n−1)/n, or approximately 2.

Comparison to other structures

Both AVL trees and red–black trees are self-balancing binary search trees and they are very similar mathematically.[9] The operations to balance the trees are different, but both occur on the average in O(1) with maximum in O(log n). The real difference between the two is the limiting height.

For a tree of size n ≥ 1

  • an AVL tree’s height is at most
where   the golden ratio,   and  .
  • a red–black tree’s height is at most
    [10]

AVL trees are more rigidly balanced than red–black trees, leading to faster retrieval but slower insertion and deletion.

See also

References

  1. ^ a b c d e f Eric Alexander. "AVL Trees".
  2. ^ Robert Sedgewick, Algorithms, Addison-Wesley, 1983, ISBN 0-201-06672-6, page 199, chapter 15: Balanced Trees.
  3. ^ Georgy Adelson-Velsky, G.; Evgenii Landis (1962). "An algorithm for the organization of information". Proceedings of the USSR Academy of Sciences (in Russian). 146: 263–266. English translation by Myron J. Ricci in Soviet Math. Doklady, 3:1259–1263, 1962.
  4. Stanford University
    .
  5. ^ AVL trees are not weight-balanced? (meaning: AVL trees are not μ-balanced?)
    Thereby: A Binary Tree is called -balanced, with , if for every node , the inequality
    holds and is minimal with this property. is the number of nodes below the tree with as root (including the root) and is the left child node of .
  6. .
  7. ^ More precisely: if the AVL balance information is kept in the child nodes – with meaning "when going upward there is an additional increment in height", this can be done with one bit. Nevertheless, the modifying operations can be programmed more efficiently if the balance information can be checked with one test.
  8. .
  9. ^ In fact, every AVL tree can be colored red–black.
  10. ^ Red–black tree#Proof of asymptotic bounds

Further reading

External links