Why is the Fibonacci Sequence Big O(2^n) instead of O(logn)?

I took discrete math (in which I learned about master theorem, Big Theta/Omega/O) a while ago and I seem to have forgotten the difference between O(logn) and O(2^n) (not in the theoretical sense of Big Oh). I generally understand that algorithms like merge and quick sort are O(nlogn) because they repeatedly divide the initial input array into sub arrays until each sub array is of size 1 before recursing back up the tree, giving a recursion tree that is of height logn + 1. But if you calculate the height of a recursive tree using n/b^x = 1 (when the size of the subproblem has become 1 as was stated in an answer here) it seems that you always get that the height of the tree is log(n).

If you solve the Fibonacci sequence using recursion, I would think that you would also get a tree of size logn, but for some reason, the Big O of the algorithm is O(2^n). I was thinking that maybe the difference is because you have to remember the all of the fib numbers for each subproblem to get the actual fib number meaning that the value at each node has to be recalled, but it seems that in merge sort, the value of each node has to be used (or at least sorted) as well. This is unlike binary search, however, where you only visit certain nodes based on comparisons made at each level of the tree so I think this is where the confusion is coming from.

So specifically, what causes the Fibonacci sequence to have a different time complexity than algorithms like merge/quick sort?


The other answers are correct, but don't make it clear - where does the large difference between the Fibonacci algorithm and divide-and-conquer algorithms come from? Indeed, the shape of the recursion tree for both classes of functions is the same - it's a binary tree.

The trick to understand is actually very simple: consider the size of the recursion tree as a function of the input size n .

Recall some basic facts about binary trees first:

  • The number of leaves n is a binary tree is equal to the the number of non-leaf nodes plus one. The size of a binary tree is therefore 2n-1.
  • In a perfect binary tree, all non-leaf nodes have two children.
  • The height h for a perfect binary tree with n leaves is equal to log(n) , for a random binary tree: h = O(log(n)) , and for a degenerate binary tree h = n-1 .
  • Intuitively:

  • For sorting an array of n elements with a recursive algorithm, the recursion tree has n leaves . It follows that the width of the tree is n , the height of the tree is O(log(n)) on the average and O(n) in the worst case.

  • For calculating a Fibonacci sequence element k with the recursive algorithm, the recursion tree has k levels (to see why, consider that fib(k) calls fib(k-1) , which calls fib(k-2) , and so on). It follows that height of the tree is k . To estimate a lower-bound on the width and number of nodes in the recursion tree, consider that since fib(k) also calls fib(k-2) , therefore there is a perfect binary tree of height k/2 as part of the recursion tree. If extracted, that perfect subtree would have 2k/2 leaf nodes. So the width of the recursion tree is at least O(2^{k/2}) or, equivalently, 2^O(k) .

  • The crucial difference is that:

  • for divide-and-conquer algorithms, the input size is the width of the binary tree.
  • for the Fibonnaci algorithm, the input size is it the height of the tree.
  • Therefore the number of nodes in the tree is O(n) in the first case, but 2^O(n) in the second. The Fibonacci tree is much larger compared to the input size.

    You mention Master theorem; however, the theorem cannot be applied to analyze the complexity of Fibonacci because it only applies to algorithms where the input is actually divided at each level of recursion. Fibonacci does not divide the input; in fact, the functions at level i produce almost twice as much input for the next level i+1 .


    To address the core of the question, that is "why Fibonacci and not Mergesort", you should focus on this crucial difference:

  • The tree you get from Mergesort has N elements for each level, and there are log(n) levels.
  • The tree you get from Fibonacci has N levels because of the presence of F(n-1) in the formula for F(N), and the number of elements for each level can vary greatly: it can be very low (near the root, or near the lowest leaves) or very high. This, of course, is because of repeated computation of the same values.
  • To see what I mean by "repeated computation", look at the tree for the computation of F(6):

    fib(6)计算树
    Fibonacci tree picture from: http://composingprograms.com/pages/28-efficiency.html

    How many times do you see F(3) being computed?


    Consider the following implementation

    int fib(int n)
    {
        if(n < 2)
          return n;
        return fib(n-1) + fib(n-2)
    }
    

    Let's denote T(n) the number of operations that fib performs to calculate fib(n) . Because fib(n) is calling fib(n-1) and fib(n-2) , it means that T(n) is at least T(n-1) + T(n-2) . This in turn means that T(n) > fib(n) . There is a direct formula of fib(n) which is some constant to the power of n . Therefore T(n) is at least exponential. QED.

    链接地址: http://www.djcxy.com/p/40038.html

    上一篇: O素数函数的符号?

    下一篇: 为什么斐波那契序列大O(2 ^ n)而不是O(logn)?