Big O notation

The Big O notation is a mathematical notation used to describe the asymptotic behavior of functions. More precisely, it is used to describe an asymptotic upper bound for the magnitude of a function in terms of another, usually simpler, function.
In mathematics, it is usually used to characterize the residual term of a truncated infinite series, especially an asymptotic series. In computer science, it is useful in the analysis of the complexity of algorithms.
It was first introduced by German number theorist Paul Bachmann in his 1892 book Analytische Zahlentheorie. The notation was popularized in the work of another German number theorist Edmund Landau, hence it is sometimes called a Landau symbol. The bigO, standing for "order of", was originally a capital omicron; today the capital letter O is used, but never the digit zero.
Contents 
Uses
There are two formally close, but noticeably different usages of this notation: infinite asymptotics and infinitesimal asymptotics. This distinction is only in application and not in principle, however—the formal definition for the "big O" is the same for both cases, only with different limits for the function argument.
Infinite asymptotics
Big O notation is useful when analyzing algorithms for efficiency. For example, the time (or the number of steps) it takes to complete a problem of size n might be found to be T(n) = 4n^{2}  2n + 2.
As n grows large, the n^{2} term will come to dominate, so that all other terms can be neglected. Further, the coefficients will depend on the precise details of the implementation and the hardware it runs on, so they should also be neglected. Big O notation captures what remains: we write
 <math>T(n)\in O(n^2)<math>
and say that the algorithm has order of n^{2} time complexity.
Infinitesimal asymptotics
Big O can also be used to describe the error term in an approximation to a mathematical function. For example,
 <math>e^x=1+x+x^2/2+\hbox{O}(x^3)\qquad\hbox{as}\ x\to 0<math>
expresses the fact that the error is smaller in absolute value than some constant times x^{3} if x is close enough to 0.
Formal definition
Suppose f(x) and g(x) are two functions defined on some subset of the real numbers. We say
 f(x) is O(g(x)) as x <math> \to <math> ∞
if and only if
 there exist numbers x_{0} and M such that f(x) ≤ M g(x) for x > x_{0}.
The notation can also be used to describe the behavior of f near some real number a: we say
 f(x) is O(g(x)) as x <math> \to <math> a
if and only if
 there exists numbers δ>0 and M such that f(x) ≤ M g(x) for x  a < δ.
If g(x) is nonzero for values of x sufficiently close to a, both of these definitions can be unified using the limit superior:
 f(x) is O(g(x)) as x <math> \to <math> a
if and only if
 <math>\limsup_{x\to a} \left\frac{f(x)}{g(x)}\right < \infty.<math>
Missing image
Yorick215.PNG
Image:Yorick215.PNG
In mathematics, both asymptotic behaviors near ∞ and near a are considered. In computational complexity theory, only asymptotics near ∞ are used; furthermore, only positive functions are considered, so the absolute value bars may be left out.
Example
Take the polynomials:
 <math> f(x) = 6x^4 2x^3 +5 \,<math>
 <math> g(x) = x^4. \,<math>
We say f(x) has order O(g(x)) or O(x^{4}). From the definition of order, f(x) ≤ C g(x) for all x>1, where C is a constant.
Proof:
 <math> 6x^4  2x^3 + 5 \le 6x^4 + 2x^3 + 5 \,<math> where x > 1
 <math> 6x^4  2x^3 + 5 \le 6x^4 + 2x^4 + 5x^4 \,<math> because x^{3} < x^{4}, and so on.
 <math> 6x^4  2x^3 + 5 \le 13x^4 \,<math>
 <math> 6x^4  2x^3 + 5 \le 13 \,x^4 . \,<math>
Matters of notation
The statement "f(x) is O(g(x))" as defined above is often written as f(x) = O(g(x)). This is a slight abuse of notation: we are not really asserting the equality of two functions. Here is an example illustrating why the equals sign is inappropriate:
 <math>O(x)=O(x^2)\ \ \mbox{but}\ \ O(x^2)\, \not=\, O(x).<math>
By this reason, some authors prefer a set notation and write f <math>\in<math> O(g), thinking of O(g) as the set of all functions dominated by g.
Furthermore, an "equation" of the form
 f(x) = h(x) + O(g(x))
is to be understood as "the difference of f(x) and h(x) is O(g(x))".
Common orders of functions
Here is a list of classes of functions that are commonly encountered when analyzing algorithms. All of these are as n increases to infinity. The slowergrowing functions are listed first. c is an arbitrary constant.
notation  name 

<math>O(1)<math>  constant 
<math>O(\log n)<math>  logarithmic 
<math>O([\log n]^c)<math>  polylogarithmic 
<math>o(n)<math>  sublinear 
<math>O(n)<math>  linear 
<math>O(n \log n)<math>  linearithmic, quasilinear or supralinear 
<math>O(n^2)<math>  quadratic 
<math>O(n^c),~c > 1<math>  polynomial, sometimes called "algebraic" 
<math>O(c^n)<math>  exponential, sometimes called "geometric" 
<math>O(n!)<math>  factorial 
Properties
If a function f(n) can be written as a finite sum of other functions, then the fastest growing one determines the order of f(n). For example
 <math>f(n) = 10 \log n + 5 (\log n)^3 + 7n + 3n^2 + 6n^3 = \hbox{O}(n^3)\,\!<math>.
In particular, if a function may be bounded by a polynomial in n, then as n tends to infinity, one may disregard lowerorder terms of the polynomial.
O(n^{c}) and O(c^{n}) are very different. The latter grows much, much faster, no matter how big the constant c is (so long as it is greater than one). A function that grows faster than any power of n is called superpolynomial. One that grows slower than an exponential function of the form c^{n} is called subexponential. An algorithm can require time that is both superpolynomial and subexponential; examples of this include the fastest algorithms known for integer factorization.
O(log n) is exactly the same as O(log(n^{c})). The logarithms differ only by a constant factor, (since log(n^{c})=c log n) and thus the big O notation ignores that. Similarly, logs with different constant bases are equivalent.
Product
 <math>O(f(n)) O(g(n)) = O(f(n)g(n)) \,<math>
Multiplication by a constant
 <math>O(K\cdot g(n)) = O(g(n))<math>
Sum
 <math>O(f(n)) + O(g(n)) = O(\max \lbrace f(n),g(n) \rbrace) \,<math>
Other useful relations are given in section Big O and little o below.
Related asymptotic notations: O, o, Ω, ω, Θ, Õ
Big O is the most commonly used asymptotic notation for comparing functions, although it is often actually an informal substitute for Θ (Theta, see below). Here, we define some related notations in terms of "big O":
Notation  Definition  Mathematical definition 

<math>f(n) \in O(g(n))<math>  asymptotic upper bound  <math>\limsup_{x \to \infty} \left\frac{f(x)}{g(x)}\right < \infty<math> 
<math>f(n) \in o(g(n))<math>  asymptotically negligible  <math>\lim_{x \to \infty} \frac{f(x)}{g(x)} = 0<math> 
<math>f(n) \in \Omega(g(n))<math>  asymptotic lower bound  <math> \liminf_{x \to \infty} \left\frac{f(x)}{g(x)}\right > 0 <math> 
<math>f(n) \in \omega(g(n))<math>  asymptotically dominant  <math>\lim_{x \to \infty} \frac{g(x)}{f(x)} = 0<math> 
<math>f(n) \in \Theta(g(n))<math>  asymptotically tight bound  <math>f\in O(g)<math> and <math>g\in O(f)<math> 
(A mnemonic for these Greek letters is that "omicron" can be read "omicron", i.e., "osmall", whereas "omega" can be read "omega" or "obig".)
The relation f(n) = o(g(n)) is read as "f(n) is littleoh of g(n)". Intuitively, it means that g(n) grows much faster than f(n). Formally, it states that the limit of f(n)/g(n) is zero.
Aside from bigO, the notations Θ and Ω are the two most often used in computer science; the lowercase o is common in mathematics but rarer in computer science. The lowercase ω is rarely used.
In casual use, O is commonly used where Θ is meant, i.e., when a tight estimate is implied. For example, one might say "heapsort is O(n log n) in the average case" when the intended meaning was "heapsort is Θ(n log n) in the average case". Both statements are true, but the latter is a stronger claim.
Another notation sometimes used in computer science is Õ (read SoftO). f(n) = Õ(g(n)) is shorthand for f(n) = O(g(n) log^{k}g(n)) for some k. Essentially, it is BigO, ignoring logarithmic factors. This notation is often used to describe a class of "nitpicking" estimates (since log^{k}n is always o(n) for any constant k).
Big O and little o
The following properties can be useful:
 <math>o(f) = O(f)<math>
 <math>O(O(f)) = O(f)<math>
 <math>o(O(f)) = o(f)<math> and <math>O(o(f)) = o(f)<math> (and thus also <math>o(o(f)) = o(f)<math>)
 <math>O(f) + O(f) = O(f)<math> (and thus also <math>O(f) + o(f) = O(f)<math>)
 <math>o(f) + o(f) = o(f)<math>
 <math>O(f) O(g) = O(fg)<math>
 <math>O(f) o(g) = o(fg)<math> (and thus also <math>o(f) o(g) = o(fg)<math>)
These formulas should be read as follows, e.g. for the last one:
 If <math>f' = O(f)<math> and <math>g' = o(g)<math>, then <math>f' g' = o(fg)<math>.
Warning: We do not have <math>O(o(f)) = o(O(f))<math>, which would mean that if <math>f' = o(f)<math> and <math>g' = O(f)<math>, then <math>h = O(f')<math> would imply <math>h = o(g')<math> (which is not true). A simple counter example is given by taking <math>g' = o<math> (the null function), and <math>h = f' = f^2, f(n)=1/n<math>.
Once again, here "=" does not means equality, but rather "∈", and there are counterexamples for each of the above relations if they are read from the right to the left.
Multiple variables
Big O (and little o, and Ω...) can also be used with multiple variables. For example, the statement
 <math>f(n,m) = n^2 + m^3 + \hbox{O}(n+m) \mbox{ as } n,m\to\infty<math>
asserts that there exist constants C and N such that
 <math>\forall n, m>N:f(n,m) \le n^2 + m^3 + C(n+m).<math>
To avoid ambiguity, the running variable should always be specified: the statement
 <math>f(n,m) = \hbox{O}(n^m) \mbox{ as } n,m\to\infty<math>
is quite different from
 <math>\forall m: f(n,m) = \hbox{O}(n^m) \mbox{ as } n\to\infty.<math>
External links
 Cprogramming.com: Algorithm Efficiency (http://www.cprogramming.com/tutorial/computersciencetheory/algorithmicefficiency1.html) An article on the Big O in Cde:LandauSymbole
es:Cota superior asintótica fr:Notations de Landau ko:대문자 O 표기법 ja:ランダウの記号 pl:Duże O sv:Ordo th:สัญกรณ์โอใหญ่ zh:大O符号