Logarithmic growth
From Wikipedia, the free encyclopedia
In mathematics, logarithmic growth describes a phenomenon that whose size or cost can be described as a logarithm function of some input. e.g. y = C log (x). Note that any logarithm base can be used, since one can be converted to another by a fixed constant. Logarithmic growth is very slow and is the inverse of exponential growth.
A familiar example of logarithmic growth is the number of digits needed to represent a number, N, in positional notation, which grows as logb (N), where b is the base of the number system used, e.g. 10 for decimal arithmetic. Another example is in cryptography, where the key size needed to protect against a brute force attack for a certain period of time grows logarithmically with the desired protection interval.
In the design of computer algorithms, logarithmic growth, and related variants, such as log-linear, or linearithmic, growth are very desirable indications of efficiency.
Logarithmic growth can lead to apparent paradoxes, as in the martingale roulette system, where the potential winnings before bankruptcy grow as the logarithm of the gambler's bankroll. It also plays a role in the St. Petersburg paradox.
[edit] See also
- Iterated logarithm - an even slower growth model