Open main menu

L-notation is an asymptotic notation analogous to big-O notation, denoted as for a bound variable tending to infinity. Like big-O notation, it is usually used to roughly convey the computational complexity of a particular algorithm.



It is defined as


where c is a positive constant, and   is a constant  .

L-notation is used mostly in computational number theory, to express the complexity of algorithms for difficult number theory problems, e.g. sieves for integer factorization and methods for solving discrete logarithms. The benefit of this notation is that it simplifies the analysis of these algorithms. The   expresses the dominant term, and the   takes care of everything smaller.

When   is 0, then


is a polynomial function of ln n; when   is 1 then


is a fully exponential function of ln n (and thereby polynomial in n).

If   is between 0 and 1, the function is subexponential of ln n (and superpolynomial).


Many general-purpose integer factorization algorithms have subexponential time complexities. The best is the general number field sieve, which has an expected running time of


for  . The best such algorithm prior to the number field sieve was the quadratic sieve which has running time


For the elliptic curve discrete logarithm problem, the fastest general purpose algorithm is the baby-step giant-step algorithm, which has a running time on the order of the square-root of the group order n. In L-notation this would be


The existence of the AKS primality test, which runs in polynomial time, means that the time complexity for primality testing is known to be at most


where c has been proven to be at most 6.[1]


L-notation has been defined in various forms throughout the literature. The first use of it came from Carl Pomerance in his paper "Analysis and comparison of some integer factoring algorithms".[2] This form had only the   parameter: the   in the formula was   for the algorithms he was analyzing. Pomerance had been using the letter   (or lower case  ) in this and previous papers for formulae that involved many logarithms.

The formula above involving two parameters was introduced by Arjen Lenstra and Hendrik Lenstra in their article on "Algorithms in Number Theory".[3] It was introduced in their analysis of a discrete logarithm algorithm of Coppersmith. This is the most commonly used form in the literature today.

The Handbook of Applied Cryptography defines the L-notation with a big   around the formula presented in this article.[4] This is not the standard definition. The big   would suggest that the running time is an upper bound. However, for the integer factoring and discrete logarithm algorithms that L-notation is commonly used for, the running time is not an upper bound, so this definition is not preferred.


  1. ^ Hendrik W. Lenstra Jr. and Carl Pomerance, "Primality testing with Gaussian periods", preprint, 2011,
  2. ^ Carl Pomerance, "Analysis and comparison of some integer factoring algorithms", In Mathematisch Centrum Computational Methods in Number Theory, Part 1, pp. 89-139, 1982,
  3. ^ Arjen K. Lenstra and Hendrik W. Lenstra, Jr, "Algorithms in Number Theory", in Handbook of Theoretical Computer Science (vol. A): Algorithms and Complexity, 1991.
  4. ^ Alfred J. Menezes, Paul C. van Oorschot and Scott A. Vanstone. Handbook of Applied Cryptography. CRC Press, 1996. ISBN 0-8493-8523-7.