Jump to content

Golomb–Dickman constant

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Ferrellswheeler (talk | contribs) at 21:01, 20 May 2009 (statement about an is expected value was wrong. cleaned up number theory). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In mathematics, the Golomb–Dickman constant arises in the theory of random permutations and in number theory. Its value is

Let an be the average — taken over all permutations of a set of size n — of the length of the longest cycle in each permutation. Then the Golomb–Dickman constant is

In the language of probability theory, is the expected length of the longest cycle in a uniformly distributed random permutation of a set of size n.

In number theory, the Golomb–Dickman constant appears in connection with the average size of the largest prime factor of an integer. More precisely,

where is the largest prime factor of k. So if k is a d digit integer, then is the asymptotic average number of digits of the largest prime factor of k.

A related result is the one hundred prisoners problem in random permutation statistics: asymptotically, the fraction of permutations with a cycle of length greater than n/2 is , or about 69%.

See also

  • Weisstein, Eric W. "Golomb-Dickman Constant". MathWorld.
  • Sloane, N. J. A. (ed.). "Sequence A084945". The On-Line Encyclopedia of Integer Sequences. OEIS Foundation.
  • Finch, Steven R. (2003). Mathematical Constants. Cambridge University Press. pp. 284–286. ISBN 0521818052.