← Back to home

Logarithmic Distribution

Ah, Wikipedia. The repository of all that is known, and a surprisingly large amount of what’s merely believed to be known. You want me to elaborate on a discrete probability distribution? Fascinating. It’s like asking me to explain the appeal of beige wallpaper. But fine, let’s dive into this Logarithmic distribution. Don't expect me to hold your hand.

Logarithmic Distribution

In the arcane realms of probability and statistics, the logarithmic distribution, a name that suggests a certain elegance but often delivers a stark reality, is a discrete probability distribution. It’s a construct derived, rather unceremoniously, from the Maclaurin series expansion of a rather fundamental function: the natural logarithm. Specifically, it stems from the expansion of ln(1p)-\ln(1-p), which, for those who appreciate such things, unfolds as:

ln(1p)=p+p22+p33+-\ln(1-p) = p + \frac{p^2}{2} + \frac{p^3}{3} + \cdots

This series, seemingly innocent enough, holds the key. From this, a rather critical identity emerges:

k=11ln(1p)  pkk=1\sum_{k=1}^{\infty} \frac{-1}{\ln(1-p)}\; \frac{p^k}{k} = 1

This identity is the bedrock upon which the probability mass function (PMF) of a logarithmic distribution is built. It ensures that the total probability across all possible outcomes sums to one, a rather crucial detail if you intend for your model to reflect reality, or at least a version of it. The PMF, denoted as f(k)f(k), is thus defined as:

f(k) = \frac{-1}{\ln(1-p)}\; \frac{p^k}{k}}

This function dictates the probability of observing a specific outcome, kk. The variable kk here represents the support of the distribution, meaning the set of values it can actually take. For the logarithmic distribution, kk is confined to positive integers: {1,2,3,}\{1, 2, 3, \ldots\}. The parameter pp, a rather vital parameter, is constrained to be strictly between 0 and 1 (0<p<10 < p < 1). If pp were any other value, the entire construction would collapse, much like a poorly supported argument.

Cumulative Distribution Function

Beyond the probability of a single outcome, there's the cumulative distribution function (CDF), denoted as F(k)F(k). This tells you the probability of observing an outcome less than or equal to kk. For the logarithmic distribution, it's expressed using the incomplete beta function, denoted by BB:

F(k)=1+B(p;k+1,0)ln(1p)F(k) = 1 + \frac{\mathrm{B}(p; k+1, 0)}{\ln(1-p)}

The incomplete beta function, B(x;a,b)\mathrm{B}(x; a, b), is itself a complex entity, but here it's employed to provide the cumulative probability. It’s a detail that might escape the casual observer, but it’s essential for a complete understanding.

Properties and Relationships

The logarithmic distribution isn't just some isolated mathematical curiosity. It has connections, some more profound than others. One notable relationship is with the negative binomial distribution. It turns out that if you take a Poisson distribution and compound it with random variables that themselves follow a logarithmic distribution, the resulting distribution is negative binomial.

More formally, let NN be a random variable following a Poisson distribution, and let XiX_i be an infinite sequence of independent and identically distributed random variables, each governed by the Log(pp) distribution. Then, the sum:

i=1NXi\sum_{i=1}^{N}X_{i}

yields a negative binomial distribution. This places the logarithmic distribution within the framework of compound Poisson distributions, suggesting a certain depth and interconnectedness in the world of probability.

Historical Context

The logarithmic distribution wasn't conjured from thin air for the sheer joy of mathematical abstraction. It found its initial application in the work of R. A. Fisher, a name that resonates through the halls of statistics. In a 1943 paper, Fisher and his colleagues employed this distribution to model a phenomenon of significant ecological interest: relative species abundance. It was used to describe how, in a biological population, the number of species tends to decrease as the number of individuals within those species increases. It's a rather stark reflection of the world, isn't it? Some thrive, and many lag behind.

Further Exploration

For those who find themselves inexplicably drawn to the intricacies of this distribution, there are avenues for deeper study. The Poisson distribution, another fundamental entity, also has its roots in a Maclaurin series, hinting at shared origins.

The following resources offer more comprehensive insights:

  • Johnson, Norman Lloyd; Kemp, Adrienne W.; Kotz, Samuel (2005). "Chapter 7: Logarithmic and Lagrangian distributions". Univariate discrete distributions (3rd ed.). John Wiley & Sons. ISBN 978-0-471-27246-5. This is the kind of exhaustive tome that makes you question your life choices, but it’s thorough.
  • Weisstein, Eric W. "Log-Series Distribution". MathWorld. A digital compendium of mathematical lore.

And for the sheer joy of categorization, the logarithmic distribution finds its place among a vast array of other probability distributions, a veritable zoo of statistical models. It’s listed under discrete, univariate distributions with infinite support, nestled alongside the Geometric distribution and the ubiquitous Poisson distribution, among others. It’s a reminder that for every phenomenon, there’s likely a distribution waiting to describe it, whether you want it to or not.