Sub-Gaussian distribution

From HandWiki
Short description: Continuous probability distribution

In probability theory, a sub-Gaussian distribution, the distribution of a sub-Gaussian random variable, is a probability distribution with strong tail decay. More specifically, the tails of a sub-Gaussian distribution are dominated by (i.e. decay at least as fast as) the tails of a Gaussian. This property gives sub-Gaussian distributions their name.

Formally, the probability distribution of a random variable X is called sub-Gaussian if there is a positive constant C such that for every t0,

P(|X|t)2exp(t2/C2).

Alternatively, a random variable is considered sub-Gaussian if its distribution function is upper bounded (up to a constant) by the distribution function of a Gaussian. Specifically, we say that X is sub-Gaussian if for all s0 we have that:

P(|X|s)cP(|Z|s),

where c0 is constant and Z is a mean zero Gaussian random variable.[1]

Definitions

The sub-Gaussian norm of X, denoted as Xgauss, is defined byXgauss=inf{c>0:E[exp(X2c2)]2},which is the Orlicz norm of X generated by the Orlicz function Φ(u)=eu21. By condition (2) below, sub-Gaussian random variables can be characterized as those random variables with finite sub-Gaussian norm.

Sub-Gaussian properties

Let X be a random variable. The following conditions are equivalent:

  1. P(|X|t)2exp(t2/K12) for all t0, where K1 is a positive constant;
  2. E[exp(X2/K22)]2, where K2 is a positive constant;
  3. E|X|p2K3pΓ(p2+1) for all p1, where K3 is a positive constant.

Proof. (1)(3) By the layer cake representation,E|X|p=0P(|X|pt)dt=0ptp1P(|X|t)dt20ptp1exp(t2K12)dtAfter a change of variables u=t2/K12, we find thatE|X|p2K1pp20up21eudu=2K1pp2Γ(p2)=2K1pΓ(p2+1).

(3)(2) Using the Taylor series for ex:ex=1+p=1xpp!, we obtain thatE[exp(λX2)]=1+p=1λpE[X2p]p!1+p=12λpK32pΓ(p+1)p!=1+2p=1λpK32p=2p=0λpK32p1=21λK321for λK32<1,which is less than or equal to 2 for λ13K32. Take K2312K3, thenE[exp(X2/K22)]2.


(2)(1) By Markov's inequality,P(|X|t)=P(exp(X2K22)exp(t2K22))E[exp(X2/K22)]exp(t2K22)2exp(t2K22).

More equivalent definitions

The following properties are equivalent:

  • The distribution of X is sub-Gaussian.
  • Laplace transform condition: for some B, b > 0, Eeλ(XE[X])Beλ2b holds for all λ.
  • Moment condition: for some K > 0, E|X|pKppp/2 for all p1.
  • Moment generating function condition: for some L>0, E[exp(λ2X2)]exp(L2λ2) for all λ such that |λ|1L. [2]
  • Union bound condition: for some c > 0,  E[max{|X1E[X]|,,|XnE[X]|}]clogn for all n > c, where X1,,Xn are i.i.d copies of X.

Examples

A standard normal random variable XN(0,1) is a sub-Gaussian random variable.

Let X be a random variable with symmetric Bernoulli distribution (or Rademacher distribution). That is, X takes values 1 and 1 with probabilities 1/2 each. Since X2=1, it follows that Xgauss=inf{c>0:E[exp(X2c2)]2}=inf{c>0:E[exp(1c2)]2}=1ln2,and hence X is a sub-Gaussian random variable.

Maximum of Sub-Gaussian Random Variables

Consider a finite collection of subgaussian random variables, X1, ..., Xn, with corresponding subgaussian parameters σ. The random variable Mn = max(X1, ..., Xn) represents the maximum of this collection. The expectation Mn can be bounded above by 2σ2logn. Note that no independence assumption is needed to form this bound.[1]

See also

Notes

  1. 1.0 1.1 Wainwright MJ. High-Dimensional Statistics: A Non-Asymptotic Viewpoint. Cambridge: Cambridge University Press; 2019. doi:10.1017/9781108627771, ISBN:9781108627771.
  2. Vershynin, R. (2018). High-dimensional probability: An introduction with applications in data science. Cambridge: Cambridge University Press. pp. 33–34. 

References