[Math] Expectation of maximum of n i.i.d random variables

expected valueprobabilityprobability distributions

I have $n$ i.i.d. random variables, $X_1,…, X_n$ which follow some arbitrary distribution. Based on experiments in Python with various distributions, it seems that $\mathbb{E}(\max(X_1,…,X_n))$ is a linear (or seemingly close to linear) function of $\mathbb{E}(X_i)$. It is indeed linear for some examples where it is possible to get a closed form solution for $\mathbb{E}(\max(X_1,…,X_n))$ or a good approximation.

Expected value of $\max\{X_1,\ldots,X_n\}$ where $X_i$ are iid uniform.

Expectation of the maximum of i.i.d. geometric random variables

I wonder if this is the case more generally? Is there some way to prove it?

Best Answer

A general technique to get a bound that is often pretty decent is to use the MGF if you have it: for all $t\geq 0$: \begin{align} \exp(t\mathbb{E}[\max_i X_i])&\leq \mathbb{E}[\exp(t\max_i X_i)]\\ &\leq\mathbb{E}[\sum_{i=1}^n \exp(t X_i)] \\ &= n\mathbb{E}[\exp(t X)], \end{align} so \begin{equation} \mathbb{E}[\max_i X_i]\leq \frac{\log(n\mathbb{E}[\exp(tX)])}{t}. \end{equation} You can then optimize in $t\geq 0$ to get a decent upper bound. For instance, doing this with Gaussians with variance $\sigma^2$ would show $\mathbb{E}[\max_i X_i]\leq \sigma\sqrt{2\log n}$, which turns out to be right up to a constant.

Related Question