Expected Value of a quite complex coin flipping game.

binomial distributionexpected valuegame theorymarkov chainsprobability

I have a rather complex game, whose expected value I need to find.
Given is a binary random value with probability for success p (lets say an unfair coin with probability of landing heads p and tails 1-p, p<<0,5)
The rules of the games are the following: you start by betting 1 euro, a multiplier = 1 and 15 tosses.
Each time you land heads (less likely option) you receive 15 more tosses (added to the remaining one you have) and your multiplier increases by 1. Landing tails doesn't give you anything, you've just lost one toss. Your multiplier can grow up to 5, and each time you land a head after achieving a multiplier of 5 adds new 15 tosses, but does not increase the multiplier further.
I am seeking for the expected multiplier of the game.

I started by building a Markov Chain table A[i,j] for the probability of spin j to have multiplier i:
A[1,1] = 1,
A[1,2] = 1-p,
A[2,2] = p,
A[1,3] = (1-p)*(1-p),
etc… We have 15 initial spins then each
A[i, 15+i] = 0.
But I am still stuck of deriving the Avg MP from the table.
Any help will be much appreciated!

Thanks!

EDIT:
I forgot to say, at each toss you win your bet x Multiplier, so actually, I am looking for the EV of the game, but I think it can be presented as finding the average MP of the game and the average number of tosses, so in the end the EV should be their product. I know how to find the Avg number of tosses, and my main issue is the AVG MP.

Best Answer

Foreword.

Probably this answer is not what you expect. It aims actually at a more complicated task of computing the probability that the game ends at the $(n+1)$-th batch of trials. Knowing the probability one can of course compute the expected value of any quantity which depends only on the number of played batches. Yet the computation of an expected value often does not require the knowledge of the individual probabilities. It may be the case also here. Nevertheless I have decided to post the answer since it reveals some beauty and can be possibly useful for other problems.

Preliminaries.

Let $\tilde n$ be a weak composition of a non-negative number $n$ into $n+1$ non-negative parts: $$ \tilde n=(n_0,n_1,\ldots,n_{n}),\tag1 $$ and define the exceedance of the composition as: $$ X(\tilde n)=\sum_{m=0}^{n}\mathbf{1}_{N_m>m}\tag2 $$ where $\mathbf1_A$ is the indicator function and $N_m=\sum_{i=0}^m n_i$ are the partial sums of $\tilde n$. Expressing simply the exceedance is the number of valid inequalities $N_m> m$ for $m$ ranging from $0$ to $n$. The exceedance can take on $n+1$ values from $0$ to $n$ (observe that $N_n=n$).

Solution.

Let $t$ be the length of a batch of trials (in your example $t=15$). Then the probability that the game ends at the $(n+1)$-th batch is:

$$W(n,t)\left(pq^{t-1}\right)^n q^t,\tag3 $$ where $q=1-p$ is the probability to toss a tail and $W(n,t)$ is the number of ways to distribute $n$ heads between $t(n+1)$ tosses in a way which ensures the condition that there is sufficient number of previously tossed heads to proceed to every next batch.

Let $n_i$ in (1) be the number of heads tossed in the $i$-th batch. Then the above mentioned condition is simply: $$ X(\tilde n)=n.\tag4 $$

It follows then that $$ W(n,t)=\frac1{n+1}\binom{t(n+1)}n.\tag5 $$

The formula can be obtained in the following way. There are in total $\binom{t(n+1)}n$ ways to distribute $n$ heads among $t(n+1)$ tosses. Consider now some particular distribution and cyclically shift it by $k$ batches ($0\le k \le n$) so that $n'_i=n_{(i+k)\pmod{n+1}}$. It appears that all $n+1$ cyclically shifted compositions have distinct exceedances so that only one of them has the required exceedance $n$ (an excellent proof can be found here). This explains the prefactor $\frac1{n+1}$.

Appendix.

The overall probability that the game ends after a finite number of batches is: $$ P_t(q)=q^t\sum_{n=0}^\infty\frac{1}{n+1}\binom{t(n+1)}n\left(pq^{t-1}\right)^n.\tag6 $$

The following figure demonstrates the behavior of the function $P_t(q)$ for $t=1,2,3,4$.

enter image description here

Observe that $P_t(q)=1$ only if $p\le\frac1t$ (the inequality is strict for $t=1$). This means that for larger $p$ there is a certain probability that the game never ends. This is however irrelevant for the expected value of a constant.

Hope this helps.

Related Question