How to find the probability of a 'mean shift' in a random process?












3














How would I go about calculating the probability that the observed mean of a collection of data generated by a random process, which is governed by a particular statistical model, would deviate from the theoretical mean by a given amount?



For example, the sum of the rolls of two six-sided dice in theory follows a statistical model that resembles a Gaussian probability distribution curve, between the numbers 2 and 12, with a mean of 7. If I was to roll and sum the results of two dice 100 times, how can I calculate the probability that the observed mean of data would be 6, rather than 7?



In other words, if I was to do that test 100 times and the observed mean was calculated to be 6, how can I calculate the probability that the deviation in the mean could be because of random chance, as opposed to the underlying (assumed) statistical model being incorrect (i.e. the dice being biased)?



I know some basic statistics, but I'm far from an expert. I would think that the observed mean of a particular statistical sample would have a probability distribution of its own, but I'm not sure how it could be calculated.










share|cite|improve this question






















  • In the most general case, where the collection elements are from potentially different distributions, and could be dependent, this is very hard to compute. If you know individual variances and covariances, you can compute the variance of the mean, and then give a bound using Chebyshev's inequality.
    – Todor Markov
    Dec 13 '18 at 8:54










  • Have also a look at Cramers theorem and large deviation theorems. ocw.mit.edu/courses/sloan-school-of-management/… and ocw.mit.edu/courses/sloan-school-of-management/…
    – Thomas
    Dec 13 '18 at 11:24












  • That's the basis of statistics !
    – G Cab
    Dec 17 '18 at 0:26
















3














How would I go about calculating the probability that the observed mean of a collection of data generated by a random process, which is governed by a particular statistical model, would deviate from the theoretical mean by a given amount?



For example, the sum of the rolls of two six-sided dice in theory follows a statistical model that resembles a Gaussian probability distribution curve, between the numbers 2 and 12, with a mean of 7. If I was to roll and sum the results of two dice 100 times, how can I calculate the probability that the observed mean of data would be 6, rather than 7?



In other words, if I was to do that test 100 times and the observed mean was calculated to be 6, how can I calculate the probability that the deviation in the mean could be because of random chance, as opposed to the underlying (assumed) statistical model being incorrect (i.e. the dice being biased)?



I know some basic statistics, but I'm far from an expert. I would think that the observed mean of a particular statistical sample would have a probability distribution of its own, but I'm not sure how it could be calculated.










share|cite|improve this question






















  • In the most general case, where the collection elements are from potentially different distributions, and could be dependent, this is very hard to compute. If you know individual variances and covariances, you can compute the variance of the mean, and then give a bound using Chebyshev's inequality.
    – Todor Markov
    Dec 13 '18 at 8:54










  • Have also a look at Cramers theorem and large deviation theorems. ocw.mit.edu/courses/sloan-school-of-management/… and ocw.mit.edu/courses/sloan-school-of-management/…
    – Thomas
    Dec 13 '18 at 11:24












  • That's the basis of statistics !
    – G Cab
    Dec 17 '18 at 0:26














3












3








3







How would I go about calculating the probability that the observed mean of a collection of data generated by a random process, which is governed by a particular statistical model, would deviate from the theoretical mean by a given amount?



For example, the sum of the rolls of two six-sided dice in theory follows a statistical model that resembles a Gaussian probability distribution curve, between the numbers 2 and 12, with a mean of 7. If I was to roll and sum the results of two dice 100 times, how can I calculate the probability that the observed mean of data would be 6, rather than 7?



In other words, if I was to do that test 100 times and the observed mean was calculated to be 6, how can I calculate the probability that the deviation in the mean could be because of random chance, as opposed to the underlying (assumed) statistical model being incorrect (i.e. the dice being biased)?



I know some basic statistics, but I'm far from an expert. I would think that the observed mean of a particular statistical sample would have a probability distribution of its own, but I'm not sure how it could be calculated.










share|cite|improve this question













How would I go about calculating the probability that the observed mean of a collection of data generated by a random process, which is governed by a particular statistical model, would deviate from the theoretical mean by a given amount?



For example, the sum of the rolls of two six-sided dice in theory follows a statistical model that resembles a Gaussian probability distribution curve, between the numbers 2 and 12, with a mean of 7. If I was to roll and sum the results of two dice 100 times, how can I calculate the probability that the observed mean of data would be 6, rather than 7?



In other words, if I was to do that test 100 times and the observed mean was calculated to be 6, how can I calculate the probability that the deviation in the mean could be because of random chance, as opposed to the underlying (assumed) statistical model being incorrect (i.e. the dice being biased)?



I know some basic statistics, but I'm far from an expert. I would think that the observed mean of a particular statistical sample would have a probability distribution of its own, but I'm not sure how it could be calculated.







statistics means






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Dec 10 '18 at 20:39









Time4Tea

27319




27319












  • In the most general case, where the collection elements are from potentially different distributions, and could be dependent, this is very hard to compute. If you know individual variances and covariances, you can compute the variance of the mean, and then give a bound using Chebyshev's inequality.
    – Todor Markov
    Dec 13 '18 at 8:54










  • Have also a look at Cramers theorem and large deviation theorems. ocw.mit.edu/courses/sloan-school-of-management/… and ocw.mit.edu/courses/sloan-school-of-management/…
    – Thomas
    Dec 13 '18 at 11:24












  • That's the basis of statistics !
    – G Cab
    Dec 17 '18 at 0:26


















  • In the most general case, where the collection elements are from potentially different distributions, and could be dependent, this is very hard to compute. If you know individual variances and covariances, you can compute the variance of the mean, and then give a bound using Chebyshev's inequality.
    – Todor Markov
    Dec 13 '18 at 8:54










  • Have also a look at Cramers theorem and large deviation theorems. ocw.mit.edu/courses/sloan-school-of-management/… and ocw.mit.edu/courses/sloan-school-of-management/…
    – Thomas
    Dec 13 '18 at 11:24












  • That's the basis of statistics !
    – G Cab
    Dec 17 '18 at 0:26
















In the most general case, where the collection elements are from potentially different distributions, and could be dependent, this is very hard to compute. If you know individual variances and covariances, you can compute the variance of the mean, and then give a bound using Chebyshev's inequality.
– Todor Markov
Dec 13 '18 at 8:54




In the most general case, where the collection elements are from potentially different distributions, and could be dependent, this is very hard to compute. If you know individual variances and covariances, you can compute the variance of the mean, and then give a bound using Chebyshev's inequality.
– Todor Markov
Dec 13 '18 at 8:54












Have also a look at Cramers theorem and large deviation theorems. ocw.mit.edu/courses/sloan-school-of-management/… and ocw.mit.edu/courses/sloan-school-of-management/…
– Thomas
Dec 13 '18 at 11:24






Have also a look at Cramers theorem and large deviation theorems. ocw.mit.edu/courses/sloan-school-of-management/… and ocw.mit.edu/courses/sloan-school-of-management/…
– Thomas
Dec 13 '18 at 11:24














That's the basis of statistics !
– G Cab
Dec 17 '18 at 0:26




That's the basis of statistics !
– G Cab
Dec 17 '18 at 0:26










2 Answers
2






active

oldest

votes


















0














Let the random variable $X_i$ denote the outcome of the $i$-th dice roll. Then the observed mean $bar X$ of $n$ dice rolls will simply be their sum $Sigma_X$ divided by the number of rolls $n$, i.e. $$bar X = frac{Sigma_X}n = frac{X_1 + X_2 + dots + X_n}n = frac1n sum_{i=1}^n X_i.$$



This implies that the distribution of $bar X$ is simply the distribution of $Sigma_X$ transformed by uniformly scaling the values of the result by $frac1n$. In particular, the probability that $bar X le 6$, for example, is equal to the probability that $Sigma_X le 6n$.



Now, in your case, each $X_i$ is the sum of the rolls of two six-sided dice, so $Sigma_X$ is simply the sum of $2n$ six-sided dice. This distribution does not (as far as I know) have a common name or any particularly nice algebraic form, although it is closely related to a multinomial distribution with $2n$ trials and six equally likely outcomes per trial. However, it's not particularly hard to calculate it numerically, e.g. by starting with the singular distribution with all probability mass concentrated at zero, and convolving it $2n$ times with the uniform distribution over ${1,2,3,4,5,6}$.



For example, this Python program calculates and prints out the distribution of the sum of 200 six-sided dice, both as the raw point probability of rolling each sum and as the cumulative probability of rolling each sum or less. From the line numbered 600, we can read that the probability of rolling a sum of 600 or less (and thus a mean of 6 or less over 100 pairs of rolls) is approximately 0.00001771.



(Alternatively, we could use a tool made specifically for this job, although we'd need to switch to the export view to get an accurate numerical value for a result so far away from the expected value.)





For large numbers of rolls (say, more than 10) we can also get pretty good results simply by approximating the distribution of $bar X$ (and/or $Sigma_X$) with a normal distribution with the same expectation and variance. This is a consequence of the central limit theorem, which says that the distribution of the sum (and the average) of a large number of identically distributed random variables tends towards a normal distribution.



By the linearity of the expected value, the expected value of $Sigma_X$ is simply the sum of the expected values of the dice rolls, and the expected value of $bar X$ is their average (i.e. the same as for each individual roll, if the rolls are identically distributed). Further, assuming that the dice rolls are independent (or at least uncorrelated), the variance of $Sigma_X$ will be the sum of the variances of the dice rolls, and the variance of $bar X$ will be $frac1{n^2}$ of the variance of $Sigma_X$ (i.e. respectively $n$ and $frac1n$ times the variance of each individual roll, if they're all the same).



Since the variance of a single $k$-sided die roll (i.e. a uniformly distributed random variable over ${1, 2, dots, k}$ is $(k^2-1)mathbin/12$, and the variance of the sum of two rolls is simply twice that, it follows that the variance of a roll of two six-sided dice is $2 times 35 mathbin/ 12 = 35 mathbin/ 6 approx 5.83$. Therefore, over 100 such double rolls, the distribution of the sum $Sigma_X$ is well approximated by a normal distribution with mean $mu = 700$ and variance $sigma^2 approx 583$, while for the average $bar X$, the corresponding parameters are $mu = 7$ and $sigma^2 approx 0.0583$.



One useful feature of this normal approximation is that, for any given deviation of the observed mean from the expected value, we can divide it by the standard deviation $sigma$ of the distribution (which is simply the square root of the variance $sigma^2$) and compare this with a precomputed table of confidence intervals, or a simple diagram like the image below, to see how likely the observed mean is to fall that far from the expectation.



Normal distribution with multiples of the standard deviation marked
Image from Wikimedia Commons by M. W. Toews, based (in concept) on figure by Jeremy Kemp; used under the CC-By 2.5 license.



For example, the average $bar X$ of 100 rolls of two six-sided dice has a standard deviation of $sigma = sqrt{35 mathbin/ 600} approx 0.2415$. Thus, an observed mean of 6 or less would be $1/sigma approx 4.14 ge 4$ standard deviations below the expected value of 7, and thus (based on the table on the Wikipedia page I linked to above) occurs with probability less than 0.00006334. A more accurate calculation (using the formula $frac12left(1+operatorname{erf}left(frac{x-mu}{sigmasqrt2}right)right)$ also given on the Wikipedia page) yields an approximate probability of 0.0000173 of the observed mean falling this far from the expectation, which is fairly close to the value calculated numerically above without using the normal approximation.






share|cite|improve this answer





























    0














    Suppose the game of backgammon is going on, and our task is to understand to what extent the cubes correspond to the ideal statistical model.



    Solutions to the problem may be different. Consider them in order of increasing complexity of obtaining the desired result.



    $$textbf{Case 1}$$
    $textbf{Can be collected separate statistics for each cube.}$



    For example, if the cubes differ in appearance.



    In this case, the binomial distribution law of
    $$P_n(k,p) = binom nm p^m(1-p)^{n-m}$$
    should be satisfied, where
    $p$ is the probability of the required cube face loss in the single test,
    $n$ is the number of throws,
    $m$ is the number of depositions of the required face.



    For each face of an ideal cube $p=p_r=dfrac16.$



    The fact of the loss of a given face $k$ times from $n$ shots denote $E_n(k).$



    Let us consider the hypotheses $H_j = left{left|p-dfrac jnright|<dfrac1{2n}right},$
    $bigg(p$ belongs to the neighbourhood of value $dfrac jnbigg),$

    then
    $$P(H_j) = F_n(H_j) = intlimits_{-1/(2n)}^{1/(2n)} P_nleft(m,dfrac jn+xright),mathrm dx,$$
    $$P(H_j) = P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)}left(dfrac{dfrac jn+x}{dfrac jn}right)^m left(dfrac{dfrac {n-j}n-x}{dfrac{n-j}n}right)^{n-m}dx$$
    $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)} left(1+dfrac{nx}jright)^m left(1-dfrac{nx}{n-j}right)^{n-m}d(nx)$$
    $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{y}jright)^m left(1-dfrac{y}{n-j}right)^{n-m}dy.$$
    If $quad j>>dfrac12,quad n-j >>dfrac12,quad$ then
    $$F_n(H_j) = {smalldfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac mj y+dfrac{m(m-1)}{2j^2}y^2+dotsright)left(1-dfrac {n-m}{n-j}y+dfrac {(n-m)(n-m-1)}{2(n-j)^2}y^2+dotsright)dy}$$
    $$approx dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{(m-j)n}{j(n-j)}y + dfrac{(n^2-n)(j-m)^2+m(n-m)}{2j^2(n-j)^2}y^2right)dy$$
    $$ = dfrac jnleft(1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}right) P_nleft(m,dfrac jnright),$$
    The probability of hypothesis for the ideal face of cube is
    $$P(H_r) = F_{6k}(k)$$
    The plot of the factor
    $$1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}$$
    for $m=10,quad n= 60,quad jin(7,15)$



    factor



    shows that prior probability of hypothesis $H_j$ is
    $$P(H_j)approx dfrac 1n P_nleft(m,dfrac jnright)$$
    with the accuracy $0.5%.$



    Let $n=6m,$ then
    begin{cases}
    P(E_k | H_r) approx P_{6m}(m,k),\[4pt]
    P(E_k | H_j) approx P_{6m}(j,k),\[4pt]
    end{cases}



    By the Bayesian formula for the posterior probability is
    $$P(H_j | E_k) = dfrac{P(E_k | H_j)P(H_j | H_r)}{sum_{j=0}^{6m}P(E_k | H_i)P(H_i |)},$$



    $$P(H_j | E_k) = dfrac{dbinom{6m}kleft(dfrac j{6m}right)^kleft(dfrac {6m-j}{6m}right)^{6m-k}{dfrac1{6m}dbinom{6m}j}left(dfrac 16right)^jleft(dfrac 56right)^{6m-j}}
    {sumlimits_{i=0}^{6m}dbinom{6m}kleft(dfrac i{6m}right)^kleft(dfrac {6m-i}{6m}right)^{6m-k}dfrac1{6m}dbinom{6m}ileft(dfrac 16right)^ileft(dfrac 56right)^{6m-i}},$$

    $$boxed{ P(H_j | E_k) = dfrac{dbinom{6m}j}
    {sumlimits_{i=0}^{6m}dbinom{6m}i 5^{j-i} left(dfrac ijright)^kleft(dfrac{6m-i}{6m-j}right)^{6m-k}}},$$



    The resulting formula indicates the posterior probability of the distribution of hypotheses about the probability of a given face of a cube falling out, provided that the given face fell $k$ times in $n$ throws



    If the given face fell $10$ times in $60$ throws ($m=10, n=60, k=10$), then Wolfram Alpha plot is



    m=10, n=60, k=10



    If the given face fell $7$ times in $60$ throws ($m=10, n=60, k=7$), then Wolfram Alpha plot is



    m=10, n=60, k=7



    If the given face fell $15$ times in $60$ throws ($m=10, n=60, k=15$), then Wolfram Alpha plot is



    m=10, n=60, k=15



    $$textbf{Case 1}$$
    $textbf{Can not be collected separate statistics for each cube.}$



    This case can not provide valid data, because summation hides individual propertied of cubes. Besides, cubes can be tuned to the combinations as "6+5". Using of this issue data cannot be recommended.






    share|cite|improve this answer





















      Your Answer





      StackExchange.ifUsing("editor", function () {
      return StackExchange.using("mathjaxEditing", function () {
      StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
      StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
      });
      });
      }, "mathjax-editing");

      StackExchange.ready(function() {
      var channelOptions = {
      tags: "".split(" "),
      id: "69"
      };
      initTagRenderer("".split(" "), "".split(" "), channelOptions);

      StackExchange.using("externalEditor", function() {
      // Have to fire editor after snippets, if snippets enabled
      if (StackExchange.settings.snippets.snippetsEnabled) {
      StackExchange.using("snippets", function() {
      createEditor();
      });
      }
      else {
      createEditor();
      }
      });

      function createEditor() {
      StackExchange.prepareEditor({
      heartbeatType: 'answer',
      autoActivateHeartbeat: false,
      convertImagesToLinks: true,
      noModals: true,
      showLowRepImageUploadWarning: true,
      reputationToPostImages: 10,
      bindNavPrevention: true,
      postfix: "",
      imageUploader: {
      brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
      contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
      allowUrls: true
      },
      noCode: true, onDemand: true,
      discardSelector: ".discard-answer"
      ,immediatelyShowMarkdownHelp:true
      });


      }
      });














      draft saved

      draft discarded


















      StackExchange.ready(
      function () {
      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3034432%2fhow-to-find-the-probability-of-a-mean-shift-in-a-random-process%23new-answer', 'question_page');
      }
      );

      Post as a guest















      Required, but never shown

























      2 Answers
      2






      active

      oldest

      votes








      2 Answers
      2






      active

      oldest

      votes









      active

      oldest

      votes






      active

      oldest

      votes









      0














      Let the random variable $X_i$ denote the outcome of the $i$-th dice roll. Then the observed mean $bar X$ of $n$ dice rolls will simply be their sum $Sigma_X$ divided by the number of rolls $n$, i.e. $$bar X = frac{Sigma_X}n = frac{X_1 + X_2 + dots + X_n}n = frac1n sum_{i=1}^n X_i.$$



      This implies that the distribution of $bar X$ is simply the distribution of $Sigma_X$ transformed by uniformly scaling the values of the result by $frac1n$. In particular, the probability that $bar X le 6$, for example, is equal to the probability that $Sigma_X le 6n$.



      Now, in your case, each $X_i$ is the sum of the rolls of two six-sided dice, so $Sigma_X$ is simply the sum of $2n$ six-sided dice. This distribution does not (as far as I know) have a common name or any particularly nice algebraic form, although it is closely related to a multinomial distribution with $2n$ trials and six equally likely outcomes per trial. However, it's not particularly hard to calculate it numerically, e.g. by starting with the singular distribution with all probability mass concentrated at zero, and convolving it $2n$ times with the uniform distribution over ${1,2,3,4,5,6}$.



      For example, this Python program calculates and prints out the distribution of the sum of 200 six-sided dice, both as the raw point probability of rolling each sum and as the cumulative probability of rolling each sum or less. From the line numbered 600, we can read that the probability of rolling a sum of 600 or less (and thus a mean of 6 or less over 100 pairs of rolls) is approximately 0.00001771.



      (Alternatively, we could use a tool made specifically for this job, although we'd need to switch to the export view to get an accurate numerical value for a result so far away from the expected value.)





      For large numbers of rolls (say, more than 10) we can also get pretty good results simply by approximating the distribution of $bar X$ (and/or $Sigma_X$) with a normal distribution with the same expectation and variance. This is a consequence of the central limit theorem, which says that the distribution of the sum (and the average) of a large number of identically distributed random variables tends towards a normal distribution.



      By the linearity of the expected value, the expected value of $Sigma_X$ is simply the sum of the expected values of the dice rolls, and the expected value of $bar X$ is their average (i.e. the same as for each individual roll, if the rolls are identically distributed). Further, assuming that the dice rolls are independent (or at least uncorrelated), the variance of $Sigma_X$ will be the sum of the variances of the dice rolls, and the variance of $bar X$ will be $frac1{n^2}$ of the variance of $Sigma_X$ (i.e. respectively $n$ and $frac1n$ times the variance of each individual roll, if they're all the same).



      Since the variance of a single $k$-sided die roll (i.e. a uniformly distributed random variable over ${1, 2, dots, k}$ is $(k^2-1)mathbin/12$, and the variance of the sum of two rolls is simply twice that, it follows that the variance of a roll of two six-sided dice is $2 times 35 mathbin/ 12 = 35 mathbin/ 6 approx 5.83$. Therefore, over 100 such double rolls, the distribution of the sum $Sigma_X$ is well approximated by a normal distribution with mean $mu = 700$ and variance $sigma^2 approx 583$, while for the average $bar X$, the corresponding parameters are $mu = 7$ and $sigma^2 approx 0.0583$.



      One useful feature of this normal approximation is that, for any given deviation of the observed mean from the expected value, we can divide it by the standard deviation $sigma$ of the distribution (which is simply the square root of the variance $sigma^2$) and compare this with a precomputed table of confidence intervals, or a simple diagram like the image below, to see how likely the observed mean is to fall that far from the expectation.



      Normal distribution with multiples of the standard deviation marked
      Image from Wikimedia Commons by M. W. Toews, based (in concept) on figure by Jeremy Kemp; used under the CC-By 2.5 license.



      For example, the average $bar X$ of 100 rolls of two six-sided dice has a standard deviation of $sigma = sqrt{35 mathbin/ 600} approx 0.2415$. Thus, an observed mean of 6 or less would be $1/sigma approx 4.14 ge 4$ standard deviations below the expected value of 7, and thus (based on the table on the Wikipedia page I linked to above) occurs with probability less than 0.00006334. A more accurate calculation (using the formula $frac12left(1+operatorname{erf}left(frac{x-mu}{sigmasqrt2}right)right)$ also given on the Wikipedia page) yields an approximate probability of 0.0000173 of the observed mean falling this far from the expectation, which is fairly close to the value calculated numerically above without using the normal approximation.






      share|cite|improve this answer


























        0














        Let the random variable $X_i$ denote the outcome of the $i$-th dice roll. Then the observed mean $bar X$ of $n$ dice rolls will simply be their sum $Sigma_X$ divided by the number of rolls $n$, i.e. $$bar X = frac{Sigma_X}n = frac{X_1 + X_2 + dots + X_n}n = frac1n sum_{i=1}^n X_i.$$



        This implies that the distribution of $bar X$ is simply the distribution of $Sigma_X$ transformed by uniformly scaling the values of the result by $frac1n$. In particular, the probability that $bar X le 6$, for example, is equal to the probability that $Sigma_X le 6n$.



        Now, in your case, each $X_i$ is the sum of the rolls of two six-sided dice, so $Sigma_X$ is simply the sum of $2n$ six-sided dice. This distribution does not (as far as I know) have a common name or any particularly nice algebraic form, although it is closely related to a multinomial distribution with $2n$ trials and six equally likely outcomes per trial. However, it's not particularly hard to calculate it numerically, e.g. by starting with the singular distribution with all probability mass concentrated at zero, and convolving it $2n$ times with the uniform distribution over ${1,2,3,4,5,6}$.



        For example, this Python program calculates and prints out the distribution of the sum of 200 six-sided dice, both as the raw point probability of rolling each sum and as the cumulative probability of rolling each sum or less. From the line numbered 600, we can read that the probability of rolling a sum of 600 or less (and thus a mean of 6 or less over 100 pairs of rolls) is approximately 0.00001771.



        (Alternatively, we could use a tool made specifically for this job, although we'd need to switch to the export view to get an accurate numerical value for a result so far away from the expected value.)





        For large numbers of rolls (say, more than 10) we can also get pretty good results simply by approximating the distribution of $bar X$ (and/or $Sigma_X$) with a normal distribution with the same expectation and variance. This is a consequence of the central limit theorem, which says that the distribution of the sum (and the average) of a large number of identically distributed random variables tends towards a normal distribution.



        By the linearity of the expected value, the expected value of $Sigma_X$ is simply the sum of the expected values of the dice rolls, and the expected value of $bar X$ is their average (i.e. the same as for each individual roll, if the rolls are identically distributed). Further, assuming that the dice rolls are independent (or at least uncorrelated), the variance of $Sigma_X$ will be the sum of the variances of the dice rolls, and the variance of $bar X$ will be $frac1{n^2}$ of the variance of $Sigma_X$ (i.e. respectively $n$ and $frac1n$ times the variance of each individual roll, if they're all the same).



        Since the variance of a single $k$-sided die roll (i.e. a uniformly distributed random variable over ${1, 2, dots, k}$ is $(k^2-1)mathbin/12$, and the variance of the sum of two rolls is simply twice that, it follows that the variance of a roll of two six-sided dice is $2 times 35 mathbin/ 12 = 35 mathbin/ 6 approx 5.83$. Therefore, over 100 such double rolls, the distribution of the sum $Sigma_X$ is well approximated by a normal distribution with mean $mu = 700$ and variance $sigma^2 approx 583$, while for the average $bar X$, the corresponding parameters are $mu = 7$ and $sigma^2 approx 0.0583$.



        One useful feature of this normal approximation is that, for any given deviation of the observed mean from the expected value, we can divide it by the standard deviation $sigma$ of the distribution (which is simply the square root of the variance $sigma^2$) and compare this with a precomputed table of confidence intervals, or a simple diagram like the image below, to see how likely the observed mean is to fall that far from the expectation.



        Normal distribution with multiples of the standard deviation marked
        Image from Wikimedia Commons by M. W. Toews, based (in concept) on figure by Jeremy Kemp; used under the CC-By 2.5 license.



        For example, the average $bar X$ of 100 rolls of two six-sided dice has a standard deviation of $sigma = sqrt{35 mathbin/ 600} approx 0.2415$. Thus, an observed mean of 6 or less would be $1/sigma approx 4.14 ge 4$ standard deviations below the expected value of 7, and thus (based on the table on the Wikipedia page I linked to above) occurs with probability less than 0.00006334. A more accurate calculation (using the formula $frac12left(1+operatorname{erf}left(frac{x-mu}{sigmasqrt2}right)right)$ also given on the Wikipedia page) yields an approximate probability of 0.0000173 of the observed mean falling this far from the expectation, which is fairly close to the value calculated numerically above without using the normal approximation.






        share|cite|improve this answer
























          0












          0








          0






          Let the random variable $X_i$ denote the outcome of the $i$-th dice roll. Then the observed mean $bar X$ of $n$ dice rolls will simply be their sum $Sigma_X$ divided by the number of rolls $n$, i.e. $$bar X = frac{Sigma_X}n = frac{X_1 + X_2 + dots + X_n}n = frac1n sum_{i=1}^n X_i.$$



          This implies that the distribution of $bar X$ is simply the distribution of $Sigma_X$ transformed by uniformly scaling the values of the result by $frac1n$. In particular, the probability that $bar X le 6$, for example, is equal to the probability that $Sigma_X le 6n$.



          Now, in your case, each $X_i$ is the sum of the rolls of two six-sided dice, so $Sigma_X$ is simply the sum of $2n$ six-sided dice. This distribution does not (as far as I know) have a common name or any particularly nice algebraic form, although it is closely related to a multinomial distribution with $2n$ trials and six equally likely outcomes per trial. However, it's not particularly hard to calculate it numerically, e.g. by starting with the singular distribution with all probability mass concentrated at zero, and convolving it $2n$ times with the uniform distribution over ${1,2,3,4,5,6}$.



          For example, this Python program calculates and prints out the distribution of the sum of 200 six-sided dice, both as the raw point probability of rolling each sum and as the cumulative probability of rolling each sum or less. From the line numbered 600, we can read that the probability of rolling a sum of 600 or less (and thus a mean of 6 or less over 100 pairs of rolls) is approximately 0.00001771.



          (Alternatively, we could use a tool made specifically for this job, although we'd need to switch to the export view to get an accurate numerical value for a result so far away from the expected value.)





          For large numbers of rolls (say, more than 10) we can also get pretty good results simply by approximating the distribution of $bar X$ (and/or $Sigma_X$) with a normal distribution with the same expectation and variance. This is a consequence of the central limit theorem, which says that the distribution of the sum (and the average) of a large number of identically distributed random variables tends towards a normal distribution.



          By the linearity of the expected value, the expected value of $Sigma_X$ is simply the sum of the expected values of the dice rolls, and the expected value of $bar X$ is their average (i.e. the same as for each individual roll, if the rolls are identically distributed). Further, assuming that the dice rolls are independent (or at least uncorrelated), the variance of $Sigma_X$ will be the sum of the variances of the dice rolls, and the variance of $bar X$ will be $frac1{n^2}$ of the variance of $Sigma_X$ (i.e. respectively $n$ and $frac1n$ times the variance of each individual roll, if they're all the same).



          Since the variance of a single $k$-sided die roll (i.e. a uniformly distributed random variable over ${1, 2, dots, k}$ is $(k^2-1)mathbin/12$, and the variance of the sum of two rolls is simply twice that, it follows that the variance of a roll of two six-sided dice is $2 times 35 mathbin/ 12 = 35 mathbin/ 6 approx 5.83$. Therefore, over 100 such double rolls, the distribution of the sum $Sigma_X$ is well approximated by a normal distribution with mean $mu = 700$ and variance $sigma^2 approx 583$, while for the average $bar X$, the corresponding parameters are $mu = 7$ and $sigma^2 approx 0.0583$.



          One useful feature of this normal approximation is that, for any given deviation of the observed mean from the expected value, we can divide it by the standard deviation $sigma$ of the distribution (which is simply the square root of the variance $sigma^2$) and compare this with a precomputed table of confidence intervals, or a simple diagram like the image below, to see how likely the observed mean is to fall that far from the expectation.



          Normal distribution with multiples of the standard deviation marked
          Image from Wikimedia Commons by M. W. Toews, based (in concept) on figure by Jeremy Kemp; used under the CC-By 2.5 license.



          For example, the average $bar X$ of 100 rolls of two six-sided dice has a standard deviation of $sigma = sqrt{35 mathbin/ 600} approx 0.2415$. Thus, an observed mean of 6 or less would be $1/sigma approx 4.14 ge 4$ standard deviations below the expected value of 7, and thus (based on the table on the Wikipedia page I linked to above) occurs with probability less than 0.00006334. A more accurate calculation (using the formula $frac12left(1+operatorname{erf}left(frac{x-mu}{sigmasqrt2}right)right)$ also given on the Wikipedia page) yields an approximate probability of 0.0000173 of the observed mean falling this far from the expectation, which is fairly close to the value calculated numerically above without using the normal approximation.






          share|cite|improve this answer












          Let the random variable $X_i$ denote the outcome of the $i$-th dice roll. Then the observed mean $bar X$ of $n$ dice rolls will simply be their sum $Sigma_X$ divided by the number of rolls $n$, i.e. $$bar X = frac{Sigma_X}n = frac{X_1 + X_2 + dots + X_n}n = frac1n sum_{i=1}^n X_i.$$



          This implies that the distribution of $bar X$ is simply the distribution of $Sigma_X$ transformed by uniformly scaling the values of the result by $frac1n$. In particular, the probability that $bar X le 6$, for example, is equal to the probability that $Sigma_X le 6n$.



          Now, in your case, each $X_i$ is the sum of the rolls of two six-sided dice, so $Sigma_X$ is simply the sum of $2n$ six-sided dice. This distribution does not (as far as I know) have a common name or any particularly nice algebraic form, although it is closely related to a multinomial distribution with $2n$ trials and six equally likely outcomes per trial. However, it's not particularly hard to calculate it numerically, e.g. by starting with the singular distribution with all probability mass concentrated at zero, and convolving it $2n$ times with the uniform distribution over ${1,2,3,4,5,6}$.



          For example, this Python program calculates and prints out the distribution of the sum of 200 six-sided dice, both as the raw point probability of rolling each sum and as the cumulative probability of rolling each sum or less. From the line numbered 600, we can read that the probability of rolling a sum of 600 or less (and thus a mean of 6 or less over 100 pairs of rolls) is approximately 0.00001771.



          (Alternatively, we could use a tool made specifically for this job, although we'd need to switch to the export view to get an accurate numerical value for a result so far away from the expected value.)





          For large numbers of rolls (say, more than 10) we can also get pretty good results simply by approximating the distribution of $bar X$ (and/or $Sigma_X$) with a normal distribution with the same expectation and variance. This is a consequence of the central limit theorem, which says that the distribution of the sum (and the average) of a large number of identically distributed random variables tends towards a normal distribution.



          By the linearity of the expected value, the expected value of $Sigma_X$ is simply the sum of the expected values of the dice rolls, and the expected value of $bar X$ is their average (i.e. the same as for each individual roll, if the rolls are identically distributed). Further, assuming that the dice rolls are independent (or at least uncorrelated), the variance of $Sigma_X$ will be the sum of the variances of the dice rolls, and the variance of $bar X$ will be $frac1{n^2}$ of the variance of $Sigma_X$ (i.e. respectively $n$ and $frac1n$ times the variance of each individual roll, if they're all the same).



          Since the variance of a single $k$-sided die roll (i.e. a uniformly distributed random variable over ${1, 2, dots, k}$ is $(k^2-1)mathbin/12$, and the variance of the sum of two rolls is simply twice that, it follows that the variance of a roll of two six-sided dice is $2 times 35 mathbin/ 12 = 35 mathbin/ 6 approx 5.83$. Therefore, over 100 such double rolls, the distribution of the sum $Sigma_X$ is well approximated by a normal distribution with mean $mu = 700$ and variance $sigma^2 approx 583$, while for the average $bar X$, the corresponding parameters are $mu = 7$ and $sigma^2 approx 0.0583$.



          One useful feature of this normal approximation is that, for any given deviation of the observed mean from the expected value, we can divide it by the standard deviation $sigma$ of the distribution (which is simply the square root of the variance $sigma^2$) and compare this with a precomputed table of confidence intervals, or a simple diagram like the image below, to see how likely the observed mean is to fall that far from the expectation.



          Normal distribution with multiples of the standard deviation marked
          Image from Wikimedia Commons by M. W. Toews, based (in concept) on figure by Jeremy Kemp; used under the CC-By 2.5 license.



          For example, the average $bar X$ of 100 rolls of two six-sided dice has a standard deviation of $sigma = sqrt{35 mathbin/ 600} approx 0.2415$. Thus, an observed mean of 6 or less would be $1/sigma approx 4.14 ge 4$ standard deviations below the expected value of 7, and thus (based on the table on the Wikipedia page I linked to above) occurs with probability less than 0.00006334. A more accurate calculation (using the formula $frac12left(1+operatorname{erf}left(frac{x-mu}{sigmasqrt2}right)right)$ also given on the Wikipedia page) yields an approximate probability of 0.0000173 of the observed mean falling this far from the expectation, which is fairly close to the value calculated numerically above without using the normal approximation.







          share|cite|improve this answer












          share|cite|improve this answer



          share|cite|improve this answer










          answered Dec 17 '18 at 0:09









          Ilmari Karonen

          19.4k25182




          19.4k25182























              0














              Suppose the game of backgammon is going on, and our task is to understand to what extent the cubes correspond to the ideal statistical model.



              Solutions to the problem may be different. Consider them in order of increasing complexity of obtaining the desired result.



              $$textbf{Case 1}$$
              $textbf{Can be collected separate statistics for each cube.}$



              For example, if the cubes differ in appearance.



              In this case, the binomial distribution law of
              $$P_n(k,p) = binom nm p^m(1-p)^{n-m}$$
              should be satisfied, where
              $p$ is the probability of the required cube face loss in the single test,
              $n$ is the number of throws,
              $m$ is the number of depositions of the required face.



              For each face of an ideal cube $p=p_r=dfrac16.$



              The fact of the loss of a given face $k$ times from $n$ shots denote $E_n(k).$



              Let us consider the hypotheses $H_j = left{left|p-dfrac jnright|<dfrac1{2n}right},$
              $bigg(p$ belongs to the neighbourhood of value $dfrac jnbigg),$

              then
              $$P(H_j) = F_n(H_j) = intlimits_{-1/(2n)}^{1/(2n)} P_nleft(m,dfrac jn+xright),mathrm dx,$$
              $$P(H_j) = P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)}left(dfrac{dfrac jn+x}{dfrac jn}right)^m left(dfrac{dfrac {n-j}n-x}{dfrac{n-j}n}right)^{n-m}dx$$
              $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)} left(1+dfrac{nx}jright)^m left(1-dfrac{nx}{n-j}right)^{n-m}d(nx)$$
              $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{y}jright)^m left(1-dfrac{y}{n-j}right)^{n-m}dy.$$
              If $quad j>>dfrac12,quad n-j >>dfrac12,quad$ then
              $$F_n(H_j) = {smalldfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac mj y+dfrac{m(m-1)}{2j^2}y^2+dotsright)left(1-dfrac {n-m}{n-j}y+dfrac {(n-m)(n-m-1)}{2(n-j)^2}y^2+dotsright)dy}$$
              $$approx dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{(m-j)n}{j(n-j)}y + dfrac{(n^2-n)(j-m)^2+m(n-m)}{2j^2(n-j)^2}y^2right)dy$$
              $$ = dfrac jnleft(1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}right) P_nleft(m,dfrac jnright),$$
              The probability of hypothesis for the ideal face of cube is
              $$P(H_r) = F_{6k}(k)$$
              The plot of the factor
              $$1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}$$
              for $m=10,quad n= 60,quad jin(7,15)$



              factor



              shows that prior probability of hypothesis $H_j$ is
              $$P(H_j)approx dfrac 1n P_nleft(m,dfrac jnright)$$
              with the accuracy $0.5%.$



              Let $n=6m,$ then
              begin{cases}
              P(E_k | H_r) approx P_{6m}(m,k),\[4pt]
              P(E_k | H_j) approx P_{6m}(j,k),\[4pt]
              end{cases}



              By the Bayesian formula for the posterior probability is
              $$P(H_j | E_k) = dfrac{P(E_k | H_j)P(H_j | H_r)}{sum_{j=0}^{6m}P(E_k | H_i)P(H_i |)},$$



              $$P(H_j | E_k) = dfrac{dbinom{6m}kleft(dfrac j{6m}right)^kleft(dfrac {6m-j}{6m}right)^{6m-k}{dfrac1{6m}dbinom{6m}j}left(dfrac 16right)^jleft(dfrac 56right)^{6m-j}}
              {sumlimits_{i=0}^{6m}dbinom{6m}kleft(dfrac i{6m}right)^kleft(dfrac {6m-i}{6m}right)^{6m-k}dfrac1{6m}dbinom{6m}ileft(dfrac 16right)^ileft(dfrac 56right)^{6m-i}},$$

              $$boxed{ P(H_j | E_k) = dfrac{dbinom{6m}j}
              {sumlimits_{i=0}^{6m}dbinom{6m}i 5^{j-i} left(dfrac ijright)^kleft(dfrac{6m-i}{6m-j}right)^{6m-k}}},$$



              The resulting formula indicates the posterior probability of the distribution of hypotheses about the probability of a given face of a cube falling out, provided that the given face fell $k$ times in $n$ throws



              If the given face fell $10$ times in $60$ throws ($m=10, n=60, k=10$), then Wolfram Alpha plot is



              m=10, n=60, k=10



              If the given face fell $7$ times in $60$ throws ($m=10, n=60, k=7$), then Wolfram Alpha plot is



              m=10, n=60, k=7



              If the given face fell $15$ times in $60$ throws ($m=10, n=60, k=15$), then Wolfram Alpha plot is



              m=10, n=60, k=15



              $$textbf{Case 1}$$
              $textbf{Can not be collected separate statistics for each cube.}$



              This case can not provide valid data, because summation hides individual propertied of cubes. Besides, cubes can be tuned to the combinations as "6+5". Using of this issue data cannot be recommended.






              share|cite|improve this answer


























                0














                Suppose the game of backgammon is going on, and our task is to understand to what extent the cubes correspond to the ideal statistical model.



                Solutions to the problem may be different. Consider them in order of increasing complexity of obtaining the desired result.



                $$textbf{Case 1}$$
                $textbf{Can be collected separate statistics for each cube.}$



                For example, if the cubes differ in appearance.



                In this case, the binomial distribution law of
                $$P_n(k,p) = binom nm p^m(1-p)^{n-m}$$
                should be satisfied, where
                $p$ is the probability of the required cube face loss in the single test,
                $n$ is the number of throws,
                $m$ is the number of depositions of the required face.



                For each face of an ideal cube $p=p_r=dfrac16.$



                The fact of the loss of a given face $k$ times from $n$ shots denote $E_n(k).$



                Let us consider the hypotheses $H_j = left{left|p-dfrac jnright|<dfrac1{2n}right},$
                $bigg(p$ belongs to the neighbourhood of value $dfrac jnbigg),$

                then
                $$P(H_j) = F_n(H_j) = intlimits_{-1/(2n)}^{1/(2n)} P_nleft(m,dfrac jn+xright),mathrm dx,$$
                $$P(H_j) = P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)}left(dfrac{dfrac jn+x}{dfrac jn}right)^m left(dfrac{dfrac {n-j}n-x}{dfrac{n-j}n}right)^{n-m}dx$$
                $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)} left(1+dfrac{nx}jright)^m left(1-dfrac{nx}{n-j}right)^{n-m}d(nx)$$
                $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{y}jright)^m left(1-dfrac{y}{n-j}right)^{n-m}dy.$$
                If $quad j>>dfrac12,quad n-j >>dfrac12,quad$ then
                $$F_n(H_j) = {smalldfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac mj y+dfrac{m(m-1)}{2j^2}y^2+dotsright)left(1-dfrac {n-m}{n-j}y+dfrac {(n-m)(n-m-1)}{2(n-j)^2}y^2+dotsright)dy}$$
                $$approx dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{(m-j)n}{j(n-j)}y + dfrac{(n^2-n)(j-m)^2+m(n-m)}{2j^2(n-j)^2}y^2right)dy$$
                $$ = dfrac jnleft(1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}right) P_nleft(m,dfrac jnright),$$
                The probability of hypothesis for the ideal face of cube is
                $$P(H_r) = F_{6k}(k)$$
                The plot of the factor
                $$1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}$$
                for $m=10,quad n= 60,quad jin(7,15)$



                factor



                shows that prior probability of hypothesis $H_j$ is
                $$P(H_j)approx dfrac 1n P_nleft(m,dfrac jnright)$$
                with the accuracy $0.5%.$



                Let $n=6m,$ then
                begin{cases}
                P(E_k | H_r) approx P_{6m}(m,k),\[4pt]
                P(E_k | H_j) approx P_{6m}(j,k),\[4pt]
                end{cases}



                By the Bayesian formula for the posterior probability is
                $$P(H_j | E_k) = dfrac{P(E_k | H_j)P(H_j | H_r)}{sum_{j=0}^{6m}P(E_k | H_i)P(H_i |)},$$



                $$P(H_j | E_k) = dfrac{dbinom{6m}kleft(dfrac j{6m}right)^kleft(dfrac {6m-j}{6m}right)^{6m-k}{dfrac1{6m}dbinom{6m}j}left(dfrac 16right)^jleft(dfrac 56right)^{6m-j}}
                {sumlimits_{i=0}^{6m}dbinom{6m}kleft(dfrac i{6m}right)^kleft(dfrac {6m-i}{6m}right)^{6m-k}dfrac1{6m}dbinom{6m}ileft(dfrac 16right)^ileft(dfrac 56right)^{6m-i}},$$

                $$boxed{ P(H_j | E_k) = dfrac{dbinom{6m}j}
                {sumlimits_{i=0}^{6m}dbinom{6m}i 5^{j-i} left(dfrac ijright)^kleft(dfrac{6m-i}{6m-j}right)^{6m-k}}},$$



                The resulting formula indicates the posterior probability of the distribution of hypotheses about the probability of a given face of a cube falling out, provided that the given face fell $k$ times in $n$ throws



                If the given face fell $10$ times in $60$ throws ($m=10, n=60, k=10$), then Wolfram Alpha plot is



                m=10, n=60, k=10



                If the given face fell $7$ times in $60$ throws ($m=10, n=60, k=7$), then Wolfram Alpha plot is



                m=10, n=60, k=7



                If the given face fell $15$ times in $60$ throws ($m=10, n=60, k=15$), then Wolfram Alpha plot is



                m=10, n=60, k=15



                $$textbf{Case 1}$$
                $textbf{Can not be collected separate statistics for each cube.}$



                This case can not provide valid data, because summation hides individual propertied of cubes. Besides, cubes can be tuned to the combinations as "6+5". Using of this issue data cannot be recommended.






                share|cite|improve this answer
























                  0












                  0








                  0






                  Suppose the game of backgammon is going on, and our task is to understand to what extent the cubes correspond to the ideal statistical model.



                  Solutions to the problem may be different. Consider them in order of increasing complexity of obtaining the desired result.



                  $$textbf{Case 1}$$
                  $textbf{Can be collected separate statistics for each cube.}$



                  For example, if the cubes differ in appearance.



                  In this case, the binomial distribution law of
                  $$P_n(k,p) = binom nm p^m(1-p)^{n-m}$$
                  should be satisfied, where
                  $p$ is the probability of the required cube face loss in the single test,
                  $n$ is the number of throws,
                  $m$ is the number of depositions of the required face.



                  For each face of an ideal cube $p=p_r=dfrac16.$



                  The fact of the loss of a given face $k$ times from $n$ shots denote $E_n(k).$



                  Let us consider the hypotheses $H_j = left{left|p-dfrac jnright|<dfrac1{2n}right},$
                  $bigg(p$ belongs to the neighbourhood of value $dfrac jnbigg),$

                  then
                  $$P(H_j) = F_n(H_j) = intlimits_{-1/(2n)}^{1/(2n)} P_nleft(m,dfrac jn+xright),mathrm dx,$$
                  $$P(H_j) = P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)}left(dfrac{dfrac jn+x}{dfrac jn}right)^m left(dfrac{dfrac {n-j}n-x}{dfrac{n-j}n}right)^{n-m}dx$$
                  $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)} left(1+dfrac{nx}jright)^m left(1-dfrac{nx}{n-j}right)^{n-m}d(nx)$$
                  $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{y}jright)^m left(1-dfrac{y}{n-j}right)^{n-m}dy.$$
                  If $quad j>>dfrac12,quad n-j >>dfrac12,quad$ then
                  $$F_n(H_j) = {smalldfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac mj y+dfrac{m(m-1)}{2j^2}y^2+dotsright)left(1-dfrac {n-m}{n-j}y+dfrac {(n-m)(n-m-1)}{2(n-j)^2}y^2+dotsright)dy}$$
                  $$approx dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{(m-j)n}{j(n-j)}y + dfrac{(n^2-n)(j-m)^2+m(n-m)}{2j^2(n-j)^2}y^2right)dy$$
                  $$ = dfrac jnleft(1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}right) P_nleft(m,dfrac jnright),$$
                  The probability of hypothesis for the ideal face of cube is
                  $$P(H_r) = F_{6k}(k)$$
                  The plot of the factor
                  $$1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}$$
                  for $m=10,quad n= 60,quad jin(7,15)$



                  factor



                  shows that prior probability of hypothesis $H_j$ is
                  $$P(H_j)approx dfrac 1n P_nleft(m,dfrac jnright)$$
                  with the accuracy $0.5%.$



                  Let $n=6m,$ then
                  begin{cases}
                  P(E_k | H_r) approx P_{6m}(m,k),\[4pt]
                  P(E_k | H_j) approx P_{6m}(j,k),\[4pt]
                  end{cases}



                  By the Bayesian formula for the posterior probability is
                  $$P(H_j | E_k) = dfrac{P(E_k | H_j)P(H_j | H_r)}{sum_{j=0}^{6m}P(E_k | H_i)P(H_i |)},$$



                  $$P(H_j | E_k) = dfrac{dbinom{6m}kleft(dfrac j{6m}right)^kleft(dfrac {6m-j}{6m}right)^{6m-k}{dfrac1{6m}dbinom{6m}j}left(dfrac 16right)^jleft(dfrac 56right)^{6m-j}}
                  {sumlimits_{i=0}^{6m}dbinom{6m}kleft(dfrac i{6m}right)^kleft(dfrac {6m-i}{6m}right)^{6m-k}dfrac1{6m}dbinom{6m}ileft(dfrac 16right)^ileft(dfrac 56right)^{6m-i}},$$

                  $$boxed{ P(H_j | E_k) = dfrac{dbinom{6m}j}
                  {sumlimits_{i=0}^{6m}dbinom{6m}i 5^{j-i} left(dfrac ijright)^kleft(dfrac{6m-i}{6m-j}right)^{6m-k}}},$$



                  The resulting formula indicates the posterior probability of the distribution of hypotheses about the probability of a given face of a cube falling out, provided that the given face fell $k$ times in $n$ throws



                  If the given face fell $10$ times in $60$ throws ($m=10, n=60, k=10$), then Wolfram Alpha plot is



                  m=10, n=60, k=10



                  If the given face fell $7$ times in $60$ throws ($m=10, n=60, k=7$), then Wolfram Alpha plot is



                  m=10, n=60, k=7



                  If the given face fell $15$ times in $60$ throws ($m=10, n=60, k=15$), then Wolfram Alpha plot is



                  m=10, n=60, k=15



                  $$textbf{Case 1}$$
                  $textbf{Can not be collected separate statistics for each cube.}$



                  This case can not provide valid data, because summation hides individual propertied of cubes. Besides, cubes can be tuned to the combinations as "6+5". Using of this issue data cannot be recommended.






                  share|cite|improve this answer












                  Suppose the game of backgammon is going on, and our task is to understand to what extent the cubes correspond to the ideal statistical model.



                  Solutions to the problem may be different. Consider them in order of increasing complexity of obtaining the desired result.



                  $$textbf{Case 1}$$
                  $textbf{Can be collected separate statistics for each cube.}$



                  For example, if the cubes differ in appearance.



                  In this case, the binomial distribution law of
                  $$P_n(k,p) = binom nm p^m(1-p)^{n-m}$$
                  should be satisfied, where
                  $p$ is the probability of the required cube face loss in the single test,
                  $n$ is the number of throws,
                  $m$ is the number of depositions of the required face.



                  For each face of an ideal cube $p=p_r=dfrac16.$



                  The fact of the loss of a given face $k$ times from $n$ shots denote $E_n(k).$



                  Let us consider the hypotheses $H_j = left{left|p-dfrac jnright|<dfrac1{2n}right},$
                  $bigg(p$ belongs to the neighbourhood of value $dfrac jnbigg),$

                  then
                  $$P(H_j) = F_n(H_j) = intlimits_{-1/(2n)}^{1/(2n)} P_nleft(m,dfrac jn+xright),mathrm dx,$$
                  $$P(H_j) = P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)}left(dfrac{dfrac jn+x}{dfrac jn}right)^m left(dfrac{dfrac {n-j}n-x}{dfrac{n-j}n}right)^{n-m}dx$$
                  $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/(2n)}^{1/(2n)} left(1+dfrac{nx}jright)^m left(1-dfrac{nx}{n-j}right)^{n-m}d(nx)$$
                  $$ = dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{y}jright)^m left(1-dfrac{y}{n-j}right)^{n-m}dy.$$
                  If $quad j>>dfrac12,quad n-j >>dfrac12,quad$ then
                  $$F_n(H_j) = {smalldfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac mj y+dfrac{m(m-1)}{2j^2}y^2+dotsright)left(1-dfrac {n-m}{n-j}y+dfrac {(n-m)(n-m-1)}{2(n-j)^2}y^2+dotsright)dy}$$
                  $$approx dfrac1n P_nleft(m,dfrac jnright)intlimits_{-1/2}^{1/2} left(1+dfrac{(m-j)n}{j(n-j)}y + dfrac{(n^2-n)(j-m)^2+m(n-m)}{2j^2(n-j)^2}y^2right)dy$$
                  $$ = dfrac jnleft(1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}right) P_nleft(m,dfrac jnright),$$
                  The probability of hypothesis for the ideal face of cube is
                  $$P(H_r) = F_{6k}(k)$$
                  The plot of the factor
                  $$1+dfrac{(n^2-n)(j-m)^2-mn(n-m)}{24j^2(n-j)^2}$$
                  for $m=10,quad n= 60,quad jin(7,15)$



                  factor



                  shows that prior probability of hypothesis $H_j$ is
                  $$P(H_j)approx dfrac 1n P_nleft(m,dfrac jnright)$$
                  with the accuracy $0.5%.$



                  Let $n=6m,$ then
                  begin{cases}
                  P(E_k | H_r) approx P_{6m}(m,k),\[4pt]
                  P(E_k | H_j) approx P_{6m}(j,k),\[4pt]
                  end{cases}



                  By the Bayesian formula for the posterior probability is
                  $$P(H_j | E_k) = dfrac{P(E_k | H_j)P(H_j | H_r)}{sum_{j=0}^{6m}P(E_k | H_i)P(H_i |)},$$



                  $$P(H_j | E_k) = dfrac{dbinom{6m}kleft(dfrac j{6m}right)^kleft(dfrac {6m-j}{6m}right)^{6m-k}{dfrac1{6m}dbinom{6m}j}left(dfrac 16right)^jleft(dfrac 56right)^{6m-j}}
                  {sumlimits_{i=0}^{6m}dbinom{6m}kleft(dfrac i{6m}right)^kleft(dfrac {6m-i}{6m}right)^{6m-k}dfrac1{6m}dbinom{6m}ileft(dfrac 16right)^ileft(dfrac 56right)^{6m-i}},$$

                  $$boxed{ P(H_j | E_k) = dfrac{dbinom{6m}j}
                  {sumlimits_{i=0}^{6m}dbinom{6m}i 5^{j-i} left(dfrac ijright)^kleft(dfrac{6m-i}{6m-j}right)^{6m-k}}},$$



                  The resulting formula indicates the posterior probability of the distribution of hypotheses about the probability of a given face of a cube falling out, provided that the given face fell $k$ times in $n$ throws



                  If the given face fell $10$ times in $60$ throws ($m=10, n=60, k=10$), then Wolfram Alpha plot is



                  m=10, n=60, k=10



                  If the given face fell $7$ times in $60$ throws ($m=10, n=60, k=7$), then Wolfram Alpha plot is



                  m=10, n=60, k=7



                  If the given face fell $15$ times in $60$ throws ($m=10, n=60, k=15$), then Wolfram Alpha plot is



                  m=10, n=60, k=15



                  $$textbf{Case 1}$$
                  $textbf{Can not be collected separate statistics for each cube.}$



                  This case can not provide valid data, because summation hides individual propertied of cubes. Besides, cubes can be tuned to the combinations as "6+5". Using of this issue data cannot be recommended.







                  share|cite|improve this answer












                  share|cite|improve this answer



                  share|cite|improve this answer










                  answered Dec 18 '18 at 22:14









                  Yuri Negometyanov

                  10.8k1726




                  10.8k1726






























                      draft saved

                      draft discarded




















































                      Thanks for contributing an answer to Mathematics Stack Exchange!


                      • Please be sure to answer the question. Provide details and share your research!

                      But avoid



                      • Asking for help, clarification, or responding to other answers.

                      • Making statements based on opinion; back them up with references or personal experience.


                      Use MathJax to format equations. MathJax reference.


                      To learn more, see our tips on writing great answers.





                      Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


                      Please pay close attention to the following guidance:


                      • Please be sure to answer the question. Provide details and share your research!

                      But avoid



                      • Asking for help, clarification, or responding to other answers.

                      • Making statements based on opinion; back them up with references or personal experience.


                      To learn more, see our tips on writing great answers.




                      draft saved


                      draft discarded














                      StackExchange.ready(
                      function () {
                      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3034432%2fhow-to-find-the-probability-of-a-mean-shift-in-a-random-process%23new-answer', 'question_page');
                      }
                      );

                      Post as a guest















                      Required, but never shown





















































                      Required, but never shown














                      Required, but never shown












                      Required, but never shown







                      Required, but never shown

































                      Required, but never shown














                      Required, but never shown












                      Required, but never shown







                      Required, but never shown







                      Popular posts from this blog

                      Bressuire

                      Cabo Verde

                      Gyllenstierna