Density and expectation of the range of a sample of uniform random variables
If the variables $alpha_1$...$alpha_n$ are distributed uniformly in $(0,1)$,
- How do I show that the spread $alpha_{(n)}$ - $alpha_{(1)}$ has density $n (n-1) x^{n-2} (1-x)$ and expectation $(n-1)/(n+1)$?
- What is the probability that all $n$ points lie within an interval of length $t$?
probability
add a comment |
If the variables $alpha_1$...$alpha_n$ are distributed uniformly in $(0,1)$,
- How do I show that the spread $alpha_{(n)}$ - $alpha_{(1)}$ has density $n (n-1) x^{n-2} (1-x)$ and expectation $(n-1)/(n+1)$?
- What is the probability that all $n$ points lie within an interval of length $t$?
probability
Homework? What have you tried? Where are you stuck?
– cardinal
Apr 19 '11 at 3:40
I'm merging your duplicate accounts. Please consider registering so the software can keep track of who you are.
– Willie Wong
Apr 19 '11 at 14:39
add a comment |
If the variables $alpha_1$...$alpha_n$ are distributed uniformly in $(0,1)$,
- How do I show that the spread $alpha_{(n)}$ - $alpha_{(1)}$ has density $n (n-1) x^{n-2} (1-x)$ and expectation $(n-1)/(n+1)$?
- What is the probability that all $n$ points lie within an interval of length $t$?
probability
If the variables $alpha_1$...$alpha_n$ are distributed uniformly in $(0,1)$,
- How do I show that the spread $alpha_{(n)}$ - $alpha_{(1)}$ has density $n (n-1) x^{n-2} (1-x)$ and expectation $(n-1)/(n+1)$?
- What is the probability that all $n$ points lie within an interval of length $t$?
probability
probability
edited Apr 19 '11 at 3:44
cardinal
6,40632743
6,40632743
asked Apr 19 '11 at 3:37
user9767
Homework? What have you tried? Where are you stuck?
– cardinal
Apr 19 '11 at 3:40
I'm merging your duplicate accounts. Please consider registering so the software can keep track of who you are.
– Willie Wong
Apr 19 '11 at 14:39
add a comment |
Homework? What have you tried? Where are you stuck?
– cardinal
Apr 19 '11 at 3:40
I'm merging your duplicate accounts. Please consider registering so the software can keep track of who you are.
– Willie Wong
Apr 19 '11 at 14:39
Homework? What have you tried? Where are you stuck?
– cardinal
Apr 19 '11 at 3:40
Homework? What have you tried? Where are you stuck?
– cardinal
Apr 19 '11 at 3:40
I'm merging your duplicate accounts. Please consider registering so the software can keep track of who you are.
– Willie Wong
Apr 19 '11 at 14:39
I'm merging your duplicate accounts. Please consider registering so the software can keep track of who you are.
– Willie Wong
Apr 19 '11 at 14:39
add a comment |
3 Answers
3
active
oldest
votes
This is a (straight forward?) exercise is order statistics.
From Wikipedia, the pdf for the joint order statistic is:
$$ f_{ X_{(i)}, X_{(j)} }(u, v) = $$
$$ frac{ n! }{ (i-1)! (j-1-i)! (n-j)! } [ F_X(u) ]^{i-1} [F_X(v) - F_X(u)]^{j-1-i} [1 - F_X(v)]^{n-j} f_X(u) f_X(v) $$
(note the change in arguments)
Where $F_X(u)$ is the cumulative distribution function (or cdf) and $f_X(u)$ is the probability density function (or pdf) of the random variable $X$.
Plugging in $i=1$ and $j=n$ with $X = alpha$, $f_{alpha}(u) = 1$ and $F_{alpha}(u) = u$, we get:
$$ f_{ alpha_{(1)}, alpha_{(n)} }(s, s+t) = frac{ n! }{ (n-2)! } [ s+t - s ]^{n-2} = n (n-1) t^{n-2} $$
for a starting point $s in [0, 1-t] $ with interval length $t$. To find the probability of finding $alpha_{(1)}$ and $alpha_{(n)}$ within some interval of $t$, we need to integrate over all (permissible) starting positions, $s$:
$$ int_0^{1-t} f_{alpha_{(1)},alpha_{(n)}}(s, s+t) ds = int_{0}^{1-t} n (n-1) t^{n-2} ds = n (n-1) t^{n-2} int_0^{1-t} ds $$
$$ = n (n-1) t^{n-2} (1-t) $$
Which gives the first part of the answer for question 1. For the second part to question 1, you could integrate the above equation for $t in [0, 1]$ or you could just take the expectation of the difference of the random variable $alpha_{(n)}$ with $alpha_{(1)}$. I will do the latter. The pdf of the $k$'th order statistic is:
$$ f_{X_{(k)}}(u) = frac{ n! }{ (k-1)! (n-k)! } [F_X(u)]^{k-1} [1-F_X(u)]^{n-k} f_X(u) $$
and plugging in $X = alpha$, $k=1$ and $k=n$ for the 1st and $n$'th order statistic (of the uniform distribution) respectively, we get:
$$ f_{alpha_{(1)}}(u) = n (1 - u)^{n-1} $$
$$ f_{alpha_{(n)}}(u) = n u^{n-1} $$
and taking the expectation of their difference:
$$ E[ alpha_{(n)} - alpha_{(1)} ] = int_{0}^{1} u (n u^{n-1} - n (1 - u)^{n-1} ) du = frac{n}{n+1} - frac{1}{n+1} = frac{n-1}{n+1} $$
As desired. (note that $alpha_{(n)}$ and $alpha_{(1)}$ are not independent but that the expectation of the sums of random variables is still the same regardless).
To answer part 2, one must find $Pr{ alpha_{(n)} - alpha_{(1)} le t } $. We just derived the pdf of the joint distribution that the difference of $alpha_{(n)}$ and $alpha_{(1)}$ is of length $t$, so now we just have to sum over all lengths less than $t$:
$$ int_0^{t} n (n-1) x^{n-2} (1-x) dx = t^{n-1} ( n - (n-1) t) $$
I must admit that I just looked up the pdf of the joint order statistic. Could anyone tell me (or give a reference) on how to derive this equation?
1
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
add a comment |
Here's a hint: The answer to part 2 is quite easy. And if you consider it, you might see a trick on how to get the answer to part 1.
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
add a comment |
Here's approach for computing the density function that doesn't rely on knowledge of order statistics. First let $R = alpha_{(n)} - alpha_{(1)}$ be the range. We compute the cumulative distribution function $F_R(x)$ by conditioning on the event $alpha_{(n)} leq x$.
begin{align}
F_R(x) &= P{R leq x} \
&= P{R leq x text{ and } alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} \
&=P{R leq x text{ | } alpha_{(n)} leq x}P{alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} label{a}tag{1}
end{align}
To compute the first term in $ref{a}$, observe that if the largest value is less than $x$, then the range $R$ must be less than x, so we have
begin{align}
P{R leq x text{ | } alpha_{(n)} leq x} = 1 \
end{align}
Also,
begin{align}
P{alpha_{(n)}leq x} &= P{alpha_1 leq x, alpha_2 leq x,ldots,alpha_n leq x} \
&= P{alpha_1 leq x}P{alpha_2 leq x}cdots P{alpha_n leq x} \
&= x^n
end{align}
To finish our computation of $F_R$, we must compute $P{R leq x text{ and } alpha_{(n)} > x} $. Let $A$ be the event that $R leq x text{ and } alpha_{(n)} > x$. Let $A_i$ be the event that $alpha_i > x$ and $alpha_i - x leq alpha_j leq alpha_i$ for all $j neq i$. Clearly $P{alpha_i} > x = (1-x)$. Because the interval $[alpha_i - x,alpha_i]$ is of width $x$, we have $P{alpha_i - x leq alpha_j leq alpha_i} = x$, which implies
$$P{alpha_i - x leq alpha_j leq alpha_i text{ for all } j neq i } = x^{n-1}$$
Therefore
$$P{A_i}= (1-x)x^{n-1} $$
Notice that $A = bigcuplimits_{i=1}^n A_i$, and the events $A_i$ are mutually exclusive, so we have
$$P(A) = P{R leq x text{ and } alpha_{(n)} > x} = n(1-x)x^{n-1}$$
Substituting these results into $ref{a}$, we get
$$F_R(x) = x^n + n(1-x)x^{n-1}$$
Differentiation and a little algebra gives the desired density function
begin{align}
f_R(x) &= F_R'(x) \
&= n(n-1)x^{n-2}(1-x)
end{align}
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f33767%2fdensity-and-expectation-of-the-range-of-a-sample-of-uniform-random-variables%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
This is a (straight forward?) exercise is order statistics.
From Wikipedia, the pdf for the joint order statistic is:
$$ f_{ X_{(i)}, X_{(j)} }(u, v) = $$
$$ frac{ n! }{ (i-1)! (j-1-i)! (n-j)! } [ F_X(u) ]^{i-1} [F_X(v) - F_X(u)]^{j-1-i} [1 - F_X(v)]^{n-j} f_X(u) f_X(v) $$
(note the change in arguments)
Where $F_X(u)$ is the cumulative distribution function (or cdf) and $f_X(u)$ is the probability density function (or pdf) of the random variable $X$.
Plugging in $i=1$ and $j=n$ with $X = alpha$, $f_{alpha}(u) = 1$ and $F_{alpha}(u) = u$, we get:
$$ f_{ alpha_{(1)}, alpha_{(n)} }(s, s+t) = frac{ n! }{ (n-2)! } [ s+t - s ]^{n-2} = n (n-1) t^{n-2} $$
for a starting point $s in [0, 1-t] $ with interval length $t$. To find the probability of finding $alpha_{(1)}$ and $alpha_{(n)}$ within some interval of $t$, we need to integrate over all (permissible) starting positions, $s$:
$$ int_0^{1-t} f_{alpha_{(1)},alpha_{(n)}}(s, s+t) ds = int_{0}^{1-t} n (n-1) t^{n-2} ds = n (n-1) t^{n-2} int_0^{1-t} ds $$
$$ = n (n-1) t^{n-2} (1-t) $$
Which gives the first part of the answer for question 1. For the second part to question 1, you could integrate the above equation for $t in [0, 1]$ or you could just take the expectation of the difference of the random variable $alpha_{(n)}$ with $alpha_{(1)}$. I will do the latter. The pdf of the $k$'th order statistic is:
$$ f_{X_{(k)}}(u) = frac{ n! }{ (k-1)! (n-k)! } [F_X(u)]^{k-1} [1-F_X(u)]^{n-k} f_X(u) $$
and plugging in $X = alpha$, $k=1$ and $k=n$ for the 1st and $n$'th order statistic (of the uniform distribution) respectively, we get:
$$ f_{alpha_{(1)}}(u) = n (1 - u)^{n-1} $$
$$ f_{alpha_{(n)}}(u) = n u^{n-1} $$
and taking the expectation of their difference:
$$ E[ alpha_{(n)} - alpha_{(1)} ] = int_{0}^{1} u (n u^{n-1} - n (1 - u)^{n-1} ) du = frac{n}{n+1} - frac{1}{n+1} = frac{n-1}{n+1} $$
As desired. (note that $alpha_{(n)}$ and $alpha_{(1)}$ are not independent but that the expectation of the sums of random variables is still the same regardless).
To answer part 2, one must find $Pr{ alpha_{(n)} - alpha_{(1)} le t } $. We just derived the pdf of the joint distribution that the difference of $alpha_{(n)}$ and $alpha_{(1)}$ is of length $t$, so now we just have to sum over all lengths less than $t$:
$$ int_0^{t} n (n-1) x^{n-2} (1-x) dx = t^{n-1} ( n - (n-1) t) $$
I must admit that I just looked up the pdf of the joint order statistic. Could anyone tell me (or give a reference) on how to derive this equation?
1
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
add a comment |
This is a (straight forward?) exercise is order statistics.
From Wikipedia, the pdf for the joint order statistic is:
$$ f_{ X_{(i)}, X_{(j)} }(u, v) = $$
$$ frac{ n! }{ (i-1)! (j-1-i)! (n-j)! } [ F_X(u) ]^{i-1} [F_X(v) - F_X(u)]^{j-1-i} [1 - F_X(v)]^{n-j} f_X(u) f_X(v) $$
(note the change in arguments)
Where $F_X(u)$ is the cumulative distribution function (or cdf) and $f_X(u)$ is the probability density function (or pdf) of the random variable $X$.
Plugging in $i=1$ and $j=n$ with $X = alpha$, $f_{alpha}(u) = 1$ and $F_{alpha}(u) = u$, we get:
$$ f_{ alpha_{(1)}, alpha_{(n)} }(s, s+t) = frac{ n! }{ (n-2)! } [ s+t - s ]^{n-2} = n (n-1) t^{n-2} $$
for a starting point $s in [0, 1-t] $ with interval length $t$. To find the probability of finding $alpha_{(1)}$ and $alpha_{(n)}$ within some interval of $t$, we need to integrate over all (permissible) starting positions, $s$:
$$ int_0^{1-t} f_{alpha_{(1)},alpha_{(n)}}(s, s+t) ds = int_{0}^{1-t} n (n-1) t^{n-2} ds = n (n-1) t^{n-2} int_0^{1-t} ds $$
$$ = n (n-1) t^{n-2} (1-t) $$
Which gives the first part of the answer for question 1. For the second part to question 1, you could integrate the above equation for $t in [0, 1]$ or you could just take the expectation of the difference of the random variable $alpha_{(n)}$ with $alpha_{(1)}$. I will do the latter. The pdf of the $k$'th order statistic is:
$$ f_{X_{(k)}}(u) = frac{ n! }{ (k-1)! (n-k)! } [F_X(u)]^{k-1} [1-F_X(u)]^{n-k} f_X(u) $$
and plugging in $X = alpha$, $k=1$ and $k=n$ for the 1st and $n$'th order statistic (of the uniform distribution) respectively, we get:
$$ f_{alpha_{(1)}}(u) = n (1 - u)^{n-1} $$
$$ f_{alpha_{(n)}}(u) = n u^{n-1} $$
and taking the expectation of their difference:
$$ E[ alpha_{(n)} - alpha_{(1)} ] = int_{0}^{1} u (n u^{n-1} - n (1 - u)^{n-1} ) du = frac{n}{n+1} - frac{1}{n+1} = frac{n-1}{n+1} $$
As desired. (note that $alpha_{(n)}$ and $alpha_{(1)}$ are not independent but that the expectation of the sums of random variables is still the same regardless).
To answer part 2, one must find $Pr{ alpha_{(n)} - alpha_{(1)} le t } $. We just derived the pdf of the joint distribution that the difference of $alpha_{(n)}$ and $alpha_{(1)}$ is of length $t$, so now we just have to sum over all lengths less than $t$:
$$ int_0^{t} n (n-1) x^{n-2} (1-x) dx = t^{n-1} ( n - (n-1) t) $$
I must admit that I just looked up the pdf of the joint order statistic. Could anyone tell me (or give a reference) on how to derive this equation?
1
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
add a comment |
This is a (straight forward?) exercise is order statistics.
From Wikipedia, the pdf for the joint order statistic is:
$$ f_{ X_{(i)}, X_{(j)} }(u, v) = $$
$$ frac{ n! }{ (i-1)! (j-1-i)! (n-j)! } [ F_X(u) ]^{i-1} [F_X(v) - F_X(u)]^{j-1-i} [1 - F_X(v)]^{n-j} f_X(u) f_X(v) $$
(note the change in arguments)
Where $F_X(u)$ is the cumulative distribution function (or cdf) and $f_X(u)$ is the probability density function (or pdf) of the random variable $X$.
Plugging in $i=1$ and $j=n$ with $X = alpha$, $f_{alpha}(u) = 1$ and $F_{alpha}(u) = u$, we get:
$$ f_{ alpha_{(1)}, alpha_{(n)} }(s, s+t) = frac{ n! }{ (n-2)! } [ s+t - s ]^{n-2} = n (n-1) t^{n-2} $$
for a starting point $s in [0, 1-t] $ with interval length $t$. To find the probability of finding $alpha_{(1)}$ and $alpha_{(n)}$ within some interval of $t$, we need to integrate over all (permissible) starting positions, $s$:
$$ int_0^{1-t} f_{alpha_{(1)},alpha_{(n)}}(s, s+t) ds = int_{0}^{1-t} n (n-1) t^{n-2} ds = n (n-1) t^{n-2} int_0^{1-t} ds $$
$$ = n (n-1) t^{n-2} (1-t) $$
Which gives the first part of the answer for question 1. For the second part to question 1, you could integrate the above equation for $t in [0, 1]$ or you could just take the expectation of the difference of the random variable $alpha_{(n)}$ with $alpha_{(1)}$. I will do the latter. The pdf of the $k$'th order statistic is:
$$ f_{X_{(k)}}(u) = frac{ n! }{ (k-1)! (n-k)! } [F_X(u)]^{k-1} [1-F_X(u)]^{n-k} f_X(u) $$
and plugging in $X = alpha$, $k=1$ and $k=n$ for the 1st and $n$'th order statistic (of the uniform distribution) respectively, we get:
$$ f_{alpha_{(1)}}(u) = n (1 - u)^{n-1} $$
$$ f_{alpha_{(n)}}(u) = n u^{n-1} $$
and taking the expectation of their difference:
$$ E[ alpha_{(n)} - alpha_{(1)} ] = int_{0}^{1} u (n u^{n-1} - n (1 - u)^{n-1} ) du = frac{n}{n+1} - frac{1}{n+1} = frac{n-1}{n+1} $$
As desired. (note that $alpha_{(n)}$ and $alpha_{(1)}$ are not independent but that the expectation of the sums of random variables is still the same regardless).
To answer part 2, one must find $Pr{ alpha_{(n)} - alpha_{(1)} le t } $. We just derived the pdf of the joint distribution that the difference of $alpha_{(n)}$ and $alpha_{(1)}$ is of length $t$, so now we just have to sum over all lengths less than $t$:
$$ int_0^{t} n (n-1) x^{n-2} (1-x) dx = t^{n-1} ( n - (n-1) t) $$
I must admit that I just looked up the pdf of the joint order statistic. Could anyone tell me (or give a reference) on how to derive this equation?
This is a (straight forward?) exercise is order statistics.
From Wikipedia, the pdf for the joint order statistic is:
$$ f_{ X_{(i)}, X_{(j)} }(u, v) = $$
$$ frac{ n! }{ (i-1)! (j-1-i)! (n-j)! } [ F_X(u) ]^{i-1} [F_X(v) - F_X(u)]^{j-1-i} [1 - F_X(v)]^{n-j} f_X(u) f_X(v) $$
(note the change in arguments)
Where $F_X(u)$ is the cumulative distribution function (or cdf) and $f_X(u)$ is the probability density function (or pdf) of the random variable $X$.
Plugging in $i=1$ and $j=n$ with $X = alpha$, $f_{alpha}(u) = 1$ and $F_{alpha}(u) = u$, we get:
$$ f_{ alpha_{(1)}, alpha_{(n)} }(s, s+t) = frac{ n! }{ (n-2)! } [ s+t - s ]^{n-2} = n (n-1) t^{n-2} $$
for a starting point $s in [0, 1-t] $ with interval length $t$. To find the probability of finding $alpha_{(1)}$ and $alpha_{(n)}$ within some interval of $t$, we need to integrate over all (permissible) starting positions, $s$:
$$ int_0^{1-t} f_{alpha_{(1)},alpha_{(n)}}(s, s+t) ds = int_{0}^{1-t} n (n-1) t^{n-2} ds = n (n-1) t^{n-2} int_0^{1-t} ds $$
$$ = n (n-1) t^{n-2} (1-t) $$
Which gives the first part of the answer for question 1. For the second part to question 1, you could integrate the above equation for $t in [0, 1]$ or you could just take the expectation of the difference of the random variable $alpha_{(n)}$ with $alpha_{(1)}$. I will do the latter. The pdf of the $k$'th order statistic is:
$$ f_{X_{(k)}}(u) = frac{ n! }{ (k-1)! (n-k)! } [F_X(u)]^{k-1} [1-F_X(u)]^{n-k} f_X(u) $$
and plugging in $X = alpha$, $k=1$ and $k=n$ for the 1st and $n$'th order statistic (of the uniform distribution) respectively, we get:
$$ f_{alpha_{(1)}}(u) = n (1 - u)^{n-1} $$
$$ f_{alpha_{(n)}}(u) = n u^{n-1} $$
and taking the expectation of their difference:
$$ E[ alpha_{(n)} - alpha_{(1)} ] = int_{0}^{1} u (n u^{n-1} - n (1 - u)^{n-1} ) du = frac{n}{n+1} - frac{1}{n+1} = frac{n-1}{n+1} $$
As desired. (note that $alpha_{(n)}$ and $alpha_{(1)}$ are not independent but that the expectation of the sums of random variables is still the same regardless).
To answer part 2, one must find $Pr{ alpha_{(n)} - alpha_{(1)} le t } $. We just derived the pdf of the joint distribution that the difference of $alpha_{(n)}$ and $alpha_{(1)}$ is of length $t$, so now we just have to sum over all lengths less than $t$:
$$ int_0^{t} n (n-1) x^{n-2} (1-x) dx = t^{n-1} ( n - (n-1) t) $$
I must admit that I just looked up the pdf of the joint order statistic. Could anyone tell me (or give a reference) on how to derive this equation?
edited May 11 '16 at 20:16
Community♦
1
1
answered Apr 20 '11 at 1:22
user4143
827610
827610
1
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
add a comment |
1
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
1
1
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
The derivation of the joint density of order statistics is as follows: we want the probability that $U_{(i)}$ is in some small interval $[u, u + du]$ and $U_{(j)}$ is in some small interval $[v, v + dv]$. This amounts to finding the probability that $i-1$ of $U_1, ldots, U_n$ is less than $u$, one is in $[u, u + du]$, $j-i-1$ are between $u+du$ and $v$, one is in $[v, v + dv]$, and the remaining $n-j$ are greater than $v$. The five clauses of the previous sentence give the five factors; the embedded combinatorial problem gives the premultiplying factor.
– Michael Lugo
Apr 20 '11 at 2:23
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
@Michael, Thanks. Obvious in retrospect.
– user4143
Apr 20 '11 at 4:08
add a comment |
Here's a hint: The answer to part 2 is quite easy. And if you consider it, you might see a trick on how to get the answer to part 1.
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
add a comment |
Here's a hint: The answer to part 2 is quite easy. And if you consider it, you might see a trick on how to get the answer to part 1.
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
add a comment |
Here's a hint: The answer to part 2 is quite easy. And if you consider it, you might see a trick on how to get the answer to part 1.
Here's a hint: The answer to part 2 is quite easy. And if you consider it, you might see a trick on how to get the answer to part 1.
answered Apr 19 '11 at 3:59
Carl Brannen
1,74811529
1,74811529
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
add a comment |
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
I understand, I think that if I can find P(Xn - X1 < t), which is P(Xn < X1 + t), then I will have the distribution for Xn - X1 and will then be able to show the density and the expectation. But I'm not exactly sure how to get there.
– user9767
Apr 19 '11 at 4:14
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
Alright, I took the integral of n(n-1)x^(n-2)(1-x) with respect to x from o to t and got the answer nt^(n-1) - (n-1)t^n as the probability. still am not seeing part 1, however
– user9767
Apr 19 '11 at 4:32
add a comment |
Here's approach for computing the density function that doesn't rely on knowledge of order statistics. First let $R = alpha_{(n)} - alpha_{(1)}$ be the range. We compute the cumulative distribution function $F_R(x)$ by conditioning on the event $alpha_{(n)} leq x$.
begin{align}
F_R(x) &= P{R leq x} \
&= P{R leq x text{ and } alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} \
&=P{R leq x text{ | } alpha_{(n)} leq x}P{alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} label{a}tag{1}
end{align}
To compute the first term in $ref{a}$, observe that if the largest value is less than $x$, then the range $R$ must be less than x, so we have
begin{align}
P{R leq x text{ | } alpha_{(n)} leq x} = 1 \
end{align}
Also,
begin{align}
P{alpha_{(n)}leq x} &= P{alpha_1 leq x, alpha_2 leq x,ldots,alpha_n leq x} \
&= P{alpha_1 leq x}P{alpha_2 leq x}cdots P{alpha_n leq x} \
&= x^n
end{align}
To finish our computation of $F_R$, we must compute $P{R leq x text{ and } alpha_{(n)} > x} $. Let $A$ be the event that $R leq x text{ and } alpha_{(n)} > x$. Let $A_i$ be the event that $alpha_i > x$ and $alpha_i - x leq alpha_j leq alpha_i$ for all $j neq i$. Clearly $P{alpha_i} > x = (1-x)$. Because the interval $[alpha_i - x,alpha_i]$ is of width $x$, we have $P{alpha_i - x leq alpha_j leq alpha_i} = x$, which implies
$$P{alpha_i - x leq alpha_j leq alpha_i text{ for all } j neq i } = x^{n-1}$$
Therefore
$$P{A_i}= (1-x)x^{n-1} $$
Notice that $A = bigcuplimits_{i=1}^n A_i$, and the events $A_i$ are mutually exclusive, so we have
$$P(A) = P{R leq x text{ and } alpha_{(n)} > x} = n(1-x)x^{n-1}$$
Substituting these results into $ref{a}$, we get
$$F_R(x) = x^n + n(1-x)x^{n-1}$$
Differentiation and a little algebra gives the desired density function
begin{align}
f_R(x) &= F_R'(x) \
&= n(n-1)x^{n-2}(1-x)
end{align}
add a comment |
Here's approach for computing the density function that doesn't rely on knowledge of order statistics. First let $R = alpha_{(n)} - alpha_{(1)}$ be the range. We compute the cumulative distribution function $F_R(x)$ by conditioning on the event $alpha_{(n)} leq x$.
begin{align}
F_R(x) &= P{R leq x} \
&= P{R leq x text{ and } alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} \
&=P{R leq x text{ | } alpha_{(n)} leq x}P{alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} label{a}tag{1}
end{align}
To compute the first term in $ref{a}$, observe that if the largest value is less than $x$, then the range $R$ must be less than x, so we have
begin{align}
P{R leq x text{ | } alpha_{(n)} leq x} = 1 \
end{align}
Also,
begin{align}
P{alpha_{(n)}leq x} &= P{alpha_1 leq x, alpha_2 leq x,ldots,alpha_n leq x} \
&= P{alpha_1 leq x}P{alpha_2 leq x}cdots P{alpha_n leq x} \
&= x^n
end{align}
To finish our computation of $F_R$, we must compute $P{R leq x text{ and } alpha_{(n)} > x} $. Let $A$ be the event that $R leq x text{ and } alpha_{(n)} > x$. Let $A_i$ be the event that $alpha_i > x$ and $alpha_i - x leq alpha_j leq alpha_i$ for all $j neq i$. Clearly $P{alpha_i} > x = (1-x)$. Because the interval $[alpha_i - x,alpha_i]$ is of width $x$, we have $P{alpha_i - x leq alpha_j leq alpha_i} = x$, which implies
$$P{alpha_i - x leq alpha_j leq alpha_i text{ for all } j neq i } = x^{n-1}$$
Therefore
$$P{A_i}= (1-x)x^{n-1} $$
Notice that $A = bigcuplimits_{i=1}^n A_i$, and the events $A_i$ are mutually exclusive, so we have
$$P(A) = P{R leq x text{ and } alpha_{(n)} > x} = n(1-x)x^{n-1}$$
Substituting these results into $ref{a}$, we get
$$F_R(x) = x^n + n(1-x)x^{n-1}$$
Differentiation and a little algebra gives the desired density function
begin{align}
f_R(x) &= F_R'(x) \
&= n(n-1)x^{n-2}(1-x)
end{align}
add a comment |
Here's approach for computing the density function that doesn't rely on knowledge of order statistics. First let $R = alpha_{(n)} - alpha_{(1)}$ be the range. We compute the cumulative distribution function $F_R(x)$ by conditioning on the event $alpha_{(n)} leq x$.
begin{align}
F_R(x) &= P{R leq x} \
&= P{R leq x text{ and } alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} \
&=P{R leq x text{ | } alpha_{(n)} leq x}P{alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} label{a}tag{1}
end{align}
To compute the first term in $ref{a}$, observe that if the largest value is less than $x$, then the range $R$ must be less than x, so we have
begin{align}
P{R leq x text{ | } alpha_{(n)} leq x} = 1 \
end{align}
Also,
begin{align}
P{alpha_{(n)}leq x} &= P{alpha_1 leq x, alpha_2 leq x,ldots,alpha_n leq x} \
&= P{alpha_1 leq x}P{alpha_2 leq x}cdots P{alpha_n leq x} \
&= x^n
end{align}
To finish our computation of $F_R$, we must compute $P{R leq x text{ and } alpha_{(n)} > x} $. Let $A$ be the event that $R leq x text{ and } alpha_{(n)} > x$. Let $A_i$ be the event that $alpha_i > x$ and $alpha_i - x leq alpha_j leq alpha_i$ for all $j neq i$. Clearly $P{alpha_i} > x = (1-x)$. Because the interval $[alpha_i - x,alpha_i]$ is of width $x$, we have $P{alpha_i - x leq alpha_j leq alpha_i} = x$, which implies
$$P{alpha_i - x leq alpha_j leq alpha_i text{ for all } j neq i } = x^{n-1}$$
Therefore
$$P{A_i}= (1-x)x^{n-1} $$
Notice that $A = bigcuplimits_{i=1}^n A_i$, and the events $A_i$ are mutually exclusive, so we have
$$P(A) = P{R leq x text{ and } alpha_{(n)} > x} = n(1-x)x^{n-1}$$
Substituting these results into $ref{a}$, we get
$$F_R(x) = x^n + n(1-x)x^{n-1}$$
Differentiation and a little algebra gives the desired density function
begin{align}
f_R(x) &= F_R'(x) \
&= n(n-1)x^{n-2}(1-x)
end{align}
Here's approach for computing the density function that doesn't rely on knowledge of order statistics. First let $R = alpha_{(n)} - alpha_{(1)}$ be the range. We compute the cumulative distribution function $F_R(x)$ by conditioning on the event $alpha_{(n)} leq x$.
begin{align}
F_R(x) &= P{R leq x} \
&= P{R leq x text{ and } alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} \
&=P{R leq x text{ | } alpha_{(n)} leq x}P{alpha_{(n)} leq x}+P{R leq x text{ and } alpha_{(n)} > x} label{a}tag{1}
end{align}
To compute the first term in $ref{a}$, observe that if the largest value is less than $x$, then the range $R$ must be less than x, so we have
begin{align}
P{R leq x text{ | } alpha_{(n)} leq x} = 1 \
end{align}
Also,
begin{align}
P{alpha_{(n)}leq x} &= P{alpha_1 leq x, alpha_2 leq x,ldots,alpha_n leq x} \
&= P{alpha_1 leq x}P{alpha_2 leq x}cdots P{alpha_n leq x} \
&= x^n
end{align}
To finish our computation of $F_R$, we must compute $P{R leq x text{ and } alpha_{(n)} > x} $. Let $A$ be the event that $R leq x text{ and } alpha_{(n)} > x$. Let $A_i$ be the event that $alpha_i > x$ and $alpha_i - x leq alpha_j leq alpha_i$ for all $j neq i$. Clearly $P{alpha_i} > x = (1-x)$. Because the interval $[alpha_i - x,alpha_i]$ is of width $x$, we have $P{alpha_i - x leq alpha_j leq alpha_i} = x$, which implies
$$P{alpha_i - x leq alpha_j leq alpha_i text{ for all } j neq i } = x^{n-1}$$
Therefore
$$P{A_i}= (1-x)x^{n-1} $$
Notice that $A = bigcuplimits_{i=1}^n A_i$, and the events $A_i$ are mutually exclusive, so we have
$$P(A) = P{R leq x text{ and } alpha_{(n)} > x} = n(1-x)x^{n-1}$$
Substituting these results into $ref{a}$, we get
$$F_R(x) = x^n + n(1-x)x^{n-1}$$
Differentiation and a little algebra gives the desired density function
begin{align}
f_R(x) &= F_R'(x) \
&= n(n-1)x^{n-2}(1-x)
end{align}
answered Dec 9 at 5:28
Chad
1857
1857
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f33767%2fdensity-and-expectation-of-the-range-of-a-sample-of-uniform-random-variables%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Homework? What have you tried? Where are you stuck?
– cardinal
Apr 19 '11 at 3:40
I'm merging your duplicate accounts. Please consider registering so the software can keep track of who you are.
– Willie Wong
Apr 19 '11 at 14:39