Probability limits of random variable sums
$begingroup$
I have $X_1, X_2, X_3, cdots$ which are independent random variables with the same non-zero mean ($mune0$) and same variance $sigma^2$.
I would like to compute $$lim_{ntoinfty} P[frac{1}nsum^n_{i=1}X_i < frac{mu}{2}]$$ for $mu<0$ and $mu>0$.
My initial thought was to use the central limit theorem but it indicates the variables must be identically distributed which I dont have here -- only have first and second order moments are similar.
Any thoughts on how to start tackling this?
convergence central-limit-theorem probability-limit-theorems
$endgroup$
add a comment |
$begingroup$
I have $X_1, X_2, X_3, cdots$ which are independent random variables with the same non-zero mean ($mune0$) and same variance $sigma^2$.
I would like to compute $$lim_{ntoinfty} P[frac{1}nsum^n_{i=1}X_i < frac{mu}{2}]$$ for $mu<0$ and $mu>0$.
My initial thought was to use the central limit theorem but it indicates the variables must be identically distributed which I dont have here -- only have first and second order moments are similar.
Any thoughts on how to start tackling this?
convergence central-limit-theorem probability-limit-theorems
$endgroup$
add a comment |
$begingroup$
I have $X_1, X_2, X_3, cdots$ which are independent random variables with the same non-zero mean ($mune0$) and same variance $sigma^2$.
I would like to compute $$lim_{ntoinfty} P[frac{1}nsum^n_{i=1}X_i < frac{mu}{2}]$$ for $mu<0$ and $mu>0$.
My initial thought was to use the central limit theorem but it indicates the variables must be identically distributed which I dont have here -- only have first and second order moments are similar.
Any thoughts on how to start tackling this?
convergence central-limit-theorem probability-limit-theorems
$endgroup$
I have $X_1, X_2, X_3, cdots$ which are independent random variables with the same non-zero mean ($mune0$) and same variance $sigma^2$.
I would like to compute $$lim_{ntoinfty} P[frac{1}nsum^n_{i=1}X_i < frac{mu}{2}]$$ for $mu<0$ and $mu>0$.
My initial thought was to use the central limit theorem but it indicates the variables must be identically distributed which I dont have here -- only have first and second order moments are similar.
Any thoughts on how to start tackling this?
convergence central-limit-theorem probability-limit-theorems
convergence central-limit-theorem probability-limit-theorems
asked Dec 21 '18 at 2:28
AvedisAvedis
647
647
add a comment |
add a comment |
2 Answers
2
active
oldest
votes
$begingroup$
Chebychev's inequality may help.
For $mu > 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| > frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0.$$ For $mu < 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i ge frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| ge frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0$$ so $Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) to 1$.
$endgroup$
1
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
add a comment |
$begingroup$
Define $M_n=frac{1}{n}sum_{i-1}^nX_i$.
$E[M_n]=mu$, $Var[M_n]=frac{sigma_X^2}{n}$
For $mu<0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(using Chebyshev inequality)
$$P[|M_n-mu| < -frac{mu}{2}]$$
$$P[|M_n-mu| < -frac{mu}{2}] geq 1 - frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity, so the Probability is $1-0=1$.
For $mu > 0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(flip inequality due to division of both sides by -1)
$$P[-M_n+mu > frac{mu}{2}]$$
(using Chebyshev inequality and fact of $|A|=|-A|$)
$$P[|M_n-mu| > frac{mu}{2}]$$
$$P[|M_n-mu| < frac{mu}{2}] leq frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity. So, the probability is 0.
$endgroup$
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3048121%2fprobability-limits-of-random-variable-sums%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Chebychev's inequality may help.
For $mu > 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| > frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0.$$ For $mu < 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i ge frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| ge frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0$$ so $Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) to 1$.
$endgroup$
1
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
add a comment |
$begingroup$
Chebychev's inequality may help.
For $mu > 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| > frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0.$$ For $mu < 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i ge frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| ge frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0$$ so $Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) to 1$.
$endgroup$
1
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
add a comment |
$begingroup$
Chebychev's inequality may help.
For $mu > 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| > frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0.$$ For $mu < 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i ge frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| ge frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0$$ so $Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) to 1$.
$endgroup$
Chebychev's inequality may help.
For $mu > 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| > frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0.$$ For $mu < 0$, we have $$Pleft(frac{1}{n} sum_{i=1}^n X_i ge frac{mu}{2}right) le Pleft(left|frac{1}{n} sum_{i=1}^n X_i - mu right| ge frac{mu}{2}right) le frac{sigma^2/n}{mu^2/4} to 0$$ so $Pleft(frac{1}{n} sum_{i=1}^n X_i < frac{mu}{2}right) to 1$.
edited Dec 21 '18 at 2:56
answered Dec 21 '18 at 2:33
angryavianangryavian
41.8k23381
41.8k23381
1
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
add a comment |
1
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
1
1
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
@ClementC. You're right, thanks for catching my mistake.
$endgroup$
– angryavian
Dec 21 '18 at 2:56
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
Why is the right side of Chebychev's inquality have n and not n^2?
$endgroup$
– Avedis
Dec 23 '18 at 1:43
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
@Avedis Think carefully about what the variance of $frac{1}{n} sum_{i=1}^n X_i$ is.
$endgroup$
– angryavian
Dec 23 '18 at 2:08
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
$begingroup$
Thank you, I just figure out what you mentioned on my own. However, when I complete the problem, I get a different answer then you do for $mu>0$. Why?
$endgroup$
– Avedis
Dec 23 '18 at 3:51
add a comment |
$begingroup$
Define $M_n=frac{1}{n}sum_{i-1}^nX_i$.
$E[M_n]=mu$, $Var[M_n]=frac{sigma_X^2}{n}$
For $mu<0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(using Chebyshev inequality)
$$P[|M_n-mu| < -frac{mu}{2}]$$
$$P[|M_n-mu| < -frac{mu}{2}] geq 1 - frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity, so the Probability is $1-0=1$.
For $mu > 0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(flip inequality due to division of both sides by -1)
$$P[-M_n+mu > frac{mu}{2}]$$
(using Chebyshev inequality and fact of $|A|=|-A|$)
$$P[|M_n-mu| > frac{mu}{2}]$$
$$P[|M_n-mu| < frac{mu}{2}] leq frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity. So, the probability is 0.
$endgroup$
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
add a comment |
$begingroup$
Define $M_n=frac{1}{n}sum_{i-1}^nX_i$.
$E[M_n]=mu$, $Var[M_n]=frac{sigma_X^2}{n}$
For $mu<0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(using Chebyshev inequality)
$$P[|M_n-mu| < -frac{mu}{2}]$$
$$P[|M_n-mu| < -frac{mu}{2}] geq 1 - frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity, so the Probability is $1-0=1$.
For $mu > 0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(flip inequality due to division of both sides by -1)
$$P[-M_n+mu > frac{mu}{2}]$$
(using Chebyshev inequality and fact of $|A|=|-A|$)
$$P[|M_n-mu| > frac{mu}{2}]$$
$$P[|M_n-mu| < frac{mu}{2}] leq frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity. So, the probability is 0.
$endgroup$
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
add a comment |
$begingroup$
Define $M_n=frac{1}{n}sum_{i-1}^nX_i$.
$E[M_n]=mu$, $Var[M_n]=frac{sigma_X^2}{n}$
For $mu<0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(using Chebyshev inequality)
$$P[|M_n-mu| < -frac{mu}{2}]$$
$$P[|M_n-mu| < -frac{mu}{2}] geq 1 - frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity, so the Probability is $1-0=1$.
For $mu > 0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(flip inequality due to division of both sides by -1)
$$P[-M_n+mu > frac{mu}{2}]$$
(using Chebyshev inequality and fact of $|A|=|-A|$)
$$P[|M_n-mu| > frac{mu}{2}]$$
$$P[|M_n-mu| < frac{mu}{2}] leq frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity. So, the probability is 0.
$endgroup$
Define $M_n=frac{1}{n}sum_{i-1}^nX_i$.
$E[M_n]=mu$, $Var[M_n]=frac{sigma_X^2}{n}$
For $mu<0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(using Chebyshev inequality)
$$P[|M_n-mu| < -frac{mu}{2}]$$
$$P[|M_n-mu| < -frac{mu}{2}] geq 1 - frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity, so the Probability is $1-0=1$.
For $mu > 0$:
$$P[M_n<frac{mu}{2}]$$
$$P[M_n-mu<frac{mu}{2}-mu]$$ (Subtracting $mu$ from both sides)
$$P[M_n-mu<frac{mu}{2}-mu]$$
$$P[M_n-mu< -frac{mu}{2}]$$
(flip inequality due to division of both sides by -1)
$$P[-M_n+mu > frac{mu}{2}]$$
(using Chebyshev inequality and fact of $|A|=|-A|$)
$$P[|M_n-mu| > frac{mu}{2}]$$
$$P[|M_n-mu| < frac{mu}{2}] leq frac{sigma_X^2/n}{mu^2/4}$$ which goes to zero when n goes to infinity. So, the probability is 0.
edited Dec 23 '18 at 14:05
answered Dec 23 '18 at 3:50
AvedisAvedis
647
647
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
add a comment |
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
In both cases you should have $(mu/2)^2=mu^2/4$ rather than $mu^2/2$. And it seems our answers agree.
$endgroup$
– angryavian
Dec 23 '18 at 4:20
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
$begingroup$
You are right and I will fix. I misread your answer. I thought the last statement was a summary of the entire problem, not the last part exclusively. Thank you!
$endgroup$
– Avedis
Dec 23 '18 at 14:04
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3048121%2fprobability-limits-of-random-variable-sums%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown