Integral difference is non-negative
$begingroup$
Problem
This is a follow-up of this question.
Let $G$ be a proper prior distribution for $thetain[0,1]$ and $ain(0,1)$ be a fixed value. I'd like to show that the following quantity is non-negative for any $a$ and for any integer $0leq xleq n-1$,
$$
g(x)=int_{a}^1 s^{x+1}(1-s)^{n-x-1}dG(s)int_{0}^{a} t^{x}(1-t)^{n-x}dG(t)-int_{a}^1 s^{x}(1-s)^{n-x}dG(s)int_{0}^{a} t^{x+1}(1-t)^{n-x-1}dG(t)
$$
I've simulated that this is true when $G$ is uniform or some beta distributions. However, I have some trouble showing this algebraically. I can lower-bound the first product of integrals using Cauchy-Schwarz and upper-bound terms in the second product of integrals by the product of expectations over $G$, noting that $s$ and $1-s$ are negatively correlated. However, none of these give me useful results for showing $g(x)geq 0$.
Here's what I have so far:
First let $f(s)=frac{s}{1-s}$ and $h(s)=s^x (1-s)^{n-x}$. Then by expressing $int_a^1 f(s)h(s)dG(s)=int_0^1 f(s)h(s)dG(s)-int_a^1 f(s)h(s)dG(s)$ (and similarly for $int_a^1 h(s) dG(s))$, we can rewrite
$$
g(x)=int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)-int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)
$$
Now the first product can be bounded below by Cauchy-Schwarz
$$
int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)geq int_0^1 f(s)^{1/2}h(s)1(sleq a) dG(s)=int_0^a f(s)^{1/2}h(s) dG(s)
$$
The second product can be bounded above by noting that on $(0,1)$, $h(s)$ attains the maximum at $s=x/n$ and $f(s)$ is increasing on $(0,a)$.
$$
int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)leq f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)
$$
Putting everything together, I get
begin{eqnarray*}
g(x)&geq& int_0^a f(s)^{1/2}h(s) dG(s)-f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)\
&=&int_0^a left[f(s)^{1/2}-f(a)hleft(frac{x}{n}right)right] h(s) dG(s)
end{eqnarray*}
Now on $sin(0,a)$, $f(s)^{1/2}-f(a)hleft(frac{x}{n}right)$ may not always be greater than 0, but I need to show that the integral is (if possible) - it may also be the case that my bounds are not tight enough.
Another thought:
Since
begin{eqnarray*}
g(x)&=&int_{0}^1 f(s)h(s)dG(s)int_{0}^{a} h(s)dG(s)-int_0^1 h(s) dG(s)int_0^a f(s)h(s)dG(s)\
&=&int_0^aleft(int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)right)h(t)dG(t)
end{eqnarray*}
then I just need to show that for $tin(0,a)$,
$$
int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)geq 0
$$
or after substitution
$$
frac{t}{1-t}leq frac{int_{0}^1 left(frac{s}{1-s}right)s^x(1-s)^{n-x}dG(s)}{int_0^1 s^x(1-s)^{n-x} dG(s)}hspace{3mm}forall tin(0,a)
$$
The LHS is just a particular value of $f(t)=frac{t}{1-t}$ on a constrained range $(0,a)$, while the RHS seems to be a weighted average of $f(t)$ over a larger range (i.e. $(0,1)$). Does this inequality always hold? Does it depend on the prior $G$?
calculus probability statistics
$endgroup$
add a comment |
$begingroup$
Problem
This is a follow-up of this question.
Let $G$ be a proper prior distribution for $thetain[0,1]$ and $ain(0,1)$ be a fixed value. I'd like to show that the following quantity is non-negative for any $a$ and for any integer $0leq xleq n-1$,
$$
g(x)=int_{a}^1 s^{x+1}(1-s)^{n-x-1}dG(s)int_{0}^{a} t^{x}(1-t)^{n-x}dG(t)-int_{a}^1 s^{x}(1-s)^{n-x}dG(s)int_{0}^{a} t^{x+1}(1-t)^{n-x-1}dG(t)
$$
I've simulated that this is true when $G$ is uniform or some beta distributions. However, I have some trouble showing this algebraically. I can lower-bound the first product of integrals using Cauchy-Schwarz and upper-bound terms in the second product of integrals by the product of expectations over $G$, noting that $s$ and $1-s$ are negatively correlated. However, none of these give me useful results for showing $g(x)geq 0$.
Here's what I have so far:
First let $f(s)=frac{s}{1-s}$ and $h(s)=s^x (1-s)^{n-x}$. Then by expressing $int_a^1 f(s)h(s)dG(s)=int_0^1 f(s)h(s)dG(s)-int_a^1 f(s)h(s)dG(s)$ (and similarly for $int_a^1 h(s) dG(s))$, we can rewrite
$$
g(x)=int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)-int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)
$$
Now the first product can be bounded below by Cauchy-Schwarz
$$
int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)geq int_0^1 f(s)^{1/2}h(s)1(sleq a) dG(s)=int_0^a f(s)^{1/2}h(s) dG(s)
$$
The second product can be bounded above by noting that on $(0,1)$, $h(s)$ attains the maximum at $s=x/n$ and $f(s)$ is increasing on $(0,a)$.
$$
int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)leq f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)
$$
Putting everything together, I get
begin{eqnarray*}
g(x)&geq& int_0^a f(s)^{1/2}h(s) dG(s)-f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)\
&=&int_0^a left[f(s)^{1/2}-f(a)hleft(frac{x}{n}right)right] h(s) dG(s)
end{eqnarray*}
Now on $sin(0,a)$, $f(s)^{1/2}-f(a)hleft(frac{x}{n}right)$ may not always be greater than 0, but I need to show that the integral is (if possible) - it may also be the case that my bounds are not tight enough.
Another thought:
Since
begin{eqnarray*}
g(x)&=&int_{0}^1 f(s)h(s)dG(s)int_{0}^{a} h(s)dG(s)-int_0^1 h(s) dG(s)int_0^a f(s)h(s)dG(s)\
&=&int_0^aleft(int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)right)h(t)dG(t)
end{eqnarray*}
then I just need to show that for $tin(0,a)$,
$$
int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)geq 0
$$
or after substitution
$$
frac{t}{1-t}leq frac{int_{0}^1 left(frac{s}{1-s}right)s^x(1-s)^{n-x}dG(s)}{int_0^1 s^x(1-s)^{n-x} dG(s)}hspace{3mm}forall tin(0,a)
$$
The LHS is just a particular value of $f(t)=frac{t}{1-t}$ on a constrained range $(0,a)$, while the RHS seems to be a weighted average of $f(t)$ over a larger range (i.e. $(0,1)$). Does this inequality always hold? Does it depend on the prior $G$?
calculus probability statistics
$endgroup$
add a comment |
$begingroup$
Problem
This is a follow-up of this question.
Let $G$ be a proper prior distribution for $thetain[0,1]$ and $ain(0,1)$ be a fixed value. I'd like to show that the following quantity is non-negative for any $a$ and for any integer $0leq xleq n-1$,
$$
g(x)=int_{a}^1 s^{x+1}(1-s)^{n-x-1}dG(s)int_{0}^{a} t^{x}(1-t)^{n-x}dG(t)-int_{a}^1 s^{x}(1-s)^{n-x}dG(s)int_{0}^{a} t^{x+1}(1-t)^{n-x-1}dG(t)
$$
I've simulated that this is true when $G$ is uniform or some beta distributions. However, I have some trouble showing this algebraically. I can lower-bound the first product of integrals using Cauchy-Schwarz and upper-bound terms in the second product of integrals by the product of expectations over $G$, noting that $s$ and $1-s$ are negatively correlated. However, none of these give me useful results for showing $g(x)geq 0$.
Here's what I have so far:
First let $f(s)=frac{s}{1-s}$ and $h(s)=s^x (1-s)^{n-x}$. Then by expressing $int_a^1 f(s)h(s)dG(s)=int_0^1 f(s)h(s)dG(s)-int_a^1 f(s)h(s)dG(s)$ (and similarly for $int_a^1 h(s) dG(s))$, we can rewrite
$$
g(x)=int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)-int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)
$$
Now the first product can be bounded below by Cauchy-Schwarz
$$
int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)geq int_0^1 f(s)^{1/2}h(s)1(sleq a) dG(s)=int_0^a f(s)^{1/2}h(s) dG(s)
$$
The second product can be bounded above by noting that on $(0,1)$, $h(s)$ attains the maximum at $s=x/n$ and $f(s)$ is increasing on $(0,a)$.
$$
int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)leq f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)
$$
Putting everything together, I get
begin{eqnarray*}
g(x)&geq& int_0^a f(s)^{1/2}h(s) dG(s)-f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)\
&=&int_0^a left[f(s)^{1/2}-f(a)hleft(frac{x}{n}right)right] h(s) dG(s)
end{eqnarray*}
Now on $sin(0,a)$, $f(s)^{1/2}-f(a)hleft(frac{x}{n}right)$ may not always be greater than 0, but I need to show that the integral is (if possible) - it may also be the case that my bounds are not tight enough.
Another thought:
Since
begin{eqnarray*}
g(x)&=&int_{0}^1 f(s)h(s)dG(s)int_{0}^{a} h(s)dG(s)-int_0^1 h(s) dG(s)int_0^a f(s)h(s)dG(s)\
&=&int_0^aleft(int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)right)h(t)dG(t)
end{eqnarray*}
then I just need to show that for $tin(0,a)$,
$$
int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)geq 0
$$
or after substitution
$$
frac{t}{1-t}leq frac{int_{0}^1 left(frac{s}{1-s}right)s^x(1-s)^{n-x}dG(s)}{int_0^1 s^x(1-s)^{n-x} dG(s)}hspace{3mm}forall tin(0,a)
$$
The LHS is just a particular value of $f(t)=frac{t}{1-t}$ on a constrained range $(0,a)$, while the RHS seems to be a weighted average of $f(t)$ over a larger range (i.e. $(0,1)$). Does this inequality always hold? Does it depend on the prior $G$?
calculus probability statistics
$endgroup$
Problem
This is a follow-up of this question.
Let $G$ be a proper prior distribution for $thetain[0,1]$ and $ain(0,1)$ be a fixed value. I'd like to show that the following quantity is non-negative for any $a$ and for any integer $0leq xleq n-1$,
$$
g(x)=int_{a}^1 s^{x+1}(1-s)^{n-x-1}dG(s)int_{0}^{a} t^{x}(1-t)^{n-x}dG(t)-int_{a}^1 s^{x}(1-s)^{n-x}dG(s)int_{0}^{a} t^{x+1}(1-t)^{n-x-1}dG(t)
$$
I've simulated that this is true when $G$ is uniform or some beta distributions. However, I have some trouble showing this algebraically. I can lower-bound the first product of integrals using Cauchy-Schwarz and upper-bound terms in the second product of integrals by the product of expectations over $G$, noting that $s$ and $1-s$ are negatively correlated. However, none of these give me useful results for showing $g(x)geq 0$.
Here's what I have so far:
First let $f(s)=frac{s}{1-s}$ and $h(s)=s^x (1-s)^{n-x}$. Then by expressing $int_a^1 f(s)h(s)dG(s)=int_0^1 f(s)h(s)dG(s)-int_a^1 f(s)h(s)dG(s)$ (and similarly for $int_a^1 h(s) dG(s))$, we can rewrite
$$
g(x)=int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)-int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)
$$
Now the first product can be bounded below by Cauchy-Schwarz
$$
int_{0}^1 f(s)h(s)dG(s)int_{0}^{1} h(s)1(sleq a)dG(s)geq int_0^1 f(s)^{1/2}h(s)1(sleq a) dG(s)=int_0^a f(s)^{1/2}h(s) dG(s)
$$
The second product can be bounded above by noting that on $(0,1)$, $h(s)$ attains the maximum at $s=x/n$ and $f(s)$ is increasing on $(0,a)$.
$$
int_0^1 h(s) dG(s)int_0^1 f(s)h(s)1(sleq a)dG(s)leq f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)
$$
Putting everything together, I get
begin{eqnarray*}
g(x)&geq& int_0^a f(s)^{1/2}h(s) dG(s)-f(a)hleft(frac{x}{n}right)int_0^a h(s) dG(s)\
&=&int_0^a left[f(s)^{1/2}-f(a)hleft(frac{x}{n}right)right] h(s) dG(s)
end{eqnarray*}
Now on $sin(0,a)$, $f(s)^{1/2}-f(a)hleft(frac{x}{n}right)$ may not always be greater than 0, but I need to show that the integral is (if possible) - it may also be the case that my bounds are not tight enough.
Another thought:
Since
begin{eqnarray*}
g(x)&=&int_{0}^1 f(s)h(s)dG(s)int_{0}^{a} h(s)dG(s)-int_0^1 h(s) dG(s)int_0^a f(s)h(s)dG(s)\
&=&int_0^aleft(int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)right)h(t)dG(t)
end{eqnarray*}
then I just need to show that for $tin(0,a)$,
$$
int_{0}^1 f(s)h(s)dG(s)- f(t)int_0^1 h(s) dG(s)geq 0
$$
or after substitution
$$
frac{t}{1-t}leq frac{int_{0}^1 left(frac{s}{1-s}right)s^x(1-s)^{n-x}dG(s)}{int_0^1 s^x(1-s)^{n-x} dG(s)}hspace{3mm}forall tin(0,a)
$$
The LHS is just a particular value of $f(t)=frac{t}{1-t}$ on a constrained range $(0,a)$, while the RHS seems to be a weighted average of $f(t)$ over a larger range (i.e. $(0,1)$). Does this inequality always hold? Does it depend on the prior $G$?
calculus probability statistics
calculus probability statistics
edited Dec 4 '17 at 5:55
stats134711
asked Dec 3 '17 at 5:46
stats134711stats134711
187215
187215
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
The answer seems to be easier than initially thought:
begin{eqnarray*}
g(x)&=&int_{a}^1 f(s)h(s)dG(s)int_{0}^{a} h(t)dG(t)-int_a^1 h(s) dG(s)int_0^a f(t)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 f(s)h(s)dG(s)- f(t)int_a^1 h(s) dG(s)right)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 [f(s)-f(t)]h(s)dG(s)right)h(t)dG(t)
end{eqnarray*}
Since $f(cdot)$ is continuous and increasing on $(0,1)$, $f(s)geq f(t)$ since $sin(a,1)$ and $tin(0,a)$. Therefore, $g(x)geq 0$ as desired.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2548483%2fintegral-difference-is-non-negative%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
The answer seems to be easier than initially thought:
begin{eqnarray*}
g(x)&=&int_{a}^1 f(s)h(s)dG(s)int_{0}^{a} h(t)dG(t)-int_a^1 h(s) dG(s)int_0^a f(t)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 f(s)h(s)dG(s)- f(t)int_a^1 h(s) dG(s)right)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 [f(s)-f(t)]h(s)dG(s)right)h(t)dG(t)
end{eqnarray*}
Since $f(cdot)$ is continuous and increasing on $(0,1)$, $f(s)geq f(t)$ since $sin(a,1)$ and $tin(0,a)$. Therefore, $g(x)geq 0$ as desired.
$endgroup$
add a comment |
$begingroup$
The answer seems to be easier than initially thought:
begin{eqnarray*}
g(x)&=&int_{a}^1 f(s)h(s)dG(s)int_{0}^{a} h(t)dG(t)-int_a^1 h(s) dG(s)int_0^a f(t)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 f(s)h(s)dG(s)- f(t)int_a^1 h(s) dG(s)right)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 [f(s)-f(t)]h(s)dG(s)right)h(t)dG(t)
end{eqnarray*}
Since $f(cdot)$ is continuous and increasing on $(0,1)$, $f(s)geq f(t)$ since $sin(a,1)$ and $tin(0,a)$. Therefore, $g(x)geq 0$ as desired.
$endgroup$
add a comment |
$begingroup$
The answer seems to be easier than initially thought:
begin{eqnarray*}
g(x)&=&int_{a}^1 f(s)h(s)dG(s)int_{0}^{a} h(t)dG(t)-int_a^1 h(s) dG(s)int_0^a f(t)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 f(s)h(s)dG(s)- f(t)int_a^1 h(s) dG(s)right)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 [f(s)-f(t)]h(s)dG(s)right)h(t)dG(t)
end{eqnarray*}
Since $f(cdot)$ is continuous and increasing on $(0,1)$, $f(s)geq f(t)$ since $sin(a,1)$ and $tin(0,a)$. Therefore, $g(x)geq 0$ as desired.
$endgroup$
The answer seems to be easier than initially thought:
begin{eqnarray*}
g(x)&=&int_{a}^1 f(s)h(s)dG(s)int_{0}^{a} h(t)dG(t)-int_a^1 h(s) dG(s)int_0^a f(t)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 f(s)h(s)dG(s)- f(t)int_a^1 h(s) dG(s)right)h(t)dG(t)\
&=&int_0^aleft(int_{a}^1 [f(s)-f(t)]h(s)dG(s)right)h(t)dG(t)
end{eqnarray*}
Since $f(cdot)$ is continuous and increasing on $(0,1)$, $f(s)geq f(t)$ since $sin(a,1)$ and $tin(0,a)$. Therefore, $g(x)geq 0$ as desired.
edited Dec 12 '18 at 18:40
answered Dec 4 '17 at 11:45
stats134711stats134711
187215
187215
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2548483%2fintegral-difference-is-non-negative%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown