Point $x in mathbb{R}^n$ that minimizes sum of distance squares $sum_{mathcal{l}=1}^{k} Vert...
Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.
How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$
I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.
So to get the center of a set of points
$$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
we can get their centroid by
$$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$
I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.
analysis norm maxima-minima average
add a comment |
Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.
How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$
I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.
So to get the center of a set of points
$$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
we can get their centroid by
$$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$
I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.
analysis norm maxima-minima average
add a comment |
Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.
How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$
I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.
So to get the center of a set of points
$$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
we can get their centroid by
$$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$
I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.
analysis norm maxima-minima average
Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.
How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$
I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.
So to get the center of a set of points
$$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
we can get their centroid by
$$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$
I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.
analysis norm maxima-minima average
analysis norm maxima-minima average
asked Nov 30 at 0:05
Ramanujan Taylor
184
184
add a comment |
add a comment |
2 Answers
2
active
oldest
votes
I believe your intuition is correct.
Consider the following:
$$
sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
$$
Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.
To find the minimum of this with respect to $x$, we can use standard differentiation procedures.
First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
$$
begin{align}
frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
&=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
&=sum_{l=1}^k 2(x_j-a_j^{(l)})
end{align}
$$
Setting this expression to zero you will obtain:
$$
begin{align}
sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
end{align}
$$
This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.
Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).
Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:
$$
begin{align}
frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
&=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
&=begin{cases}
2 & i=j\
0 & ineq j
end{cases}
end{align}
$$
This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.
Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.
add a comment |
You can use this relation:
$$
sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
$$
where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$
If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
$$
x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
$$
To prove the first relation, you must notice that:
$$
sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
$$
then simply develop
$$
sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
$$
(I can write the details if you want, just ask)
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3019429%2fpoint-x-in-mathbbrn-that-minimizes-sum-of-distance-squares-sum-mathca%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
2 Answers
2
active
oldest
votes
2 Answers
2
active
oldest
votes
active
oldest
votes
active
oldest
votes
I believe your intuition is correct.
Consider the following:
$$
sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
$$
Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.
To find the minimum of this with respect to $x$, we can use standard differentiation procedures.
First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
$$
begin{align}
frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
&=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
&=sum_{l=1}^k 2(x_j-a_j^{(l)})
end{align}
$$
Setting this expression to zero you will obtain:
$$
begin{align}
sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
end{align}
$$
This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.
Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).
Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:
$$
begin{align}
frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
&=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
&=begin{cases}
2 & i=j\
0 & ineq j
end{cases}
end{align}
$$
This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.
Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.
add a comment |
I believe your intuition is correct.
Consider the following:
$$
sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
$$
Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.
To find the minimum of this with respect to $x$, we can use standard differentiation procedures.
First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
$$
begin{align}
frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
&=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
&=sum_{l=1}^k 2(x_j-a_j^{(l)})
end{align}
$$
Setting this expression to zero you will obtain:
$$
begin{align}
sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
end{align}
$$
This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.
Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).
Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:
$$
begin{align}
frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
&=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
&=begin{cases}
2 & i=j\
0 & ineq j
end{cases}
end{align}
$$
This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.
Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.
add a comment |
I believe your intuition is correct.
Consider the following:
$$
sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
$$
Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.
To find the minimum of this with respect to $x$, we can use standard differentiation procedures.
First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
$$
begin{align}
frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
&=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
&=sum_{l=1}^k 2(x_j-a_j^{(l)})
end{align}
$$
Setting this expression to zero you will obtain:
$$
begin{align}
sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
end{align}
$$
This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.
Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).
Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:
$$
begin{align}
frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
&=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
&=begin{cases}
2 & i=j\
0 & ineq j
end{cases}
end{align}
$$
This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.
Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.
I believe your intuition is correct.
Consider the following:
$$
sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
$$
Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.
To find the minimum of this with respect to $x$, we can use standard differentiation procedures.
First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
$$
begin{align}
frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
&=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
&=sum_{l=1}^k 2(x_j-a_j^{(l)})
end{align}
$$
Setting this expression to zero you will obtain:
$$
begin{align}
sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
end{align}
$$
This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.
Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).
Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:
$$
begin{align}
frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
&=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
&=begin{cases}
2 & i=j\
0 & ineq j
end{cases}
end{align}
$$
This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.
Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.
edited Nov 30 at 12:15
answered Nov 30 at 12:08
Sean Lee
1387
1387
add a comment |
add a comment |
You can use this relation:
$$
sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
$$
where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$
If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
$$
x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
$$
To prove the first relation, you must notice that:
$$
sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
$$
then simply develop
$$
sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
$$
(I can write the details if you want, just ask)
add a comment |
You can use this relation:
$$
sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
$$
where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$
If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
$$
x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
$$
To prove the first relation, you must notice that:
$$
sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
$$
then simply develop
$$
sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
$$
(I can write the details if you want, just ask)
add a comment |
You can use this relation:
$$
sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
$$
where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$
If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
$$
x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
$$
To prove the first relation, you must notice that:
$$
sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
$$
then simply develop
$$
sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
$$
(I can write the details if you want, just ask)
You can use this relation:
$$
sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
$$
where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$
If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
$$
x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
$$
To prove the first relation, you must notice that:
$$
sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
$$
then simply develop
$$
sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
$$
(I can write the details if you want, just ask)
edited Nov 30 at 14:59
answered Nov 30 at 13:29
Picaud Vincent
1,20838
1,20838
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Some of your past answers have not been well-received, and you're in danger of being blocked from answering.
Please pay close attention to the following guidance:
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3019429%2fpoint-x-in-mathbbrn-that-minimizes-sum-of-distance-squares-sum-mathca%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown