Point $x in mathbb{R}^n$ that minimizes sum of distance squares $sum_{mathcal{l}=1}^{k} Vert...












1














Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.



How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$



I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.



So to get the center of a set of points



$$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
we can get their centroid by
$$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$



I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.










share|cite|improve this question



























    1














    Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.



    How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$



    I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.



    So to get the center of a set of points



    $$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
    we can get their centroid by
    $$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$



    I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.










    share|cite|improve this question

























      1












      1








      1







      Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.



      How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$



      I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.



      So to get the center of a set of points



      $$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
      we can get their centroid by
      $$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$



      I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.










      share|cite|improve this question













      Let $a^{(1)},...,a^{(k)} in mathbb{R}^n$.



      How can one find the point $x in mathbb{R}^n$, which minimizes the sum of distance squares $sum_{mathcal{l}=1}^{k} Vert x-a^{mathcal{(l)}} Vert _2^2$



      I know that a function $h(t) = sum_{l=1}^{k}(t-x_i)^2$ has its minimum when $t = overline{x}$. So for the average $overline{x}$ of the numbers $x_1,...,x_n$ the sum of squares of the deviations $overline{x}-x_i$ is minimal.



      So to get the center of a set of points



      $$ S={(x_1,y_1),(x_2,y_2),dots (x_n,y_n)}$$
      we can get their centroid by
      $$(bar x,bar y) = left(frac{1}{n}sum_{i=0}^n x_i, frac{1}{n}sum_{i=0}^n y_iright).$$



      I don't really know if this point actually minimizes the sum of distance squares given above. I'd also like to know if only one such point exists or if there are more.







      analysis norm maxima-minima average






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Nov 30 at 0:05









      Ramanujan Taylor

      184




      184






















          2 Answers
          2






          active

          oldest

          votes


















          1














          I believe your intuition is correct.



          Consider the following:



          $$
          sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
          $$



          Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.



          To find the minimum of this with respect to $x$, we can use standard differentiation procedures.



          First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
          $$
          begin{align}
          frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
          &=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
          &=sum_{l=1}^k 2(x_j-a_j^{(l)})
          end{align}
          $$

          Setting this expression to zero you will obtain:
          $$
          begin{align}
          sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
          hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
          end{align}
          $$



          This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.



          Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).



          Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:



          $$
          begin{align}
          frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
          &=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
          &=begin{cases}
          2 & i=j\
          0 & ineq j
          end{cases}
          end{align}
          $$



          This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.



          Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.






          share|cite|improve this answer































            0














            You can use this relation:
            $$
            sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
            $$

            where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$



            If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
            $$
            x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
            $$





            To prove the first relation, you must notice that:
            $$
            sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
            $$

            then simply develop
            $$
            sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
            $$

            (I can write the details if you want, just ask)






            share|cite|improve this answer























              Your Answer





              StackExchange.ifUsing("editor", function () {
              return StackExchange.using("mathjaxEditing", function () {
              StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
              StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
              });
              });
              }, "mathjax-editing");

              StackExchange.ready(function() {
              var channelOptions = {
              tags: "".split(" "),
              id: "69"
              };
              initTagRenderer("".split(" "), "".split(" "), channelOptions);

              StackExchange.using("externalEditor", function() {
              // Have to fire editor after snippets, if snippets enabled
              if (StackExchange.settings.snippets.snippetsEnabled) {
              StackExchange.using("snippets", function() {
              createEditor();
              });
              }
              else {
              createEditor();
              }
              });

              function createEditor() {
              StackExchange.prepareEditor({
              heartbeatType: 'answer',
              autoActivateHeartbeat: false,
              convertImagesToLinks: true,
              noModals: true,
              showLowRepImageUploadWarning: true,
              reputationToPostImages: 10,
              bindNavPrevention: true,
              postfix: "",
              imageUploader: {
              brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
              contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
              allowUrls: true
              },
              noCode: true, onDemand: true,
              discardSelector: ".discard-answer"
              ,immediatelyShowMarkdownHelp:true
              });


              }
              });














              draft saved

              draft discarded


















              StackExchange.ready(
              function () {
              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3019429%2fpoint-x-in-mathbbrn-that-minimizes-sum-of-distance-squares-sum-mathca%23new-answer', 'question_page');
              }
              );

              Post as a guest















              Required, but never shown

























              2 Answers
              2






              active

              oldest

              votes








              2 Answers
              2






              active

              oldest

              votes









              active

              oldest

              votes






              active

              oldest

              votes









              1














              I believe your intuition is correct.



              Consider the following:



              $$
              sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
              $$



              Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.



              To find the minimum of this with respect to $x$, we can use standard differentiation procedures.



              First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
              $$
              begin{align}
              frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
              &=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
              &=sum_{l=1}^k 2(x_j-a_j^{(l)})
              end{align}
              $$

              Setting this expression to zero you will obtain:
              $$
              begin{align}
              sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
              hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
              end{align}
              $$



              This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.



              Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).



              Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:



              $$
              begin{align}
              frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
              &=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
              &=begin{cases}
              2 & i=j\
              0 & ineq j
              end{cases}
              end{align}
              $$



              This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.



              Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.






              share|cite|improve this answer




























                1














                I believe your intuition is correct.



                Consider the following:



                $$
                sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
                $$



                Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.



                To find the minimum of this with respect to $x$, we can use standard differentiation procedures.



                First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
                $$
                begin{align}
                frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
                &=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
                &=sum_{l=1}^k 2(x_j-a_j^{(l)})
                end{align}
                $$

                Setting this expression to zero you will obtain:
                $$
                begin{align}
                sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
                hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
                end{align}
                $$



                This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.



                Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).



                Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:



                $$
                begin{align}
                frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
                &=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
                &=begin{cases}
                2 & i=j\
                0 & ineq j
                end{cases}
                end{align}
                $$



                This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.



                Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.






                share|cite|improve this answer


























                  1












                  1








                  1






                  I believe your intuition is correct.



                  Consider the following:



                  $$
                  sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
                  $$



                  Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.



                  To find the minimum of this with respect to $x$, we can use standard differentiation procedures.



                  First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
                  $$
                  begin{align}
                  frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
                  &=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
                  &=sum_{l=1}^k 2(x_j-a_j^{(l)})
                  end{align}
                  $$

                  Setting this expression to zero you will obtain:
                  $$
                  begin{align}
                  sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
                  hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
                  end{align}
                  $$



                  This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.



                  Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).



                  Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:



                  $$
                  begin{align}
                  frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
                  &=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
                  &=begin{cases}
                  2 & i=j\
                  0 & ineq j
                  end{cases}
                  end{align}
                  $$



                  This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.



                  Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.






                  share|cite|improve this answer














                  I believe your intuition is correct.



                  Consider the following:



                  $$
                  sum_{l=1}^k || x - a^{(l)}||_2^2 = sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
                  $$



                  Where $a_m^{(l)}$ is the $m$th component of the $l$th vector, and $x_m$ is the $m$th component of $x$.



                  To find the minimum of this with respect to $x$, we can use standard differentiation procedures.



                  First, differentiate with respect to one direction (i.e. differentiate with respect to some $x_j$):
                  $$
                  begin{align}
                  frac{partial}{partial x_j} sum_{l=1}^k sum_{m=1}^n (x_i - a^{(l)}_m)^2
                  &=sum_{l=1}^k sum_{m=1}^n frac{partial}{partial x_j} (x_m - a^{(l)}_m)^2 \
                  &=sum_{l=1}^k 2(x_j-a_j^{(l)})
                  end{align}
                  $$

                  Setting this expression to zero you will obtain:
                  $$
                  begin{align}
                  sum_{l=1}^k 2(hat{x}_j-a_j^{(l)}) &= 0\
                  hat{x}_j &= frac{1}{k}sum_{l=1}^k a_j^{(l)}
                  end{align}
                  $$



                  This shows that for any direction $j$, there is a stationary point of the function $sum_{l=1}^k || x - a^{(l)}||_2^2$ (which we will show is a global minimum) given by the average of the $j$th component of the vectors ${a}_{l=1}^k$.



                  Now, to show that this is a unique global minimum, we will use the fact that a function $f(x)$ is strongly convex (which implies that $f(x)$ has a unique minimum point) if its Hessian $H$ (matrix containing the second derivatives) is positive definite (We say that $H$ is positive definite if $forall c in mathbb{R}^n/{0}, c^THc > 0$).



                  Consider the second derivatives of $sum_{l=1}^k || x - a^{(l)}||_2^2$:



                  $$
                  begin{align}
                  frac{partial^2}{partial_ipartial x_j} sum_{l=1}^k sum_{m=1}^n (x_m - a^{(l)}_m)^2
                  &=sum_{l=1}^k frac{partial}{partial x_i}2(hat{x}_j-a_j^{(l)})\
                  &=begin{cases}
                  2 & i=j\
                  0 & ineq j
                  end{cases}
                  end{align}
                  $$



                  This means that the Hessian $H$ will be diagonal, and that there will be strictly positive entries on the diagonal. Thus $H$ is positive definite, implying that $sum_{l=1}^k || x - a^{(l)}||_2^2$ is strongly convex with respect to $x$.



                  Hence, we can conclude that the stationary point which we found above is indeed a unique minimum.







                  share|cite|improve this answer














                  share|cite|improve this answer



                  share|cite|improve this answer








                  edited Nov 30 at 12:15

























                  answered Nov 30 at 12:08









                  Sean Lee

                  1387




                  1387























                      0














                      You can use this relation:
                      $$
                      sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
                      $$

                      where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$



                      If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
                      $$
                      x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
                      $$





                      To prove the first relation, you must notice that:
                      $$
                      sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
                      $$

                      then simply develop
                      $$
                      sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
                      $$

                      (I can write the details if you want, just ask)






                      share|cite|improve this answer




























                        0














                        You can use this relation:
                        $$
                        sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
                        $$

                        where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$



                        If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
                        $$
                        x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
                        $$





                        To prove the first relation, you must notice that:
                        $$
                        sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
                        $$

                        then simply develop
                        $$
                        sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
                        $$

                        (I can write the details if you want, just ask)






                        share|cite|improve this answer


























                          0












                          0








                          0






                          You can use this relation:
                          $$
                          sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
                          $$

                          where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$



                          If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
                          $$
                          x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
                          $$





                          To prove the first relation, you must notice that:
                          $$
                          sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
                          $$

                          then simply develop
                          $$
                          sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
                          $$

                          (I can write the details if you want, just ask)






                          share|cite|improve this answer














                          You can use this relation:
                          $$
                          sum_{i=1}^n |x-a_i|^2=n|bar{a}-x|^2+sum_{i=1}^n |bar{a}-a_i|^2
                          $$

                          where $bar{a}=frac{1}{n}sum_{i=1}^n a_i$



                          If you want to minimize w.r.t. $x$, just take $x=bar{a}$ to cancel the $|bar{a}-x|^2$ term. Therefore the answer is:
                          $$
                          x=bar{a}=frac{1}{n}sum_{i=1}^n a_i
                          $$





                          To prove the first relation, you must notice that:
                          $$
                          sum_i^{n}|a_i-bar{a}|^2=dots=left(sum_i^{n}|a_i|^2right)-n|bar{a}|^2
                          $$

                          then simply develop
                          $$
                          sum_{i=1}^n |x-a_i|^2=sum_{i=1}^n (|x|^2-2langle x,a_i rangle+|a_i|^2)=dots
                          $$

                          (I can write the details if you want, just ask)







                          share|cite|improve this answer














                          share|cite|improve this answer



                          share|cite|improve this answer








                          edited Nov 30 at 14:59

























                          answered Nov 30 at 13:29









                          Picaud Vincent

                          1,20838




                          1,20838






























                              draft saved

                              draft discarded




















































                              Thanks for contributing an answer to Mathematics Stack Exchange!


                              • Please be sure to answer the question. Provide details and share your research!

                              But avoid



                              • Asking for help, clarification, or responding to other answers.

                              • Making statements based on opinion; back them up with references or personal experience.


                              Use MathJax to format equations. MathJax reference.


                              To learn more, see our tips on writing great answers.





                              Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


                              Please pay close attention to the following guidance:


                              • Please be sure to answer the question. Provide details and share your research!

                              But avoid



                              • Asking for help, clarification, or responding to other answers.

                              • Making statements based on opinion; back them up with references or personal experience.


                              To learn more, see our tips on writing great answers.




                              draft saved


                              draft discarded














                              StackExchange.ready(
                              function () {
                              StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3019429%2fpoint-x-in-mathbbrn-that-minimizes-sum-of-distance-squares-sum-mathca%23new-answer', 'question_page');
                              }
                              );

                              Post as a guest















                              Required, but never shown





















































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown

































                              Required, but never shown














                              Required, but never shown












                              Required, but never shown







                              Required, but never shown







                              Popular posts from this blog

                              To store a contact into the json file from server.js file using a class in NodeJS

                              Redirect URL with Chrome Remote Debugging Android Devices

                              Dieringhausen