Recursively updating the MLE as new observations stream inSimple MLE Question4 cases of Maximum Likelihood Estimation of Gaussian distribution parameterssimulating random samples with a given MLEFor the family of distributions, $f_theta(x) = theta x^theta-1$, what is the sufficient statistic corresponding to the monotone likelihood ratio?Prove that MLE does not depend on the dominating measureDetermining an MLEMLE of $f(xmidtheta) = theta x^theta−1e^−x^thetaI_(0,infty)(x)$Sufficient statistic when $Xsim U(theta,2 theta)$Estimating the MLE where the parameter is also the constraintTrouble with MLE

How are passwords stolen from companies if they only store hashes?

Asserting that Atheism and Theism are both faith based positions

Unfrosted light bulb

Do native speakers use "ultima" and "proxima" frequently in spoken English?

Why is indicated airspeed rather than ground speed used during the takeoff roll?

pipe commands inside find -exec?

Print a physical multiplication table

Have any astronauts/cosmonauts died in space?

Can "few" be used as a subject? If so, what is the rule?

CLI: Get information Ubuntu releases

How to find the largest number(s) in a list of elements?

Help with identifying unique aircraft over NE Pennsylvania

Weird lines in Microsoft Word

Why is this tree refusing to shed its dead leaves?

Why is "la Gestapo" feminine?

Recursively updating the MLE as new observations stream in

How do researchers send unsolicited emails asking for feedback on their works?

Is a square zero matrix positive semidefinite?

What will the Frenchman say?

Difficulty understanding group delay concept

What is the tangent at a sharp point on a curve?

Imaginary part of expression too difficult to calculate

Would mining huge amounts of resources on the Moon change its orbit?

"Marked down as someone wanting to sell shares." What does that mean?



Recursively updating the MLE as new observations stream in


Simple MLE Question4 cases of Maximum Likelihood Estimation of Gaussian distribution parameterssimulating random samples with a given MLEFor the family of distributions, $f_theta(x) = theta x^theta-1$, what is the sufficient statistic corresponding to the monotone likelihood ratio?Prove that MLE does not depend on the dominating measureDetermining an MLEMLE of $f(xmidtheta) = theta x^theta−1e^−x^thetaI_(0,infty)(x)$Sufficient statistic when $Xsim U(theta,2 theta)$Estimating the MLE where the parameter is also the constraintTrouble with MLE













3












$begingroup$


General Question



Say we have iid data $x_1$, $x_2$, ... $sim f(x,|,boldsymboltheta)$ streaming in. We want to recursively compute the maximum likelihood estimate of $boldsymboltheta$. That is, having computed
$$hatboldsymboltheta_n-1=undersetboldsymbolthetainmathbbR^pargmaxprod_i=1^n-1f(x_i,|,boldsymboltheta),$$
we observe a new $x_n$, and wish to somehow incrementally update our estimate
$$hatboldsymboltheta_n-1,,x_n to hatboldsymboltheta_n$$
without having to start from scratch. Are there generic algorithms for this?



Toy Example



If $x_1$, $x_2$, ... $sim N(x,|,mu, 1)$, then
$$hatmu_n-1 = frac1n-1sumlimits_i=1^n-1x_iquadtextandquadhatmu_n = frac1nsumlimits_i=1^nx_i,$$
so
$$hatmu_n=frac1nleft[(n-1)hatmu_n-1 + x_nright].$$










share|cite|improve this question











$endgroup$











  • $begingroup$
    Awesome question!
    $endgroup$
    – dlnB
    2 hours ago















3












$begingroup$


General Question



Say we have iid data $x_1$, $x_2$, ... $sim f(x,|,boldsymboltheta)$ streaming in. We want to recursively compute the maximum likelihood estimate of $boldsymboltheta$. That is, having computed
$$hatboldsymboltheta_n-1=undersetboldsymbolthetainmathbbR^pargmaxprod_i=1^n-1f(x_i,|,boldsymboltheta),$$
we observe a new $x_n$, and wish to somehow incrementally update our estimate
$$hatboldsymboltheta_n-1,,x_n to hatboldsymboltheta_n$$
without having to start from scratch. Are there generic algorithms for this?



Toy Example



If $x_1$, $x_2$, ... $sim N(x,|,mu, 1)$, then
$$hatmu_n-1 = frac1n-1sumlimits_i=1^n-1x_iquadtextandquadhatmu_n = frac1nsumlimits_i=1^nx_i,$$
so
$$hatmu_n=frac1nleft[(n-1)hatmu_n-1 + x_nright].$$










share|cite|improve this question











$endgroup$











  • $begingroup$
    Awesome question!
    $endgroup$
    – dlnB
    2 hours ago













3












3








3


2



$begingroup$


General Question



Say we have iid data $x_1$, $x_2$, ... $sim f(x,|,boldsymboltheta)$ streaming in. We want to recursively compute the maximum likelihood estimate of $boldsymboltheta$. That is, having computed
$$hatboldsymboltheta_n-1=undersetboldsymbolthetainmathbbR^pargmaxprod_i=1^n-1f(x_i,|,boldsymboltheta),$$
we observe a new $x_n$, and wish to somehow incrementally update our estimate
$$hatboldsymboltheta_n-1,,x_n to hatboldsymboltheta_n$$
without having to start from scratch. Are there generic algorithms for this?



Toy Example



If $x_1$, $x_2$, ... $sim N(x,|,mu, 1)$, then
$$hatmu_n-1 = frac1n-1sumlimits_i=1^n-1x_iquadtextandquadhatmu_n = frac1nsumlimits_i=1^nx_i,$$
so
$$hatmu_n=frac1nleft[(n-1)hatmu_n-1 + x_nright].$$










share|cite|improve this question











$endgroup$




General Question



Say we have iid data $x_1$, $x_2$, ... $sim f(x,|,boldsymboltheta)$ streaming in. We want to recursively compute the maximum likelihood estimate of $boldsymboltheta$. That is, having computed
$$hatboldsymboltheta_n-1=undersetboldsymbolthetainmathbbR^pargmaxprod_i=1^n-1f(x_i,|,boldsymboltheta),$$
we observe a new $x_n$, and wish to somehow incrementally update our estimate
$$hatboldsymboltheta_n-1,,x_n to hatboldsymboltheta_n$$
without having to start from scratch. Are there generic algorithms for this?



Toy Example



If $x_1$, $x_2$, ... $sim N(x,|,mu, 1)$, then
$$hatmu_n-1 = frac1n-1sumlimits_i=1^n-1x_iquadtextandquadhatmu_n = frac1nsumlimits_i=1^nx_i,$$
so
$$hatmu_n=frac1nleft[(n-1)hatmu_n-1 + x_nright].$$







maximum-likelihood online






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited 1 hour ago







bamts

















asked 2 hours ago









bamtsbamts

753312




753312











  • $begingroup$
    Awesome question!
    $endgroup$
    – dlnB
    2 hours ago
















  • $begingroup$
    Awesome question!
    $endgroup$
    – dlnB
    2 hours ago















$begingroup$
Awesome question!
$endgroup$
– dlnB
2 hours ago




$begingroup$
Awesome question!
$endgroup$
– dlnB
2 hours ago










2 Answers
2






active

oldest

votes


















4












$begingroup$

See the concept of sufficiency and in particular, minimal sufficient statistics. In many cases you need the whole sample to compute the estimate at a given sample size, with no trivial way to update from a sample one size smaller (i.e. there's no convenient general result).



If the distribution is exponential family (and in some other cases besides; the uniform is a neat example) there's a nice sufficient statistic that can in many cases be updated in the manner you seek (i.e. with a number of commonly used distributions there would be a fast update).



One example I'm not aware of any direct way to either calculate or update is the estimate for the location of the Cauchy distribution (e.g. with unit scale, to make the problem a simple one-parameter problem). There may be a faster update, however, that I simply haven't noticed - I can't say I've really done more than glance at it for considering the updating case.



On the other hand, with MLEs that are obtained via numerical optimization methods, the previous estimate would in many cases be a great starting point, since typically the previous estimate would be very close to the updated estimate; in that sense at least, rapid updating should often be possible. Even this isn't the general case, though -- with multimodal likelihood functions (again, see the Cauchy for an example), a new observation might lead to the highest mode being some distance from the previous one (even if the locations of each of the biggest few modes didn't shift much, which one is highest could well change).






share|cite|improve this answer











$endgroup$




















    0












    $begingroup$

    In machine learning, this is referred to as online learning.



    As @Glen_b pointed out, there are special cases in which the MLE can be updated without needing to access all the previous data. As he also points out, I don't believe there's a generic solution for finding the MLE.



    A fairly generic approach for finding the approximate solution is to use something like stochastic gradient descent. In this case, as each observation comes in, we compute the gradient with respect to this individual observation and move the parameter values a very small amount in this direction. Under certain conditions, we can show that this will converge to a neighborhood of the MLE with high probability; the neighborhood is tighter and tighter as we reduce the step size, but more data is required for convergence. However, these stochastic methods in general require much more fiddling to obtain good performance than, say, closed form updates.






    share|cite









    $endgroup$












      Your Answer





      StackExchange.ifUsing("editor", function ()
      return StackExchange.using("mathjaxEditing", function ()
      StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix)
      StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
      );
      );
      , "mathjax-editing");

      StackExchange.ready(function()
      var channelOptions =
      tags: "".split(" "),
      id: "65"
      ;
      initTagRenderer("".split(" "), "".split(" "), channelOptions);

      StackExchange.using("externalEditor", function()
      // Have to fire editor after snippets, if snippets enabled
      if (StackExchange.settings.snippets.snippetsEnabled)
      StackExchange.using("snippets", function()
      createEditor();
      );

      else
      createEditor();

      );

      function createEditor()
      StackExchange.prepareEditor(
      heartbeatType: 'answer',
      autoActivateHeartbeat: false,
      convertImagesToLinks: false,
      noModals: true,
      showLowRepImageUploadWarning: true,
      reputationToPostImages: null,
      bindNavPrevention: true,
      postfix: "",
      imageUploader:
      brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
      contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
      allowUrls: true
      ,
      onDemand: true,
      discardSelector: ".discard-answer"
      ,immediatelyShowMarkdownHelp:true
      );



      );













      draft saved

      draft discarded


















      StackExchange.ready(
      function ()
      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f398220%2frecursively-updating-the-mle-as-new-observations-stream-in%23new-answer', 'question_page');

      );

      Post as a guest















      Required, but never shown

























      2 Answers
      2






      active

      oldest

      votes








      2 Answers
      2






      active

      oldest

      votes









      active

      oldest

      votes






      active

      oldest

      votes









      4












      $begingroup$

      See the concept of sufficiency and in particular, minimal sufficient statistics. In many cases you need the whole sample to compute the estimate at a given sample size, with no trivial way to update from a sample one size smaller (i.e. there's no convenient general result).



      If the distribution is exponential family (and in some other cases besides; the uniform is a neat example) there's a nice sufficient statistic that can in many cases be updated in the manner you seek (i.e. with a number of commonly used distributions there would be a fast update).



      One example I'm not aware of any direct way to either calculate or update is the estimate for the location of the Cauchy distribution (e.g. with unit scale, to make the problem a simple one-parameter problem). There may be a faster update, however, that I simply haven't noticed - I can't say I've really done more than glance at it for considering the updating case.



      On the other hand, with MLEs that are obtained via numerical optimization methods, the previous estimate would in many cases be a great starting point, since typically the previous estimate would be very close to the updated estimate; in that sense at least, rapid updating should often be possible. Even this isn't the general case, though -- with multimodal likelihood functions (again, see the Cauchy for an example), a new observation might lead to the highest mode being some distance from the previous one (even if the locations of each of the biggest few modes didn't shift much, which one is highest could well change).






      share|cite|improve this answer











      $endgroup$

















        4












        $begingroup$

        See the concept of sufficiency and in particular, minimal sufficient statistics. In many cases you need the whole sample to compute the estimate at a given sample size, with no trivial way to update from a sample one size smaller (i.e. there's no convenient general result).



        If the distribution is exponential family (and in some other cases besides; the uniform is a neat example) there's a nice sufficient statistic that can in many cases be updated in the manner you seek (i.e. with a number of commonly used distributions there would be a fast update).



        One example I'm not aware of any direct way to either calculate or update is the estimate for the location of the Cauchy distribution (e.g. with unit scale, to make the problem a simple one-parameter problem). There may be a faster update, however, that I simply haven't noticed - I can't say I've really done more than glance at it for considering the updating case.



        On the other hand, with MLEs that are obtained via numerical optimization methods, the previous estimate would in many cases be a great starting point, since typically the previous estimate would be very close to the updated estimate; in that sense at least, rapid updating should often be possible. Even this isn't the general case, though -- with multimodal likelihood functions (again, see the Cauchy for an example), a new observation might lead to the highest mode being some distance from the previous one (even if the locations of each of the biggest few modes didn't shift much, which one is highest could well change).






        share|cite|improve this answer











        $endgroup$















          4












          4








          4





          $begingroup$

          See the concept of sufficiency and in particular, minimal sufficient statistics. In many cases you need the whole sample to compute the estimate at a given sample size, with no trivial way to update from a sample one size smaller (i.e. there's no convenient general result).



          If the distribution is exponential family (and in some other cases besides; the uniform is a neat example) there's a nice sufficient statistic that can in many cases be updated in the manner you seek (i.e. with a number of commonly used distributions there would be a fast update).



          One example I'm not aware of any direct way to either calculate or update is the estimate for the location of the Cauchy distribution (e.g. with unit scale, to make the problem a simple one-parameter problem). There may be a faster update, however, that I simply haven't noticed - I can't say I've really done more than glance at it for considering the updating case.



          On the other hand, with MLEs that are obtained via numerical optimization methods, the previous estimate would in many cases be a great starting point, since typically the previous estimate would be very close to the updated estimate; in that sense at least, rapid updating should often be possible. Even this isn't the general case, though -- with multimodal likelihood functions (again, see the Cauchy for an example), a new observation might lead to the highest mode being some distance from the previous one (even if the locations of each of the biggest few modes didn't shift much, which one is highest could well change).






          share|cite|improve this answer











          $endgroup$



          See the concept of sufficiency and in particular, minimal sufficient statistics. In many cases you need the whole sample to compute the estimate at a given sample size, with no trivial way to update from a sample one size smaller (i.e. there's no convenient general result).



          If the distribution is exponential family (and in some other cases besides; the uniform is a neat example) there's a nice sufficient statistic that can in many cases be updated in the manner you seek (i.e. with a number of commonly used distributions there would be a fast update).



          One example I'm not aware of any direct way to either calculate or update is the estimate for the location of the Cauchy distribution (e.g. with unit scale, to make the problem a simple one-parameter problem). There may be a faster update, however, that I simply haven't noticed - I can't say I've really done more than glance at it for considering the updating case.



          On the other hand, with MLEs that are obtained via numerical optimization methods, the previous estimate would in many cases be a great starting point, since typically the previous estimate would be very close to the updated estimate; in that sense at least, rapid updating should often be possible. Even this isn't the general case, though -- with multimodal likelihood functions (again, see the Cauchy for an example), a new observation might lead to the highest mode being some distance from the previous one (even if the locations of each of the biggest few modes didn't shift much, which one is highest could well change).







          share|cite|improve this answer














          share|cite|improve this answer



          share|cite|improve this answer








          edited 50 mins ago

























          answered 1 hour ago









          Glen_bGlen_b

          214k22414764




          214k22414764























              0












              $begingroup$

              In machine learning, this is referred to as online learning.



              As @Glen_b pointed out, there are special cases in which the MLE can be updated without needing to access all the previous data. As he also points out, I don't believe there's a generic solution for finding the MLE.



              A fairly generic approach for finding the approximate solution is to use something like stochastic gradient descent. In this case, as each observation comes in, we compute the gradient with respect to this individual observation and move the parameter values a very small amount in this direction. Under certain conditions, we can show that this will converge to a neighborhood of the MLE with high probability; the neighborhood is tighter and tighter as we reduce the step size, but more data is required for convergence. However, these stochastic methods in general require much more fiddling to obtain good performance than, say, closed form updates.






              share|cite









              $endgroup$

















                0












                $begingroup$

                In machine learning, this is referred to as online learning.



                As @Glen_b pointed out, there are special cases in which the MLE can be updated without needing to access all the previous data. As he also points out, I don't believe there's a generic solution for finding the MLE.



                A fairly generic approach for finding the approximate solution is to use something like stochastic gradient descent. In this case, as each observation comes in, we compute the gradient with respect to this individual observation and move the parameter values a very small amount in this direction. Under certain conditions, we can show that this will converge to a neighborhood of the MLE with high probability; the neighborhood is tighter and tighter as we reduce the step size, but more data is required for convergence. However, these stochastic methods in general require much more fiddling to obtain good performance than, say, closed form updates.






                share|cite









                $endgroup$















                  0












                  0








                  0





                  $begingroup$

                  In machine learning, this is referred to as online learning.



                  As @Glen_b pointed out, there are special cases in which the MLE can be updated without needing to access all the previous data. As he also points out, I don't believe there's a generic solution for finding the MLE.



                  A fairly generic approach for finding the approximate solution is to use something like stochastic gradient descent. In this case, as each observation comes in, we compute the gradient with respect to this individual observation and move the parameter values a very small amount in this direction. Under certain conditions, we can show that this will converge to a neighborhood of the MLE with high probability; the neighborhood is tighter and tighter as we reduce the step size, but more data is required for convergence. However, these stochastic methods in general require much more fiddling to obtain good performance than, say, closed form updates.






                  share|cite









                  $endgroup$



                  In machine learning, this is referred to as online learning.



                  As @Glen_b pointed out, there are special cases in which the MLE can be updated without needing to access all the previous data. As he also points out, I don't believe there's a generic solution for finding the MLE.



                  A fairly generic approach for finding the approximate solution is to use something like stochastic gradient descent. In this case, as each observation comes in, we compute the gradient with respect to this individual observation and move the parameter values a very small amount in this direction. Under certain conditions, we can show that this will converge to a neighborhood of the MLE with high probability; the neighborhood is tighter and tighter as we reduce the step size, but more data is required for convergence. However, these stochastic methods in general require much more fiddling to obtain good performance than, say, closed form updates.







                  share|cite












                  share|cite



                  share|cite










                  answered 5 mins ago









                  Cliff ABCliff AB

                  13.6k12567




                  13.6k12567



























                      draft saved

                      draft discarded
















































                      Thanks for contributing an answer to Cross Validated!


                      • Please be sure to answer the question. Provide details and share your research!

                      But avoid


                      • Asking for help, clarification, or responding to other answers.

                      • Making statements based on opinion; back them up with references or personal experience.

                      Use MathJax to format equations. MathJax reference.


                      To learn more, see our tips on writing great answers.




                      draft saved


                      draft discarded














                      StackExchange.ready(
                      function ()
                      StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f398220%2frecursively-updating-the-mle-as-new-observations-stream-in%23new-answer', 'question_page');

                      );

                      Post as a guest















                      Required, but never shown





















































                      Required, but never shown














                      Required, but never shown












                      Required, but never shown







                      Required, but never shown

































                      Required, but never shown














                      Required, but never shown












                      Required, but never shown







                      Required, but never shown







                      Popular posts from this blog

                      How should I use the fbox command correctly to avoid producing a Bad Box message?How to put a long piece of text in a box?How to specify height and width of fboxIs there an arrayrulecolor-like command to change the rule color of fbox?What is the command to highlight bad boxes in pdf?Why does fbox sometimes place the box *over* the graphic image?how to put the text in the boxHow to create command for a box where text inside the box can automatically adjust?how can I make an fbox like command with certain color, shape and width of border?how to use fbox in align modeFbox increase the spacing between the box and it content (inner margin)how to change the box height of an equationWhat is the use of the hbox in a newcommand command?

                      Doxepinum Nexus interni Notae | Tabula navigationis3158DB01142WHOa682390"Structural Analysis of the Histamine H1 Receptor""Transdermal and Topical Drug Administration in the Treatment of Pain""Antidepressants as antipruritic agents: A review"

                      inputenc: Unicode character … not set up for use with LaTeX The Next CEO of Stack OverflowEntering Unicode characters in LaTeXHow to solve the `Package inputenc Error: Unicode char not set up for use with LaTeX` problem?solve “Unicode char is not set up for use with LaTeX” without special handling of every new interesting UTF-8 characterPackage inputenc Error: Unicode character ² (U+B2)(inputenc) not set up for use with LaTeX. acroI2C[I²C]package inputenc error unicode char (u + 190) not set up for use with latexPackage inputenc Error: Unicode char u8:′ not set up for use with LaTeX. 3′inputenc Error: Unicode char u8: not set up for use with LaTeX with G-BriefPackage Inputenc Error: Unicode char u8: not set up for use with LaTeXPackage inputenc Error: Unicode char ́ (U+301)(inputenc) not set up for use with LaTeX. includePackage inputenc Error: Unicode char ̂ (U+302)(inputenc) not set up for use with LaTeX. … $widehatleft (OA,AA' right )$Package inputenc Error: Unicode char â„¡ (U+2121)(inputenc) not set up for use with LaTeX. printbibliography[heading=bibintoc]Package inputenc Error: Unicode char − (U+2212)(inputenc) not set up for use with LaTeXPackage inputenc Error: Unicode character α (U+3B1) not set up for use with LaTeXPackage inputenc Error: Unicode characterError: ! Package inputenc Error: Unicode char ⊘ (U+2298)(inputenc) not set up for use with LaTeX