{"id":481,"date":"2022-12-14T21:54:21","date_gmt":"2022-12-14T21:54:21","guid":{"rendered":"http:\/\/statisticalbiophysicsblog.org\/?p=481"},"modified":"2022-12-14T21:54:24","modified_gmt":"2022-12-14T21:54:24","slug":"maximum-likelihood-vs-bayesian-estimation-of-uncertainty","status":"publish","type":"post","link":"https:\/\/statisticalbiophysicsblog.org\/?p=481","title":{"rendered":"Maximum Likelihood vs. Bayesian estimation of uncertainty"},"content":{"rendered":"<p>When we want to estimate parameters from data (e.g., from binding, kinetics, or electrophysiology experiments), there are two tasks: (i) estimate the most likely values, and (ii) equally importantly, estimate the uncertainty in those values.  After all, if the uncertainty is huge, it\u2019s hard to say we really know the parameters.  We also need to choose the model in the first place, which is an extremely important task, but that is beyond the scope of this discussion.<br \/>\n<!--more--><\/p>\n<p>While maximum likelihood (ML) estimates are clearly a sensible choice for parameter values, sometimes the ML approach is extended to provide confidence intervals, i.e., uncertainty ranges.  In this discussion, I will explain why ML can be problematic for uncertainty in parameter values.  As I have seen it applied (see Further Reading, below), ML is an approximation that is formally equivalent to a physical description of a system that ignores entropy.  To be clear, ignoring entropy in a probabilistic sense is intrinsic to ML, and has nothing to do with whether entropy is being considered among physical parameters for data analysis.  More on this below.<\/p>\n<p>I believe that using ML is not a good idea for uncertainty, even though it may work \u2018well enough\u2019 in some cases.  A key limitation is that the ML approach on its own cannot reveal when its approximation is a reasonable one.<\/p>\n<p>Before getting into the critique, I will say that the right approach is Bayesian inference (BI).  If you find BI confusing, let\u2019s make clear at the outset that BI is simply a combination of likelihood \u2013 the very same ingredient that\u2019s in ML already \u2013 and prior assumptions, which often are merely common-sense and\/or empirical limits on parameter ranges \u2026 and such limits may be in place for ML estimates too.  So don\u2019t be afraid of BI theory \u2026 it\u2019s quite similar to ML until the approximations come in.  In fact, subtleties that enter prior selection in BI actually are implicit in ML approaches \u2013 see below.<\/p>\n<p>You may know that BI calculations tend to be more computationally demanding than ML and certainly can be trickier.  But I would say that in cases where the parameter likelihood or error landscape (defined below) is simple \u2013 both ML and BI will be straightforward.  In cases where that landscape is challenging, both ML and BI will require care.  In my opinion, typical BI practices are more likely to reveal a challenging landscape.  But anyway, even in a simple landscape ML has the missing-entropy approximation problem.<\/p>\n<p><em>Into the weeds: Words and pictures<\/em><\/p>\n<p>Before writing down any equations, let\u2019s make clear the conceptual error that occurs with ML uncertainty estimation (as I have seen it practiced).  The error comes from restricting attention to the <em>subset<\/em> of maximum likelihood solutions in a multi-dimensional space instead of considering <em>all<\/em> likelihood.  In the simplest non-trivial case with two parameters <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>  and <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  (see figure), the ML estimate of uncertainty for <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>   considers <em>only<\/em> the set of <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>   values that maximize the likelihood for each value of <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/> .  That is, for each<img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/> , the ML analysis considers only a single <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-577bb293dfdc79522df12596a147779b_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;&#94;&#42;\" title=\"Rendered by QuickLaTeX.com\" height=\"17\" width=\"15\" style=\"vertical-align: -5px;\"\/>    value which maximizes likelihood for the given <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/> .  This ignores that the likelihood of any <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>   is determined not only by this maximum but by all possible <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>   values.  In more technical terms, the likelihood associated with each <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>   is determined by the well-known <em>marginal distribution,<\/em> not by what might be dubbed the \u2018silhouette\u2019 distribution based only on the maxima (see figure).  The ML approach ignores potentially significant variability that may occur in orthogonal variables (<img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  in the simple example), and such variability is expected whenever there are non-linear correlations.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" width=\"681\" height=\"752\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-1-6.png\" class=\"wp-image-511\" srcset=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-1-6.png 681w, https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-1-6-272x300.png 272w\" sizes=\"auto, (max-width: 681px) 100vw, 681px\" \/><\/p>\n<p>In statistical mechanics, the ML approximation is equivalent to generating a free energy surface that ignores entropy \u2013 i.e., is determined only by the lowest-energy configuration at each value of the coordinate(s) considered.  (See Zuckerman textbook.)  For complex systems not at low temperatures, such an approximation would not generally be reliable.<\/p>\n<p>There\u2019s no question that the ML approach just outlined is not correct and must be viewed as an approximation.  The only rigorous way to check the approximation that I can see is to consider the full likelihood distribution \u2026 which is just Bayesian inference.  In other words, it seems that we need to do BI to check if the ML approximation is ok.<\/p>\n<p><em>Into the weeds: Equations<\/em><\/p>\n<p>Let\u2019s dive into the math, which fortunately isn\u2019t too terrible.  On the one hand, you have your data, which we\u2019ll denote as <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-bb3c186e5c65fcd066bb23dec8f4e48a_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#121;&#95;&#105;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"14\" style=\"vertical-align: -4px;\"\/>  values.  On the other, you have a mathematical model with some unknown parameters <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-0cfcf53857a36b6c78bcbdfc8bb2d028_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#32;&#61;&#32;&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;&#44;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;&#44;&#32;&#92;&#108;&#100;&#111;&#116;&#115;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"114\" style=\"vertical-align: -5px;\"\/>  that are to be compared to the data: the model leads to predicted <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-365557f5f12703d4de96202aa97b83e8_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#104;&#97;&#116;&#123;&#121;&#125;&#95;&#105;\" title=\"Rendered by QuickLaTeX.com\" height=\"16\" width=\"14\" style=\"vertical-align: -4px;\"\/>  that depend on the specific set of parameters.<\/p>\n<p>The difference between the data and the model can be quantified for any given set of parameters by examining a quantity related to the sum of squared deviations, given in the simplest case by<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" width=\"370\" height=\"131\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-2-5.png\" class=\"wp-image-512\" srcset=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-2-5.png 370w, https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-2-5-300x106.png 300w\" sizes=\"auto, (max-width: 370px) 100vw, 370px\" \/><\/p>\n<p>where <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-5793832f979c2268e3694c246d53b1bb_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#78;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"16\" style=\"vertical-align: 0px;\"\/>  is the number of data points with values <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-bb3c186e5c65fcd066bb23dec8f4e48a_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#121;&#95;&#105;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"14\" style=\"vertical-align: -4px;\"\/>  and where <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-365557f5f12703d4de96202aa97b83e8_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#104;&#97;&#116;&#123;&#121;&#125;&#95;&#105;\" title=\"Rendered by QuickLaTeX.com\" height=\"16\" width=\"14\" style=\"vertical-align: -4px;\"\/>  are the values generated by the model (for any given choice of parameters: <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6566b599305d06bd59ff610d596d22ce_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#104;&#97;&#116;&#123;&#121;&#125;&#95;&#105;&#32;&#61;&#32;&#92;&#104;&#97;&#116;&#123;&#121;&#125;&#95;&#105;&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"74\" style=\"vertical-align: -5px;\"\/> ) and <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6a987274197f5fb6bfd3855d351bc2af_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#115;&#105;&#103;&#109;&#97;&#94;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"18\" style=\"vertical-align: 0px;\"\/>  is the noise variance, which is often considered a separate (\u2018nuisance\u2019) parameter.  We will consider <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-356a08e839ab6974a16448e16e56745d_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"9\" style=\"vertical-align: 0px;\"\/>  to be a vector listing the full set of parameters.  In one case of particular interest to us, isothermal titration (ITC) calorimetry for simple binding, we have <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-fe32749b885ad7f598cc1ac1cb113d48_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#32;&#61;&#32;&#40;&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#71;&#44;&#32;&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#72;&#44;&#32;&#92;&#115;&#105;&#103;&#109;&#97;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"131\" style=\"vertical-align: -5px;\"\/> .<\/p>\n<p>How do we get our \u2018fitted\u2019 parameters and uncertainties?  In the absence of usable prior information, the Bayesian and Maximum Likelihood pictures agree that the most likely set of parameters is the one that minimizes the squared error function <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-18afa9b3f6adb29e9cab2ab71d89a876_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#101;&#112;&#115;&#105;&#108;&#111;&#110;&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"29\" style=\"vertical-align: -5px;\"\/> .  The agreement between the two approaches stems from the common (sense) view that the probability that a given set of parameters <em>decreases<\/em> as error <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-18afa9b3f6adb29e9cab2ab71d89a876_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#101;&#112;&#115;&#105;&#108;&#111;&#110;&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"29\" style=\"vertical-align: -5px;\"\/>   <em>increases<\/em>.  In fact, this is often made explicit through a Gaussian noise assumption which leads to a simple equation for the likelihood <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-66a9f474fc3c52efdfb0ba6a70199ee8_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#76;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"12\" style=\"vertical-align: 0px;\"\/>  (i.e., probability \u2013 not normalized for simplicity) of a set of parameters as<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" width=\"291\" height=\"86\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-3-5.png\" class=\"wp-image-513\"><\/p>\n<p>We can assume this equation is exact for the purposes of our discussion.  A different form of the likelihood function doesn\u2019t change the conclusions we\u2019ll reach.<\/p>\n<p>Let\u2019s think about the \u2018space\u2019 of parameters, which is a fancy way of saying we\u2019ll consider all possible sets of parameters, e.g., <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-141a37146d8ed806ea831780a9e21e85_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#40;&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#71;&#44;&#32;&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#72;&#44;&#32;&#92;&#115;&#105;&#103;&#109;&#97;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"97\" style=\"vertical-align: -5px;\"\/>   in the (simplified) ITC case.  We can consider this space in the context of error <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-729568734d87ffb0f88cf42b1bc6828a_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#101;&#112;&#115;&#105;&#108;&#111;&#110;\" title=\"Rendered by QuickLaTeX.com\" height=\"8\" width=\"7\" style=\"vertical-align: 0px;\"\/>  or likelihood <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-66a9f474fc3c52efdfb0ba6a70199ee8_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#76;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"12\" style=\"vertical-align: 0px;\"\/>  \u2013 which give the same information because one is the negative log of the other.  In particular, think of what might be termed an \u2018error surface\u2019, a schematic of which is sketched in two dimensions <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6aff2f5b933dc19726a678ac42c343d8_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;&#44;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"52\" style=\"vertical-align: -5px;\"\/>  in the figure so we can visualize it.<\/p>\n<p>This error surface\/probability distribution\/energy landscape shows how using a ML strategy on a subset of variables (<img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>  in this case) artifactually shifts probability in the space.  By taking only the most probable value of <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  for each <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>  value, we discard some \u2013 potentially most \u2013 of the information contained in the <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  coordinate.  This approximation is equivalent to looking at a <strong>silhouette<\/strong> of the joint distribution (looking in the <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  direction), rather than integrating (summing) over all pertinent values of <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/> . Mathematically, the ML approach amounts to considering <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-9ae9ee9b6cfd2f83b07dac96c3615660_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#109;&#97;&#120;&#95;&#123;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;&#125;&#32;&#76;&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;&#44;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"115\" style=\"vertical-align: -5px;\"\/> , i.e., for each <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>  we consider only the <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  value which maximizes the likelihood.<\/p>\n<p>Generically, the ML approach will lead to errors in predicted confidence intervals.  Formulating this in mathematical language, the true probabilistic marginal which integrates (sums) over all <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  probability is not the same as the (implied) ML distribution, i.e., the silhouette distribution:<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" width=\"1304\" height=\"93\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-4-5.png\" class=\"wp-image-514\" srcset=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-4-5.png 1304w, https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-4-5-300x21.png 300w, https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-4-5-1024x73.png 1024w, https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-4-5-768x55.png 768w, https:\/\/statisticalbiophysicsblog.org\/wp-content\/uploads\/2022\/12\/word-image-481-4-5-788x56.png 788w\" sizes=\"auto, (max-width: 1304px) 100vw, 1304px\" \/><\/p>\n<p>Here, \u201cNormalized{ \u2026 }\u201d simply means the function in braces is normalized to a probability of 1 so we are making an apples-to-apples fair comparison.  The left side represents the true marginal \u2013 used in BI \u2013which includes all probability consistent with each <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>  value (integrated over <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>) whereas the right side is the \u201csilhouette\u201d distribution implicit in the ML approach which reflects only a single <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  for each <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>.  Equality is not expected in general, though it will hold exactly for uncorrelated variables, and I would expect it to hold approximately for linearly correlated variables.  If the two normalized probability densities are not the same, confidence intervals computed from them will generally differ.<\/p>\n<p>In a statistical mechanics sense, the error function <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-18afa9b3f6adb29e9cab2ab71d89a876_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#101;&#112;&#115;&#105;&#108;&#111;&#110;&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"29\" style=\"vertical-align: -5px;\"\/>  is analogous to the energy, so the error surface is like an energy landscape.  The scaled noise variance <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-4956c0134421345dbb45ab2f81828a36_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#115;&#105;&#103;&#109;&#97;&#94;&#50;&#47;&#78;\" title=\"Rendered by QuickLaTeX.com\" height=\"20\" width=\"42\" style=\"vertical-align: -5px;\"\/>  is analogous to the temperature: with less noise or more data, the distribution will become narrower.  Ignoring all energy values except the minimum is equivalent to ignoring entropy.<\/p>\n<p>In the equation above, the variables <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-6b23c893ca360d51e9b018181adfc8e3_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"14\" style=\"vertical-align: -3px;\"\/>  and <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-1f187e55d57f6da1e7cc07693994c3be_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"15\" style=\"vertical-align: -3px;\"\/>  can be considered vectors when considering the case of more than two dimensions.  Although I do not know of a rigorous result regarding higher dimensions, my intuition is that ignoring more dimensions in a higher dimensional problem will increase the severity of the error in the ML approximation.<\/p>\n<p>From my perspective, why would you use an approximation when there is a correct approach available?<\/p>\n<p>Regarding the estimation of other features of the distribution, the ML process will likewise lead to approximation.  This applies, for example, to the correlation structure, which fits into the formulation above: say we want to understand correlations among two parameters <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-e97346ffa2082f53364b2e4cbb9e31d9_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#97;&#44;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#98;&#41;&#32;&#61;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#49;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"90\" style=\"vertical-align: -5px;\"\/>  which may be considered as a vector for the equations above.  In exact analogy to the two-parameter description, we can contrast integrating over all other parameters <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-c5ebe55a0837849dfd2de3f9408ba44e_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#40;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#99;&#44;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#99;&#44;&#32;&#92;&#108;&#100;&#111;&#116;&#115;&#41;&#32;&#61;&#32;&#92;&#116;&#104;&#101;&#116;&#97;&#95;&#50;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"118\" style=\"vertical-align: -5px;\"\/>   in a BI picture vs. maximizing over those additional parameters.  From these different mathematical operations, we expect different results \u2013 though the nature of the difference will depend on the precise joint parameter distribution of the system \u2026 which is unknown in advance.<\/p>\n<p><em>Practical considerations in Bayesian vs. Maximum Likelihood error estimation<\/em><\/p>\n<p>There\u2019s no doubt that Bayesian computations require care and expertise \u2013 though we and others are working to reduce that burden.  On the other hand, based on some experience, I believe that ML calculations require more care than some practitioners may realize \u2013 potentially compounding the approximation noted above with numerical errors.  The problem for ML is that in a challenging system where the error landscape is not smooth, the well-known problem of start-point bias may corrupt the maximization process.  That is, different starting values for parameters in the maximization may lead to different apparent ML values, and this won\u2019t be detected unless the practitioner is extremely careful \u2026 and even if detected may be challenging to fix.  (This issue is independent of the missing entropy, which can\u2019t be fixed in ML.)<\/p>\n<p>Bayesian inference, by contrast, attempts to sample the likelihood landscape \u2013 all of it that contributes appreciably to the likelihood.  This requires a different mindset, but I would say a better awareness of the full space.  BI also requires more computing, but to my mind this is completely worthwhile.  After all, experimentalists may spend weeks or months preparing samples, calibrating equipment, and gathering data, so what\u2019s the big deal about spending a few hours or days to analyze it?<\/p>\n<p>A more speculative point about using BI is my belief that it promotes \u201clandscape awareness.\u201d  That is, serious BI will involve critical examination of sampling \u2013 hopefully through replicate sampling runs \u2013 that forces the user to reckon with challenges in the system.  A sampling mindset in which one worries about sampling quality\/replicability is extremely valuable for obtaining reliable results.<\/p>\n<p>I will note that some very complex systems may be out of reach for current BI software, but sampling checks should reveal whether this is the case, and then ML can be attempted.  On the other hand, if you do sample the landscape well under BI, you can also extract ML values, but without the start-point bias noted above.  I would also point out that BI developers are working actively to make software that is more powerful and reliable.<\/p>\n<p><em>A word about priors: they\u2019re not just for Bayesian inference<\/em><\/p>\n<p>Prior assumptions and knowledge are part of doing science, and the quantification of such information is always a part of doing parameter uncertainty analysis whether we realize it or not.  That is, the \u2018prior\u2019 (assumed distribution of possible outcomes in the absence of data) is an implicit part of maximum likelihood <em>and <\/em>Bayesian uncertainty quantification.<\/p>\n<p>In the simplest case, we may assume a range of possible values for a given parameter.  We hope that the likelihood function becomes negligible near the limits of our assumed range, and then we may imagine the prior had no bearing on our uncertainty range.  This is not quite true, either for BI or ML.<\/p>\n<p>The complication for both BI and ML is that typically there is no unique, fundamental representation of a parameter.  In ITC analysis, for example, we are interested in binding free energies <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-a930c3edefd57fae159d8f76cdeb4816_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#71;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"29\" style=\"vertical-align: 0px;\"\/>  but these can often be expressed in terms of the dissociation constant <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-7639a351cce72c9e3a9b113e7c9d5c69_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#75;&#95;&#100;&#32;&#61;&#32;&#92;&#101;&#120;&#112;&#40;&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#71;&#47;&#82;&#84;&#41;\" title=\"Rendered by QuickLaTeX.com\" height=\"19\" width=\"151\" style=\"vertical-align: -5px;\"\/> .  In a different system, we might be interested in an angle measured in degrees, but we also might describe it via the cosine of the angle.  Those familiar with Jacobians will recognize the problem: our representation of the parameter will affect how much probability is allotted to different parts of the parameter range of interest.  Consider the difference between a certain parameter and its log, where we confine ourselves to the range from 0.0001 to 1.  If we assume a uniform distribution in log space, there is equal probability (25%) assigned to each of four intervals: 0.0001 to 0.001, 0.001 to 0.01, 0.01 to 0.1, and 0.1 to 1.  However, you can immediately see that this assigns more than 75% of the probability to values below 0.5, so it\u2019s certainly not uniform in the linear representation of the coordinate, where we\u2019d expect just about 50% of the probability above and below 0.5.  It\u2019s not exactly 50% because the interval started at 0.0001 and not zero.<\/p>\n<p>This parameter representation or Jacobian issue will affect BI because \u2013 in the ITC case, for example \u2013 a prior that is uniform in <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-a930c3edefd57fae159d8f76cdeb4816_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#92;&#68;&#101;&#108;&#116;&#97;&#32;&#71;\" title=\"Rendered by QuickLaTeX.com\" height=\"12\" width=\"29\" style=\"vertical-align: 0px;\"\/>  is not uniform in <img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/statisticalbiophysicsblog.org\/wp-content\/ql-cache\/quicklatex.com-c6a66f7efd79d61532658051f677c8e6_l3.png\" class=\"ql-img-inline-formula quicklatex-auto-format\" alt=\"&#75;&#95;&#100;\" title=\"Rendered by QuickLaTeX.com\" height=\"15\" width=\"23\" style=\"vertical-align: -3px;\"\/>.  The issue may not affect the ML procedure if it is based purely on error or likelihood values (rather than integrals over likelihood) but there is an underlying implicit assumption of a prior distribution based on the parameter characterization.<\/p>\n<p>Bottom line: Priors are part of parameter uncertainty estimation whether we like it or not, and whether you apply Bayesian inference or not.<\/p>\n<p><em>Summing up<\/em><\/p>\n<p>The table below compares BI and ML.  I favor BI because it avoids approximation and forces the practitioner to be more aware of all the issues affecting uncertainty, including landscape roughness which affects both BI and ML.<\/p>\n<table>\n<tbody>\n<tr>\n<td colspan=\"2\"><strong>Model Parameter Uncertainty Estimation from Data<\/strong><\/td>\n<\/tr>\n<tr>\n<td><strong>Bayesian Inference<\/strong><\/td>\n<td><strong>Maximum Likelihood<\/strong><\/td>\n<\/tr>\n<tr>\n<td>\n  Agrees with ML on the most likely set of parameters, although prior assumptions may shift values.<\/td>\n<td>\n  Uses likelihood function alone for most likely parameters, aside from possible prior ranges.<\/td>\n<\/tr>\n<tr>\n<td>\n  Computations require more time and expertise.  Must check sampling.<\/td>\n<td>\n  Computations require less time but do require care.  Must check sensitivity to initial points for optimization.<\/td>\n<\/tr>\n<tr>\n<td>\n  Uncertainty ranges account for all likelihood in full parameter space, i.e., they use true marginals<\/td>\n<td>\n  Approximates likelihood for parameter(s) of interest by optimizing excluded parameters, i.e,. uses \u2018silhouette\u2019 distributions<\/td>\n<\/tr>\n<tr>\n<td>\n  Correlation structure includes all likelihood via true marginals<\/td>\n<td>\n  Correlation structure approximated by optimization of excluded parameters<\/td>\n<\/tr>\n<tr>\n<td>\n  BI can (in)validate ML<\/td>\n<td>\n  Does not reveal when ML approximation works well<\/td>\n<\/tr>\n<tr>\n<td>\n  Explicitly quantifies prior assumptions (beware subtleties)<\/td>\n<td>\n  Includes prior assumptions implicitly (beware subtleties)<\/td>\n<\/tr>\n<tr>\n<td>\n  Accounts for unknown magnitude of noise<\/td>\n<td>\n  Typically assumes fixed  value for noise<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p><em>Note<\/em><\/p>\n<p>This post was previously \u2018published\u2019 as a preprint, with the same content.  See <a href=\"https:\/\/osf.io\/ajuvf\/\">https:\/\/osf.io\/ajuvf\/<\/a><\/p>\n<p><em>Acknowledgements<\/em><\/p>\n<p>Aidan Estelle, August George, Douglas Walker, and Mike Harms offered helpful comments on this discussion.  I am grateful for support from the NSF (Grant MCB 2119837) and NIH (Grant GM141733).<\/p>\n<p><em>Further reading<\/em><\/p>\n<p>Bevington, Philip R., and D. Keith Robinson. &#8220;Data reduction and error analysis.&#8221; <em>McGraw-Hill, New York<\/em> (2003).  Describes ML uncertainty analysis.<\/p>\n<p>Duvvuri, Hiranmayi, Lucas C. Wheeler, and Michael J. Harms. &#8220;PYTC: open-source Python software for global analyses of isothermal titration calorimetry data.&#8221; <a href=\"https:\/\/pubs.acs.org\/doi\/full\/10.1021\/acs.biochem.7b01264\"><em>Biochemistry<\/em> 57.18 (2018): 2578-2583<\/a>.  Uses Bayesian and other methods of uncertainty analysis.<\/p>\n<p>Estelle, Aidan B., et al. &#8220;Quantifying cooperative multisite binding through Bayesian inference.&#8221; <a href=\"https:\/\/www.biorxiv.org\/content\/10.1101\/2022.06.29.498022v2.abstract\"><em>bioRxiv<\/em> (2022)<\/a>.<\/p>\n<p>Nguyen, Trung Hai, et al. &#8220;Bayesian analysis of isothermal titration calorimetry for binding thermodynamics.&#8221; <a href=\"https:\/\/journals.plos.org\/plosone\/article?id=10.1371\/journal.pone.0203224\"><em>PLoS ONE<\/em> 13.9 (2018): e0203224<\/a>.<\/p>\n<p>Zhao, Huaying, Grzegorz Piszczek, and Peter Schuck. &#8220;SEDPHAT\u2013a platform for global ITC analysis and global multi-method analysis of molecular interactions.&#8221; <a href=\"https:\/\/www.sciencedirect.com\/science\/article\/abs\/pii\/S1046202314003752\"><em>Methods<\/em> 76 (2015): 137-148<\/a>.  Employs ML uncertainty analysis.<\/p>\n<p>Zuckerman, Daniel M. <em>Statistical physics of biomolecules: an introduction<\/em>. CRC press, 2010.  Explains entropy and probability.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>When we want to estimate parameters from data (e.g., from binding, kinetics, or electrophysiology experiments), there are two tasks: (i) estimate the most likely values, and (ii) equally importantly, estimate the uncertainty in those values. After all, if the uncertainty is huge, it\u2019s hard to say we really know the parameters. We also need to [&hellip;]<\/p>\n","protected":false},"author":6,"featured_media":483,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20,12,18],"tags":[],"class_list":["post-481","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-bayesian-statistics","category-binding","category-statistical-uncertainty"],"_links":{"self":[{"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/posts\/481","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=481"}],"version-history":[{"count":8,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/posts\/481\/revisions"}],"predecessor-version":[{"id":515,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/posts\/481\/revisions\/515"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=\/wp\/v2\/media\/483"}],"wp:attachment":[{"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=481"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=481"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/statisticalbiophysicsblog.org\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=481"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}