Abstract
Given a finite and noisy dataset generated with a closed-form mathematical model, when is it possible to learn the true generating model from the data alone? This is the question we investigate here. We show that this model-learning problem displays a transition from a low-noise phase in which the true model can be learned, to a phase in which the observation noise is too high for the true model to be learned by any method. Both in the low-noise phase and in the high-noise phase, probabilistic model selection leads to optimal generalization to unseen data. This is in contrast to standard machine learning approaches, including artificial neural networks, which in this particular problem are limited, in the low-noise phase, by their ability to interpolate. In the transition region between the learnable and unlearnable phases, generalization is hard for all approaches including probabilistic model selection.
Similar content being viewed by others
Introduction
For a few centuries, scientists have described natural phenomena by means of relatively simple mathematical models such as Newtonâs law of gravitation or Snellâs law of refraction. Sometimes, they arrived to these models deductively, starting from fundamental considerations; more frequently, however, they derived the models inductively from data. With increasing amounts of data available for all sorts of (natural and social) systems, one may argue that we are now in a position to inductively uncover new interpretable models for these systems. To this end, machine learning approaches that can automatically uncover closed-form models from data have been recently developed1,2,3,4,5,6. (Here and throughout this article, we refer to closed-form models as those that are expressed in terms of a relatively small number of basic functions, such as addition, multiplication, trigonometric functions, etc.) In physics alone, such approaches have been applied successfully to quantum systems7, non-linear and chaotic systems2,4, fluid mechanics8, and astrophysics9, among others6.
A central assumption implicit in these approaches is that, given data, it is always possible to identify the correct underlying model. Here, we investigate the validity of this assumption. In particular, consider a dataset \(D=\left\{({y}_{i},{{{{{{{{\bf{x}}}}}}}}}_{i})\right\}\), with iâ=â1,ââ¦,âN, generated using the closed-form model m*, so that yiâ=âm*(xi,âθ*)â+âϵi with θ* being the parameters of the model, and ϵi a random unbiased observation noise drawn from the normal distribution with variance \({s}_{\epsilon }^{2}\). The assumption of Gaussian noise is standard in regression and symbolic regression problems. In principle, one could assume other noise structures (for example, multiplicative noise) or even more general likelihoods, but these would be hard to justify in the context of regression and symbolic regression/model discovery. The question we are interested in is: Assuming that m* can be expressed in closed form, when is it possible to identify it as the true generating model among all possible closed-form mathematical models, for someone who does not know the true model beforehand? Note that our focus is on learning the structure of the model m* and not the values of the parameters θ*, a problem that has received much more attention from the theoretical point of view10. Additionally, we are interested in situations in which the dimension of the feature space \({{{{{{{\bf{x}}}}}}}}\in {{\mathbb{R}}}^{k}\) is relatively small (compared to typical feature spaces in machine learning settings), which is the relevant regime for symbolic regression and model discovery.
To address the model-learning question above, we formulate the problem of identifying the true generating model probabilistically, and show that probabilistic model selection is quasi-optimal at generalization, that is, at making predictions about unobserved data. This is in contrast to standard machine learning approaches, which, in this case, are suboptimal in the region of low observation noise. We then investigate the transition occurring between: (i) a learnable phase at low observation noise, in which the true model can in principle be learned from the data; and (ii) an unlearnable phase, in which the observation noise is too large for the true model to be learned from the data by any method. Finally, we provide an upper bound for the noise at which the learnability transition takes place, that is, the noise beyond which the true generating model cannot be learned by any method. This bound corresponds to the noise at which most plausible a priori models, which we call trivial, become more parsimonious descriptions of the data than the true generating model. Despite the simplicity of the approach, the bound provides a good approximation to the actual transition point.
Results
Probabilistic formulation of the problem
The complete probabilistic solution to the problem of identifying a model from observations is encapsulated in the posterior distribution over models p(mâ£D). The posterior gives the probability that each model mâ=âm(x,âθ), with parameters θ, is the true generating model given the data D. Again, notice that we are interested on the posterior over model structures m rather than model parameters θ; thus, we obtain p(mâ£D) by marginalizing p(m,âθâ£D) over possible parameter values Î
where p(Dâ£m,âθ) is the model likelihood, and p(θâ£m) and p(m) are the prior distributions over the parameters of a given model and the models themselves, respectively. The posterior over models can always be rewritten as
with \(Z=p(D)={\sum }_{\{m\}}\exp \left[-{{{{{{{\mathcal{H}}}}}}}}(m)\right]\) and
Although the integral over parameters in Eq. (3) cannot, in general, be calculated exactly, it can be approximated as5,11
where B(m) is the Bayesian information criterion (BIC) of the model11. This approximation results from using Laplaceâs method to integrate the distribution p(Dâ£m,âθ)p(θâ£m) over the parameters θ. Thus, the calculation assumes that: (i) the likelihood p(Dâ£m,âθ) is peaked around \({\theta }^{*}=\arg \mathop{\max}\limits_{\theta }p(D|m,\theta )\), so that it can be approximated by a Gaussian around θ*; (ii) the prior p(θâ£m) is smooth around θ* so that it can be assumed to be approximately constant within the Gaussian. Unlike other contexts, in regression-like problems these assumptions are typically mild.
Within an information-theoretical interpretation of model selection, \({{{{{{{\mathcal{H}}}}}}}}(m)\) is the description length, that is, the number of nats (or bits if one uses base 2 logarithms instead of natural logarithms) necessary to jointly encode the data and the model with an optimal encoding12. Therefore, the most plausible modelâthat is, the model with maximum p(mâ£D)âhas the minimum description length, that is, compresses the data optimally.
The posterior in Eq. (2) can also be interpreted as in the canonical ensemble in statistical mechanics, with \({{{{{{{\mathcal{H}}}}}}}}(m)\) playing the role of the energy of a physical system, models playing the role of configurations of the system, and the most plausible model corresponding to the ground state. Here, we sample the posterior distribution over models p(mâ£D) by generating a Markov chain with the Metropolis algorithm, as one would do for physical systems. To do this, we use the âBayesian machine scientistâ introduced in ref. 5 (Methods). We then select, among the sampled models, the most plausible one (that is, the maximum a posteriori model, the minimum description length model, or the ground state, in each interpretation).
Probabilistic model selection yields quasi-optimal predictions for unobserved data
Probabilistic model selection as described above follows directly (and exactly, except for the explicit approximation in Eq. (4)) from the postulates of Coxâs theorem13,14 and is therefore (Bayes) optimal when models are truly drawn from the prior p(m). In particular, the minimum description length (MDL) model is the most compressive and the most plausible one, and any approach selecting, from D, a model that is not the MDL model violates those basic postulates.
We start by investigating whether this optimality in model selection also leads to the ability of the MDL model to generalize well, that is, to make accurate predictions about unobserved data. We show that the MDL model yields quasi-optimal generalization despite the fact that the best possible generalization is achieved by averaging over models5, and despite the BIC approximation in the calculation of the description length (Fig. 1). Specifically, we sample a model m* from the prior p(m) described in ref. 5 and in the Methods; for a model not drawn from the prior, see Supplementary Text and Fig. S1. From this model, we generate synthetic datasets \(D=\left\{({y}_{i},{{{{{{{{\bf{x}}}}}}}}}_{i})\right\}\), iâ=â1,ââ¦,âN (where yiâ=âm*(xi,âθ*)â+âϵi and ϵiâ~âGaussian(0,âsϵ)) with different number of points N and different levels of noise sϵ. Then, for each dataset D, we sample models from p(mâ£D) using the Bayesian machine scientist5, select the MDL model among those sampled, and use it to make predictions on a test dataset \({D}^{{\prime} }\).
Because \({D}^{{\prime} }\) is, like D, subject to observation noise, the irreducible error is sϵ, that is, the root mean squared error (RMSE) of predictions cannot be, on average, smaller than sϵ. As we show in Fig. 1, the predictions of the MDL model achieve this optimal prediction limit except for small N and some intermediate values of the observation noise. This is in contrast to standard machine learning algorithms, such as artificial neural networks. These algorithms achieve the optimal prediction error at large values of the noise, but below certain values of sϵ the prediction error stops decreasing and predictions become distinctly suboptimal (Fig. 1).
In the limit sϵââââ of high noise, all models make predictions whose errors are small compared to sϵ. Thus, the prediction error is similar to sϵ regardless of the chosen model, which also means that it is impossible to correctly identify the model that generated the data. Conversely, in the sϵâââ0 limit, the limiting factor for standard machine learning methods is their ability to interpolate between points in D, and thus the prediction error becomes independent of the observation error. By contrast, because Eqs. (2)â(4) provide consistent model selection, in this limit the MDL should coincide with the true generating model m* and interpolate perfectly. This is exactly what we observeâthe only error in the predictions of the MDL model is, again, the irreducible error sϵ. Therefore, our observations show that probabilistic model selection leads to quasi-optimal generalization in the limits of high and low observation noise.
Phenomenology of the learnability transition
Next, we establish the existence of the learnable and unlearnable regimes, and clarify how the transition between them happens. Again, we generate synthetic data D using a known model m* as in the previous section, and sample models from the posterior p(mâ£D). To investigate whether a model is learnable we consider the ensemble of sampled models (Fig. 2); if no model in the ensemble is more plausible than the true generating model (that is, if the true model is also the MDL model), then we conclude that the model is learnable. Otherwise, the model is unlearnable because, lacking any additional information about the generating model, it is impossible to identify it; other models provide more parsimonious descriptions of the data D.
To this end, we first consider the gap \(\Delta {{{{{{{\mathcal{H}}}}}}}}(m)={{{{{{{\mathcal{H}}}}}}}}(m)-{{{{{{{\mathcal{H}}}}}}}}({m}^{*})\) between the description length of each sampled model m and that of the true generating model m* (Fig. 2a). For low observation error sϵ, all sampled models have positive gaps, that is, description lengths that are longer than or equal to the true model. This indicates that the most plausible model is the true model; therefore, the true model is learnable. By contrast, when observation error grows (sϵââ³â0.6 in Fig. 2a), some of the models sampled for some training datasets D start to have negative gaps, which means that, for those datasets, these models become more plausible than the true model. When this happens, the true model becomes unlearnable. For even larger values of sϵ (sϵâ>â2), the true model is unlearnable for virtually all training sets D.
To better characterize the sampled models, we next compute description length of the model; that is, the term \({{{{{{{{\mathcal{H}}}}}}}}}_{M}(m)=-\ln p(m)\) in the description length, which measures the complexity of the model regardless of the data (Fig. 2b). For convenience we set an arbitrary origin for \({{{{{{{{\mathcal{H}}}}}}}}}_{M}(m)\) so that \({{{{{{{{\mathcal{H}}}}}}}}}_{M}({m}^{c})=0\) for the models \({m}^{c}=\arg {\max }_{m}p(m)=\arg {\min }_{m}{{{{{{{{\mathcal{H}}}}}}}}}_{M}(m)\) that are most plausible a priori, which we refer to as trivial models; all other models have \({{{{{{{{\mathcal{H}}}}}}}}}_{M}(m) \; > \;0\). We observe that, for low observation error, most of the sampled models are more complex than the true model. Above a certain level of noise (sϵââ³â3), however, most sampled models are trivial or very simple models with \({{{{{{{{\mathcal{H}}}}}}}}}_{M}(m)\;\approx \;0\). These trivial models appear suddenlyâat sϵâ<â2 almost no trivial models are sampled at all.
Altogether, our analysis shows that for low enough observation error the true model is always recovered. Conversely, in the opposite limit only trivial models (those that are most plausible in the prior distribution) are considered reasonable descriptions of the data. Importantly, our analysis shows that trivial models appear suddenly as one increases the level of noise, which suggests that the transition to the unlearnable regime may be akin to a phase transition driven by changes in the model plausibility (or description length) landscape, similarly to what happens in other problems in inference, constraint satisfaction, and optimization10,15.
To further probe whether that is the case, we analyze the transition in more detail. Our previous analysis shows that learnability is a property of the training dataset D, so that, for the same level of observation noise, some instances of D enable learning of the true model whereas others do not. Thus, by analogy with satisfiability transitions15, we define the learnability Ï(sϵ) as the fraction of training datasets D at a given noise sϵ for which the true generating model is learnable. In Fig. 3, we show the behavior of the learnability Ï(sϵ) for different sizes N, for the two models in Fig. 1 (see Supplementary Fig. S1 for another model). Consistent with the qualitative description above, we observe that the learnability transition occurs abruptly at a certain value of the observation noise; the transition shifts towards higher values of sϵ with increasing size of D.
Remarkably, the difficult region identified in the previous section, in which prediction error deviates from the theoretical irreducible error sϵ, coincides with the transition region. Therefore, our results paint a picture with three qualitatively distinct regimes: a learnable regime in which the true model can always be learned and predictions about unobserved data are optimal; an unlearnable regime in which the true model can never be learned but predictions are still optimal because error is driven by observation noise and not by the model itself; and a transition regime in which the true model is learnable only sometimes, and in which predictions are, on average, suboptimal. This, again, is reminiscent of the hard phase in satisfiability transitions and of other statistical learning problems15,16,17.
Learnability conditions
Finally, we obtain an upper bound to the learnability transition point by assuming that all the phenomenology of the transtion can be explained by the competition between just two minima of the description length landscape \({{{{{{{\mathcal{H}}}}}}}}(m)\): the minimum corresponding to the true model m*; and the minimum corresponding to the trivial model mc that is most plausible a priori
As noted earlier, mc is such that \({{{{{{{{\mathcal{H}}}}}}}}}_{M}({m}^{c})=0\) by our choice of origin, and for our choice of priors it corresponds to trivial models without any operation, such as \(m({{{{{{{\bf{x}}}}}}}})={{{{{{{\rm{const.}}}}}}}}\) or m(x)â=âx1. (Note that, although the origin of descriptions lengths is not arbitrary, this choice is innocuous as long as we are only concerned with comparisons between models.) Below, we focus on one of these models \(m({{{{{{{\bf{x}}}}}}}})={{{{{{{\rm{const.}}}}}}}}\), but our arguments (not the precise calculations) remain valid for any other choice of prior and the corresponding mc.
As we have also noted above, the true model is learnable when the description length gap \(\Delta {{{{{{{\mathcal{H}}}}}}}}(m)={{{{{{{\mathcal{H}}}}}}}}(m)-{{{{{{{\mathcal{H}}}}}}}}({m}^{*})\) is strictly positive for all models mââ âm*. Conversely, the true model m* becomes unlearnable when \(\Delta {{{{{{{\mathcal{H}}}}}}}}(m)=0\) for some model mââ âm*. As per our assumption that the only relevant minima are those corresponding to m* and mc, we therefore postulate that the transition occurs when the description length gap of the trivial model becomes, on average (over datasets), \(\Delta {{{{{{{\mathcal{H}}}}}}}}({m}^{c})=0\). In fact, when \({{{{{{{\mathcal{H}}}}}}}}({m}^{c})\le {{{{{{{\mathcal{H}}}}}}}}({m}^{*})\) the true model is certainly unlearnable; but other models m could fullfill the condition \({{{{{{{\mathcal{H}}}}}}}}(m)\le {{{{{{{\mathcal{H}}}}}}}}({m}^{*})\) earlier, thus making the true model unlearnable at lower observation noises. Therefore, the condition \(\Delta {{{{{{{\mathcal{H}}}}}}}}({m}^{c})=0\) yields an upper bound to the value of the noise at which the learnability transition happens. We come back to this issue later, but for now we ignore all potential intermediate models.
Within the BIC approximation, the description length of each of these two models is (âMethodsâ section)
where k* is the number of parameters of the true model, \({\delta }_{i}={m}_{i}^{c}-{m}_{i}^{*}\) is the reducible error of the trivial model, and the averages ãâ¯ãD are over the observations i in D. Then, over many realizations of D, the transition occurs on average at
where ãδ2ã is now the variance of m* over the observation interval rather than in any particular dataset. For large Nââ³â100 and model description lengths of the true model such that \({\Delta }_{M}^{*}\equiv {{{{{{{{\mathcal{H}}}}}}}}}_{M}({m}^{*})-{{{{{{{{\mathcal{H}}}}}}}}}_{M}({m}^{c}) \sim O(1)\), the transition noise is well approximated by (âMethodsâ section)
Therefore, since \({s}_{\epsilon }^{\times }\) diverges for Nââââ, the true model is learnable for any finite observation noise, provided that N is large enough (which is just another way of seeing that probabilistic model selection with the BIC approximation is consistent11).
In Fig. 3, we show that the upper bound in Eq. (8) is close to the exact transition point from the learnable to the unlearnable phase, as well as to the peak of the prediction error relative to the irreducible error. Moreover, once represented as a function of the scaled noise \({s}_{\epsilon }/{s}_{\epsilon }^{\times }\), the learnability curves of both models collapse into a single curve (Fig. 3e), suggesting that the behavior at the transition may be universal. Additionally, the transition between m* and mc becomes more abrupt with increasing N, as the fluctuations in ãδãD become smaller. If this upper bound became tighter with increasing N, this would be indicative of a true, discontinuous phase transition between the learnable and the unlearnable phases.
To further understand the transition and further test the validity of the assumption leading to the upper bound \({s}_{\epsilon }^{\times }\), we plot the average description length (over datasets D) of the MDL model at each level of observation noise sϵ (Fig. 4). We observe that the description length of the MDL model coincides with the description length \({{{{{{{\mathcal{H}}}}}}}}({m}^{*})\) of the true model below the transition observation noise, and with the description length \({{{{{{{\mathcal{H}}}}}}}}({m}^{c})\) of the trivial model above the transition observation noise. Around \({s}_{\epsilon }^{\times }\), and especially for smaller N, the observed MDL is lower than both \({{{{{{{\mathcal{H}}}}}}}}({m}^{*})\) and \({{{{{{{\mathcal{H}}}}}}}}({m}^{c})\), suggesting that, in the transition region, a multiplicity of models (or Rashomon set18) other than m* and mc become relevant.
Discussion
The bidirectional cross fertilization between statistical learning theory and statistical physics goes back to the 1980âs, when the paradigm in artificial intelligence shifted from rule-based approaches to statistical learning approaches19. Nonetheless, the application of statistical physics ideas and tools to learning problems has so far focused on learning parameter values17,20,21,22, or on specific problems such as learning probabilistic graphical models and, more recently, network models23,24,25,26,27. Thus, despite the existence of rigorous probabilistic model selection approaches28, the issue of learning the structure of models, and especially closed-form mathematical models, has received much less attention from the statistical physics point of view10.
Our approach shows that, once the model-selection problem is formalized in such a way that models can be represented and enumerated, and their posteriors p(mâ£D) can be estimated and sampled in the same way we do for discrete configurations in a physical system5, there is no fundamental difference between learning models and learning parameters (except, perhaps, for the discreteness of model structures). Therefore, the same richness that has been described at length in parameter learning can be expected in model learning, including the learnability transition that we have described here, but perhaps others related, for example, to the precise characterization of the description length landscape in the hard phase in which generalization is difficult. We may also expect model-learning and parameter-learning transitions to interact. For example, there are limits to our ability to learn model parameters, even for noiseless data17,29,30; in this unlearnable phase of the parameter learning problem, it seems that the true model should also be unlearnable, even in this noiseless limit, something we have not considered here. Our findings are thus the first step in the characterization of the rich phenomenology arising from the interplay between data size, noise, and parameter and model identification from data.
Our results on closed-form mathematical models may also open the door to addressing some problems of learning that have so far remained difficult to tackle. In deep learning, for example, the question of how many training examples are necessary to learn a certain task with precision is still open19, whereas, as we have shown, in our context of (closed-form) model selection the answer to this question arises naturally.
Finally, we have shown that, when data are generated with a (relatively simple) closed-form expression, probabilistic model selection generalizes quasi-optimally, especially in the ideal regime of low observation noise. This is in contrast, as we have seen, to what happens in this case with some machine learning approaches such as artificial neural networks. Conversely, it would be interesting to understand how expressive closed-form mathematical models are, that is, to what extent they are able to describe and generalize when data are not generated using a closed-form mathematical model (for example, for the solution of a differential equation that cannot be expressed in closed form). We hope that our work encourages more research on machine learning approaches geared towards learning such interpretable models18, and on the statistical physics of such approaches.
Methods
Prior over models
The probabilistic formulation of the model selection problem outlined in Eqs. 1â4 requires the choice of a prior distribution p(m) over modelsâalthough the general, qualitative results described in the body of the text do not depend on a particular choice of prior. Indeed, no matter how priors are chosen, some model \({m}^{c}=\arg {\max }_{m}p(m)\) (or, at most, a finite subset of models, since uniform priors cannot be used5) different from the true generating model m* will be most plausible a priori, so the key arguments in the main text hold.
However, the quantitative results must be obtained for one specification of p(m). Here, as in the previous literature5,8, we choose p(m) to be the maximum entropy distribution that is compatible with an empirical corpus of mathematical equations5. In particular, we impose that the mean number ãnoã of occurrences of each operation \(o \in \{+,*,\exp, \ldots \}\) per equation is the same as in the empirical corpus; and that the fluctuations of these numbers \(\langle {n}_{o}^{2}\rangle\) are also as in the corpus. Therefore, the prior is
where \({{{{\mathcal{O}}}}}=\{+,*, \exp , \ldots \}\), and the hyperparameters αoââ¥â0 and βoââ¥â0 are chosen so as to reproduce the statistical features of the empirical corpus5. For this particular choice of prior, the models \({m}^{c}=\arg {\max }_{m}p(m)\) are those that contain no operations at all, for example \({m}^{c}={{{{{{{\rm{const.}}}}}}}}\); hence the use of the term trivial to denote these models in the text.
Sampling models with the Bayesian machine scientist
For a given dataset D, each model m has a description length \({{{{{{{\mathcal{H}}}}}}}}(m)\) given by Eq. (4). The Bayesian machine scientist5 generates a Markov chain of models {m0,âm1,ââ¦,âmT} using the Metropolis algorithm as described next.
Each model is represented as an expression tree, and each new model mt+1 in the Markov chain is proposed from the previous one mt by changing the corresponding expression tree: changing an operation or a variable in the expression tree (for example, proposing a new model mt+1â=âθ0â+âx1 from mtâ=âθ0*x1); adding a new term to the tree (for example, mt+1â=âθ0*x1â+âx2 from mtâ=âθ0*x1); or replacing one block of the tree (for example, \({m}_{t+1}={\theta }_{0}*\exp ({x}_{2})\) from mtâ=âθ0*x1) (see ref. 5 for details). Once a new model mt+1 is proposed from mt, the new model is accepted using the Metropolis rule.
Note that the only input to the Bayesian machine is the observed data D. In particular, the observational noise sϵ is unknown and must be estimated, via maximum likelihood, to calculate the Bayesian information criterion B(m) of each model m.
Artificial neural network benchmarks
For the analysis of predictions on unobserved data, we use as benchmarks the following artificial neural network architectures and training procedures. The networks consist of: an input layer with two inputs corresponding to x1 and x2; (ii) four hidden fully connected feed-forward layers, with 10 units each and ReLU activation functions; and (iii) a linear output layer with a single output y.
Each network was trained with a dataset D containing N points, just as in the probabilistic model selection experiments. Training errors and validation errors (computed on an independent set) were calculated, and the training process stopped when the validation error increased, on average, for 100 epochs; this typically entailed training for 1000â2000 epochs. This procedure was repeated three times, and the model with the overall lowest validation error was kept for making predictions on a final test set \({D}^{{\prime} }\). The predictions on \({D}^{{\prime} }\) are those reported in Fig. 1.
Model description lengths
The description length of a model is given by Eq. (4), and the BIC is
where \({\left.{{{{{{{\mathcal{L}}}}}}}}(m)\right|}_{\hat{\theta }}=p(D|m,\hat{\theta })\) is the likelihood of the model calculated at the maximum likelihood estimator of the parameters, \(\hat{\theta }=\arg {\max }_{\theta }p(D|m,\theta )\), and k is the number of parameters in m. In a model-selection setting, one would typically assume that the deviations of the observed data are normally distributed independent variables, so that
where miââ¡âm(xi), and sy is the observed error, which in general differs from sϵ when mââ âm*. We can obtain the maximum likelihood estimator for \({{s}_{y}}^{2}\), which gives \({{s}_{y}}^{2}=\frac{1}{N}{\sum }_{i}{({y}_{i}-{m}_{i}^{*})}^{2}\), and replacing into Eqs. (4)â(12) gives
For mâ=âm* we have that \({{s}_{y}}^{2}=\frac{1}{N}{\sum }_{i}{({y}_{i}-{m}_{i}^{*})}^{2}={\langle {\epsilon }^{2}\rangle }_{D}\). For any other model m, we define for each point the deviation from the original model \({\delta }_{i}:={m}_{i}-{m}_{i}^{*}\) so that \({{s}_{y}}^{2}=\frac{1}{N}{\sum }_{i}{({\epsilon }_{i}-{\delta }_{i})}^{2}={\langle {\epsilon }^{2}\rangle }_{D}+{\langle {\delta }^{2}\rangle }_{D}-2{\langle \delta \epsilon \rangle }_{D}\). Plugging these expressions into Eq. (13), we obtain the expressions in Eqs. (6) and (7).
Approximation of \({s}_{\epsilon }^{\times }\) for large N
Defining xâ=â1/N in Eq. (8) and using the Puiseux series expansion
around xâ=â0, we obtain Eq. (9).
Data availability
We did not use any data beyond the synthetically generated data described in the manuscript.
Code availability
The code for the Bayesian machine scientist is publicly available as a repository from the following URL: https://bitbucket.org/rguimera/machine-scientist/.
References
Džeroski, S. & Todorovski, L (eds). Computational Discovery of Scientific Knowledge. Lecture Notes in Artificial Intelligence (Springer, 2007).
Schmidt, M. & Lipson, H. Distilling free-form natural laws from experimental data. Science 324, 81â5 (2009).
Evans, J. & Rzhetsky, A. Machine science,. Science 329, 399â400 (2010).
Brunton, S. L., Proctor, J. L. & Kutz, J. N. Discovering governing equations from data by sparse identification of nonlinear dynamical systems,. Proc. Natl Acad. Sci. USA 113, 3932â3937 (2016).
Guimerà , R. et al. A Bayesian machine scientist to aid in the solution of challenging scientific problems,. Sci. Adv. 6, eaav6971 (2020).
Udrescu, S.-M. & Tegmark, M. AI Feynman: a physics-inspired method for symbolic regression,. Sci. Adv. 6, eaay2631 (2020).
Gentile, A. A. et al. Learning models of quantum systems from experiments,. Nat. Phys. 17, 837â843 (2021).
Reichardt, I., Pallarès, J., Sales-Pardo, M. & Guimerà , R. Bayesian machine scientist to compare data collapses for the Nikuradse dataset,. Phys. Rev. Lett. 124, 084503 (2020).
Cranmer, M. et al. Advances in Neural Information Processing Systems, Vol. 33 (eds. H Larochelle, H., Ranzato, M. Hadsell, R. Balcan, M. F. & Lin, H.) (Curran Associates, Inc., 2020) pp. 17429â17442.
Zdeborová, L. & Krzakala, F. Statistical physics of inference: Thresholds and algorithms,. Adv. Phys. 65, 453â552 (2016).
Schwarz, G. Estimating the dimension of a model,. Ann. Stat. 6, 461â464 (1978).
Grünwald, P. D. The Minimum Description Length Principle (The MIT Press, Cambridge, Massachusetts, 2007).
Cox, R. T. Probability, frequency and reasonable expectation,. Am. J. Phys. 14, 1â10 (1946).
Jaynes, E. T. Probability Theory: The Logic of Science (Cambridge University Press, 2003).
Mézard, M. & Montanari, A. Information, Physics, and Computation (Oxford University Press, 2009).
Ricci-Tersenghi, F., Semerjian, G. & Zdeborová, L. Typology of phase transitions in bayesian inference problems,. Phys. Rev. E 99, 042109 (2019).
Maillard, A., Loureiro, B., Krzakala, F. & Zdeborová, L. Advances in Neural Information Processing Systems, Vol. 33 (eds Larochelle, H., Ranzatom, M., Hadsell, R., Balcan, M. F. & Lin, H. (Curran Associates, Inc., 2020) pp. 11071â11082.
Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead,. Nat. Mach. Intell. 1, 206â215 (2019).
Carleo, G. et al. Machine learning and the physical sciences,. Rev. Mod. Phys. 91, 45002 (2019).
Engel, A. & Van den Broeck, C. Statistical Mechanics of Learning (Cambridge University Press, Cambridge, UK, 2004).
Aubin, B., Krzakala, F., Lu, Y. and Zdeborová, L. Advances in Neural Information Processing Systems. Vol. 33 (eds Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M. F. & Lin, H.) (Curran Associates, Inc., 2020) pp. 12199â12210.
Mignacco, F., Krzakala, F., Lu, Y. M., Urbani, P. & Zdeborová L. The role of regularization in classification of high-dimensional noisy gaussian mixture, in Proceedings of the 37th International Conference on Machine Learning, ICMLâ20 (JMLR.org, 2022).
Guimerà , R. & Sales-Pardo, M. Missing and spurious interactions and the reconstruction of complex networks. Proc. Natl Acad. Sci. USA 106, 22073â22078 (2009).
Decelle, A., Krzakala, F., Moore, C. & Zdeborová, L. Inference and phase transitions in the detection of modules in sparse networks,. Phys. Rev. Lett. 107, 065701 (2011).
Krzakala, F., Ricci-Tersenghi, F., Zdeborová, L., Zecchina, R., Tramel, E. W. & Cugliandolo, L. F. (eds) Statistical Physics, Optimization, Inference, and Message-Passing Algorithms. 1st ed (Oxford, UK, 2016).
Vallès-Català , T., Peixoto, T. P., Sales-Pardo, M. & Guimerà , R. Consistencies and inconsistencies between model selection and link prediction in networks,. Phys. Rev. E 97, 062316 (2018).
Peixoto, T. P. Advances in Network Clustering And Blockmodeling (John Wiley & Sons Ltd, 2020).
Ando, T. Bayesian Model Selection And Statistical Modeling (CRC Press, 2010).
Mondelli, M. & Montanari, A. M. âM. mondelli, m. montanari,â Foundations of Computational Mathematics. Vol. 19, p. 703â773 (2019)..
Barbier, J., Krzakala, F., Macris, N., Miolane, L. & Zdeborová, L. Optimal errors and phase transitions in high-dimensional generalized linear models,. Proc. Natl Acad. Sci. USA 116, 5451â5460 (2019).
Acknowledgements
This research was funded by project PID2019-106811GB-C31 from the Spanish MCIN/AEI/10.13039/501100011033 and by the Government of Catalonia (2017SGR-896).
Author information
Authors and Affiliations
Contributions
M.S-.P. and R.G. designed the research. O.F.-F., I.R., H.D.L.R., and R.G. wrote code and run computational experiments. O.F.-F., I.R., H.D.L.R., J.D., M.S.-P., and R.G. analyzed data. O.F.-F., I.R., M.S.-P., and R.G. wrote the manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Communications thanks the other anonymous reviewer(s) for their contribution to the peer review of this work. Peer review reports are available.
Additional information
Publisherâs note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the articleâs Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the articleâs Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Fajardo-Fontiveros, O., Reichardt, I., De Los RÃos, H.R. et al. Fundamental limits to learning closed-form mathematical models from data. Nat Commun 14, 1043 (2023). https://doi.org/10.1038/s41467-023-36657-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-023-36657-z
This article is cited by
-
Generative learning for nonlinear dynamics
Nature Reviews Physics (2024)
-
AI-driven research in pure mathematics and theoretical physics
Nature Reviews Physics (2024)
-
Probabilistic grammars for modeling dynamical systems from coarse, noisy, and partial data
Machine Learning (2024)