logo
ResearchBunny Logo
Abstract
This paper investigates the fundamental limits of learning closed-form mathematical models from finite, noisy datasets. The authors demonstrate a phase transition in model learning, from a low-noise phase where the true model is learnable to a high-noise phase where it is not. Probabilistic model selection proves quasi-optimal for generalization in both phases, unlike standard machine learning approaches which are limited by interpolation in the low-noise phase. The transition region presents a challenge for all methods.
Publisher
Nature Communications
Published On
Feb 24, 2023
Authors
Oscar Fajardo-Fontiveros, Ignasi Reichardt, Harry R. De Los Ríos, Jordi Duch, Marta Sales-Pardo, Roger Guimerà
Tags
mathematical models
noise
model learning
probabilistic model selection
phase transition
generalization
Listen, Learn & Level Up
Over 10,000 hours of research content in 25+ fields, available in 12+ languages.
No more digging through PDFs—just hit play and absorb the world's latest research in your language, on your time.
listen to research audio papers with researchbunny