Reconciling modern machine learning and the bias-variance trade-off

preview_player
Показать описание
It turns out that the classic view of generalization and overfitting is incomplete! If you add parameters beyond the number of points in your dataset, generalization performance might increase again due to the increased smoothness of overparameterized functions.

Abstract:
The question of generalization in machine learning---how algorithms are able to learn predictors from a training sample to make accurate predictions out-of-sample---is revisited in light of the recent breakthroughs in modern machine learning technology.
The classical approach to understanding generalization is based on bias-variance trade-offs, where model complexity is carefully calibrated so that the fit on the training sample reflects performance out-of-sample.
However, it is now common practice to fit highly complex models like deep neural networks to data with (nearly) zero training error, and yet these interpolating predictors are observed to have good out-of-sample accuracy even for noisy data.
How can the classical understanding of generalization be reconciled with these observations from modern machine learning practice?
In this paper, we bridge the two regimes by exhibiting a new "double descent" risk curve that extends the traditional U-shaped bias-variance curve beyond the point of interpolation.
Specifically, the curve shows that as soon as the model complexity is high enough to achieve interpolation on the training sample---a point that we call the "interpolation threshold"---the risk of suitably chosen interpolating predictors from these models can, in fact, be decreasing as the model complexity increases, often below the risk achieved using non-interpolating models.
The double descent risk curve is demonstrated for a broad range of models, including neural networks and random forests, and a mechanism for producing this behavior is posited.

Authors: Mikhail Belkin, Daniel Hsu, Siyuan Ma, Soumik Mandal

Рекомендации по теме
Комментарии
Автор

I struggled with understanding this paper due to lack of knowledge (conceptually spoken), but after seeing your explanation, everything is clear.
thank you very much

PeterJMPuyneers
Автор

Mind blown. Super cool! I have so many tests to rerun with higher parameter count now

AntonCode
Автор

you did a great job. This just left me speechless!!!

MLDawn
Автор

Fantastic video -- thank you! Fascinating...

danielbigham
Автор

Mind blown. Very interesting paper! Does this mean that if you are in the regime where the test loss has started to decrease (as a function of parameters) again and you add more training examples, your test accuracy will get worse because it makes it harder for the optimizer to find a simple function that perfectly mahces the training data? In theory, this could make it beneficial to reduce the number of training examples, but intuitively, that feels wrong.

kristoferkrus
Автор

This is such an amazing study. So many synergies with the Deep Double Descent paper.

sayakpaul
Автор

I started to read this paper during the last days and I confirm that it is really interesting! However, I have some doubts on the way they evaluate the MSE (how do they deal with the fact the function h(x) is complex?) and the zero-one loss/norm of coefficients (since it is a multi-class classification problem, they probably use one-hot encoding, but again how do they deal with the complex h(x)? Moreover, if they use one-hot encoding, the regressor is a 2D matrix, thus what norm are they plotting? L2 norm for matrices?). Did you try to reproduce their plots with the MNIST database? Are these technical passages clear to you? Thank you again for the video!

Fede
Автор

A high-complexity solution be like "Braaaah! Brrraah!" 😂👍

DasGrosseFressen
Автор

This is an interesting paper; I wonder if this applies to boosting/bagging with models that don't have many parameter options like multinomial naive bayes. Would parameter optimization on ensemble models have the same effect when the baseline model within are linear? Interesting option for some testing here.

DrAhdol
Автор

Can you elaborate on the Hilbert space thing? What does Hilbert space to do with neural networks?

herp_derpingson
Автор

Is complexity of H means no of features here ?

ujjwalkar
join shbcf.ru