filmov
tv
Regularisation of Neural Networks by Enforcing Lipschitz Continuity
Показать описание
In this video we continue on the topic of Lipschitz continuity by presenting a paper which proposes a projection method to enforce it!
Abstract: We investigate the effect of explicitly enforcing the Lipschitz continuity of neural networks with respect to their inputs. To this end, we provide a simple technique for computing an upper bound to the Lipschitz constant---for multiple p-norms---of a feed forward neural network composed of commonly used layer types. Our technique is then used to formulate training a neural network with a bounded Lipschitz constant as a constrained optimisation problem that can be solved using projected stochastic gradient methods. Our evaluation study shows that the performance of the resulting models exceeds that of models trained with other common regularisers. We also provide evidence that the hyperparameters are intuitive to tune, demonstrate how the choice of norm for computing the Lipschitz constant impacts the resulting model, and show that the performance gains provided by our method are particularly noticeable when only a small amount of training data is available.
Abstract: We investigate the effect of explicitly enforcing the Lipschitz continuity of neural networks with respect to their inputs. To this end, we provide a simple technique for computing an upper bound to the Lipschitz constant---for multiple p-norms---of a feed forward neural network composed of commonly used layer types. Our technique is then used to formulate training a neural network with a bounded Lipschitz constant as a constrained optimisation problem that can be solved using projected stochastic gradient methods. Our evaluation study shows that the performance of the resulting models exceeds that of models trained with other common regularisers. We also provide evidence that the hyperparameters are intuitive to tune, demonstrate how the choice of norm for computing the Lipschitz constant impacts the resulting model, and show that the performance gains provided by our method are particularly noticeable when only a small amount of training data is available.
Regularization in a Neural Network | Dealing with overfitting
Regularisation of Neural Networks by Enforcing Lipschitz Continuity
Why Regularization Reduces Overfitting (C2W1L05)
CS 152 NN—12: Regularization: Dropout
How to Implement Regularization on Neural Networks
Tutorial 9- Drop Out Layers in Multi Neural Network
Regularization (C2W1L04)
NN - 16 - L2 Regularization / Weight Decay (Theory + @PyTorch code)
L10.0 Regularization Methods for Neural Networks -- Lecture Overview
Regularization of Big Neural Networks
Deep Neural Network Regularization - Part 1
Regularization in Neural Network || Lesson 20 || Deep Learning || Learning Monkey ||
Dropout Regularization (C2W1L06)
Regularization in a Neural Network explained
What is Dropout Regularization | How is it different?
L10.4 L2 Regularization for Neural Nets
CS 152 NN—6: Regularization—Neural-network-specific
Why Does Regularization Reduce Overfitting in Deep Neural Networks?
Regularization in machine learning | L1 and L2 Regularization | Lasso and Ridge Regression
Lipschitz Regularization of Neural Networks - Intriguing Properties of Neural Networks
Dropout Regularization
Regularization - Explained!
Deep Learning - Lecture 5.4 (Regularization: Dropout)
Dropout Regularization | Deep Learning Tutorial 20 (Tensorflow2.0, Keras & Python)
Комментарии