filmov
tv
Sparse Neural Networks: From Practice to Theory
Показать описание
Atlas Wang
Assistant Professor, Electrical and Computer Engineering
The University of Texas at Austin
Abstract: A sparse neural network (NN) has most of its parameters set to zero and is traditionally considered as the product of NN compression (i.e., pruning). Yet recently, sparsity has exposed itself as an important bridge for modeling the underlying low dimensionality of NNs, for understanding their generalization, optimization dynamics, implicit regularization, expressivity, and robustness. Deep NNs learned with sparsity-aware priors have also demonstrated significantly improved performances through a full stack of applied work on algorithms, systems, and hardware. In this talk, I plan to cover some of our recent progress on the practical, theoretical, and scientific aspects of sparse NNs. I will try scratching the surface of three aspects: (1) practically, why one should love a sparse NN, beyond just a post-training NN compression tool; (2) theoretically, what are some guarantees that one can expect from sparse NNs; and (3) what is future prospect of exploiting sparsity.
Assistant Professor, Electrical and Computer Engineering
The University of Texas at Austin
Abstract: A sparse neural network (NN) has most of its parameters set to zero and is traditionally considered as the product of NN compression (i.e., pruning). Yet recently, sparsity has exposed itself as an important bridge for modeling the underlying low dimensionality of NNs, for understanding their generalization, optimization dynamics, implicit regularization, expressivity, and robustness. Deep NNs learned with sparsity-aware priors have also demonstrated significantly improved performances through a full stack of applied work on algorithms, systems, and hardware. In this talk, I plan to cover some of our recent progress on the practical, theoretical, and scientific aspects of sparse NNs. I will try scratching the surface of three aspects: (1) practically, why one should love a sparse NN, beyond just a post-training NN compression tool; (2) theoretically, what are some guarantees that one can expect from sparse NNs; and (3) what is future prospect of exploiting sparsity.
Sparse Neural Networks: From Practice to Theory
Sparse Training of Neural Networks Using AC/DC
[AUTOML23] Learning Activation Functions for Sparse Neural Networks Teaser
SPARSE: a sparse hypergraph neural network for... - Duc Anh Nguyen - MLCSB - Proceedings - ISMB 2022
Joan Bruna: 'On Sparse Linear Programming and (simple) neural networks'
Intro to Sparse Tensors and Spatially Sparse Neural Networks
Michael Elad - Sparse Modelling of Data and its Relation to Deep Learning
Introduction | Sparse Learning in Neural Networks | CVPR'22 Tutorial
Numenta Demonstrates 50x Speed-Up on Deep Learning Networks Using Brain-Derived Sparse Algorithms
SPARSE CONVOLUTIONAL NEURAL NETWORKS
Neural Networks explained in 60 seconds!
How to Design a Neural Network | 2020 Edition
Neural networks [8.6] : Sparse coding - online dictionary learning algorithm
The Sparse Manifold Transform
Learning From Incomplete Features by Simultaneous Training of Neural Networks and Sparse Coding
Sparse connectivity in convolutional layers of a neural network
SuperLU and STRUMPACK Sparse Direct Solver and Preconditioner | X. Sherry Li, LBNL
Predictive Uncertainty Estimation of Neural Networks with Sparse Gaussian Processes
Neural networks [8.7] : Sparse coding - ZCA preprocessing
The Dark Matter of AI [Mechanistic Interpretability]
Sparse Activation- Game-changer for the Future of Deep Learning. Devansh Machine Learning Techniques
ICAI MMLL webinar IV: From Sparse Modeling to Sparse Communication
Sparse Expert Models: Past and Future
[ECCV 2020] Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution
Комментарии