filmov
tv
Explaining Hyperparameter Optimization via Partial Dependence Plots (NeurIPS'21)
Показать описание
Automated hyperparameter optimization (HPO) can support practitioners to obtain peak performance in machine learning models. However, there is often a lack of valuable insights into the effects of different hyperparameters on the final model performance. This lack of explainability makes it difficult to trust and understand the automated HPO process and its results.
We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO).
BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the \emph{partial dependence plot} (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions.
We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO).
BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the \emph{partial dependence plot} (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions.
Explaining Hyperparameter Optimization via Partial Dependence Plots (NeurIPS'21)
Practical approaches for efficient hyperparameter optimization with Oríon | SciPy 2021
Better and Faster Hyper Parameter Optimization with Dask | SciPy 2019 | Scott Sievert
Massively Parallel Hyperparameter Tuning
Hyperparameter Optimization: This Tutorial Is All You Need
Self-Tuning Networks: Amortizing the Hypergradient Computation for Hyperparameter Optimization
Martin Wistuba: Hyperparameter optimization for the impatient
Practical Approaches for Efficient Hyperparameter Optimization
Hyperparameter Optimization Using Dask with ORÍON | Xavier Bouthillier | Dask Summit 2021
The Kernel Trick in Support Vector Machine (SVM)
MedAI #37: Federated Hyperparameters Tuning: Challenges, Baselines & Connections | Mikhail Khoda...
Hyperparameter Tuning with W&B Sweeps
Efficient Distributed Hyperparameter Tuning with Apache Spark
PyData Madison Meetup: Better and Faster Hyperparameter Optimization with Dask
Deep Learning Tutorial - Lesson 07 - Coding Example - Model Optimization II: Hyperparameter-Tuning
Sherpa Tutorial (partial) at Chandra Data Science
Hyperparameter Optimization with Ray Tune
Learn Particle Swarm Optimization (PSO) in 20 minutes
10c Machine Learning: Optimization Basics
Hyperparameter tuning using Optuna | Climate Data Science lecture 30
Hyperparameter Tuning
Resource-Efficient Deep Learning Model Selection on Apache Spark
Regularizing Trajectory Optimization with Denoising Autoencoders (Paper Explained)
Interpretable Machine Learning - Feature Effects - Partial Dependence (PD) Plot
Комментарии