Hyperparameter Transfer Learning with Adaptive Complexity
dc.contributor.author | Horvath, Samuel | |
dc.contributor.author | Klein, Aaron | |
dc.contributor.author | Richtarik, Peter | |
dc.contributor.author | Archambeau, Cedric | |
dc.date.accessioned | 2021-08-30T06:33:42Z | |
dc.date.available | 2021-03-03T06:45:52Z | |
dc.date.available | 2021-08-30T06:33:42Z | |
dc.date.issued | 2021 | |
dc.identifier.issn | 2640-3498 | |
dc.identifier.uri | http://hdl.handle.net/10754/667827 | |
dc.description.abstract | Bayesian optimization (BO) is a data-efficient approach to automatically tune the hyperparameters of machine learning models. In practice, one frequently has to solve similar hyperparameter tuning problems sequentially. For example, one might have to tune a type of neural network learned across a series of different classification problems. Recent work on multi-task BO exploits knowledge gained from previous hyperparameter tuning tasks to speed up a new tuning task. However, previous approaches do not account for the fact that BO is a sequential decision making procedure. Hence, there is in general a mismatch between the number of evaluations collected in the current tuning task compared to the number of evaluations accumulated in all previously completed tasks. In this work, we enable multi-task BO to compensate for this mismatch, such that the transfer learning procedure is able to handle different data regimes in a principled way. We propose a new multi-task BO method that learns a set of ordered, non-linear basis functions of increasing complexity via nested drop-out and automatic relevance determination. Experiments on a variety of hyperparameter tuning problems show that our method improves the sample efficiency of recently published multi-task BO methods. | |
dc.publisher | MLResearchPress | |
dc.relation.url | http://proceedings.mlr.press/v130/horvath21a.html | |
dc.rights | Copyright 2021 by the author(s). | |
dc.title | Hyperparameter Transfer Learning with Adaptive Complexity | |
dc.type | Conference Paper | |
dc.contributor.department | Computer Science Program | |
dc.contributor.department | Computer, Electrical and Mathematical Science and Engineering (CEMSE) Division | |
dc.contributor.department | Statistics | |
dc.contributor.department | Statistics Program | |
dc.conference.date | APR 13-15, 2021 | |
dc.conference.name | 24th International Conference on Artificial Intelligence and Statistics (AISTATS) | |
dc.conference.location | Virtual | |
dc.identifier.wosut | WOS:000659893801067 | |
dc.eprint.version | Publisher's Version/PDF | |
dc.contributor.institution | Amazon Web Services | |
dc.identifier.volume | 130 | |
dc.identifier.arxivid | 2102.12810 | |
kaust.person | Horvath, Samuel | |
kaust.person | Richtarik, Peter | |
refterms.dateFOA | 2021-03-03T06:48:49Z |
Files in this item
This item appears in the following Collection(s)
-
Conference Papers
-
Computer Science Program
For more information visit: https://cemse.kaust.edu.sa/cs -
Statistics Program
For more information visit: https://stat.kaust.edu.sa/ -
Computer, Electrical and Mathematical Science and Engineering (CEMSE) Division
For more information visit: https://cemse.kaust.edu.sa/