Publication:
Evolutionary Algorithms for Hyperparameter Search in Machine Learning

dc.contributor.advisor Bain, Michael
dc.contributor.advisor Wobcke, Wayne
dc.contributor.author Bai, Yu
dc.date.accessioned 2022-02-17T04:08:15Z
dc.date.available 2022-02-17T04:08:15Z
dc.date.issued 2021
dc.description.abstract Machine learning algorithms usually have a number of hyperparameters. The choice of values for these hyperparameters may have a significant impact on the performance of an algorithm. In practice, for most learning algorithms the hyperparameter values are determined empirically, typically by search. From the research that has been done in this area, approaches for automating the search of hyperparameters mainly fall into the following categories: manual search, greedy search, random search, Bayesian model-based optimization, and evolutionary algorithm-based search. However, all these approaches have drawbacks — for example, manual and random search methods are undirected, greedy search is very inefficient, Bayesian model-based optimization is complicated and performs poorly with large numbers of hyperparameters, and classic evolutionary algorithm-based search can be very slow and risks falling into local optima. In this thesis we introduce three improved evolutionary algorithms applied to search for high-performing hyperparameter values for different learning algorithms. The first, named EWLNB, combines Naive Bayes and lazy instance-weighted learning. The second, EMLNB, extends this approach to multiple label classification. Finally, we further develop similar methods in an algorithm, named SEODP, for optimizing hyperparameters of deep networks, and report its usefulness on a real-world application of machine learning for philanthropy. EWLNB is a differential evolutionary algorithm which can automatically adapt to different datasets without human intervention by searching for the best hyperparameters for the models based on the characteristics of the datasets to which it is applied. To validate the EWLNB algorithm, we first use it to optimize two key parameters for a locally-weighted Naive Bayes model. Experimental evaluation of this approach on 56 of the benchmark UCI machine learning datasets demonstrate that EWLNB significantly outperforms Naive Bayes as well as several other improved versions of the Naive Bayes algorithms both in terms of classification accuracy and class probability estimation. We then extend the EWLNB approach in the form of the Evolutionary Multi-label Lazy Naive Bayes (EMLNB) algorithm to enable hyperparameter search for multi-label classification problems. Lastly, we revise the above algorithms to propose a method, SEODP, for optimizing deep learning (DL) architecture and hyperparameters. SEODP uses a semi-evolutionary and semi-random approach to search for hyperparameter values, which is designed to evolve a solution automatically over different datasets. SEODP is much faster than other methods, and can adaptively determine different deep network architectures automatically. Experimental results show that compared with manual search, SEODP is much more effective, and compared with grid search, SEODP can achieve optimal performance using only approximately 2% of the running time of greedy search. We also use SEODP on a real-world social-behavioral dataset from a charity organization for a philanthropy application. This dataset contains comprehensive real-time attributes on potential indicators for candidates to be donors. The results show that SEODP is a promising approach for optimizing deep network (DN) architectures over different types of datasets, including a real-world dataset. In summary, the results in this thesis indicate that our methods address the main drawback of evolutionary algorithms, which is the convergence time, and show experimentally that evolutionary-based algorithms can achieve good results in optimizing the hyperparameters for a range of different machine learning algorithms.
dc.identifier.uri http://hdl.handle.net/1959.4/100092
dc.language English
dc.language.iso en
dc.publisher UNSW, Sydney
dc.rights CC BY 4.0
dc.rights.uri https://creativecommons.org/licenses/by/4.0/
dc.subject.other Hyperparameters Optimizing
dc.subject.other Deep Learning
dc.subject.other Evolutionary Algorithms
dc.title Evolutionary Algorithms for Hyperparameter Search in Machine Learning
dc.type Thesis
dcterms.accessRights open access
dcterms.rightsHolder Bai, Yu
dspace.entity.type Publication
unsw.accessRights.uri https://purl.org/coar/access_right/c_abf2
unsw.identifier.doi https://doi.org/10.26190/unsworks/2002
unsw.isDatasetRelatedToPublication https://archive.ics.uci.edu/ml/datasets.php
unsw.relation.faculty Engineering
unsw.relation.school School of Computer Science and Engineering
unsw.relation.school School of Computer Science and Engineering
unsw.relation.school School of Computer Science and Engineering
unsw.subject.fieldofresearchcode 4611 Machine learning
unsw.thesis.degreetype PhD Doctorate
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
public version.pdf
Size:
2.58 MB
Format:
application/pdf
Description:
Resource type