Please use this identifier to cite or link to this item: https://hdl.handle.net/2440/138713
Citations
Scopus Web of Science® Altmetric
?
?
Full metadata record
DC FieldValueLanguage
dc.contributor.authorBossek, J.-
dc.contributor.authorDoerr, C.-
dc.contributor.authorKerschke, P.-
dc.contributor.authorNeumann, A.-
dc.contributor.authorNeumann, F.-
dc.contributor.editorBäck, T.-
dc.contributor.editorPreuss, M.-
dc.contributor.editorDeutz, A.H.-
dc.contributor.editorWang, H.-
dc.contributor.editorDoerr, C.-
dc.contributor.editorEmmerich, M.T.M.-
dc.contributor.editorTrautmann, H.-
dc.date.issued2020-
dc.identifier.citationLecture Notes in Artificial Intelligence, 2020 / Bäck, T., Preuss, M., Deutz, A.H., Wang, H., Doerr, C., Emmerich, M.T.M., Trautmann, H. (ed./s), vol.12269, pp.111-124-
dc.identifier.isbn9783030581114-
dc.identifier.issn0302-9743-
dc.identifier.issn1611-3349-
dc.identifier.urihttps://hdl.handle.net/2440/138713-
dc.description.abstractOne-shot optimization tasks require to determine the set of solution candidates prior to their evaluation, i.e., without possibility for adaptive sampling. We consider two variants, classic one-shot optimization (where our aim is to find at least one solution of high quality) and one-shot regression (where the goal is to fit a model that resembles the true problem as well as possible). For both tasks it seems intuitive that well-distributed samples should perform better than uniform or grid-based samples, since they show a better coverage of the decision space. In practice, quasi-random designs such as Latin Hypercube Samples and low-discrepancy point sets are indeed very commonly used designs for one-shot optimization tasks. We study in this work how well low star discrepancy correlates with performance in one-shot optimization. Our results confirm an advantage of low-discrepancy designs, but also indicate the correlation between discrepancy values and overall performance is rather weak. We then demonstrate that commonly used designs may be far from optimal. More precisely, we evolve 24 very specific designs that each achieve good performance on one of our benchmark problems. Interestingly, we find that these specifically designed samples yield surprisingly good performance across the whole benchmark set. Our results therefore give strong indication that significant performance gains over state-of-the-art one-shot sampling techniques are possible, and that evolutionary algorithms can be an efficient means to evolve these.-
dc.description.statementofresponsibilityJakob Bossek, Carola Doerr, Pascal Kerschke, Aneta Neumann, and Frank Neumann-
dc.language.isoen-
dc.publisherSpringer-
dc.relation.ispartofseriesLecture Notes in Computer Science; 12269-
dc.rights© Springer Nature Switzerland AG 2020-
dc.source.urihttps://doi.org/10.1007/978-3-030-58112-1-
dc.subjectOne-shot optimization; Regression; Fully parallel search; Surrogate-assisted optimization; Continuous optimization-
dc.titleEvolving Sampling Strategies for One-Shot Optimization Tasks-
dc.typeConference paper-
dc.contributor.conference16th International Conference on Parallel Problem Solving from Nature (PPSN) (5 Sep 2020 - 9 Sep 2020 : Leiden, The Netherlands)-
dc.identifier.doi10.1007/978-3-030-58112-1_8-
dc.publisher.placeCham, Switzerland-
dc.relation.granthttp://purl.org/au-research/grants/arc/DP190103894-
pubs.publication-statusPublished-
dc.identifier.orcidBossek, J. [0000-0002-4121-4668]-
dc.identifier.orcidNeumann, A. [0000-0002-0036-4782]-
dc.identifier.orcidNeumann, F. [0000-0002-2721-3618]-
Appears in Collections:Computer Science publications

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.