Repository logo
  • Log In
    Log in via Symplectic to deposit your publication(s).
Repository logo
  • Communities & Collections
  • Research Outputs
  • Statistics
  • Log In
    Log in via Symplectic to deposit your publication(s).
  1. Home
  2. Faculty of Engineering
  3. Computing
  4. Computing
  5. Improving the data efficiency of multi-objective quality-diversity through gradient assistance and crowding exploration
 
  • Details
Improving the data efficiency of multi-objective quality-diversity through gradient assistance and crowding exploration
File(s)
pap516s3-file2.pdf (2.02 MB)
Supporting information
3583131.3590470.pdf (1.91 MB)
Published version
Author(s)
Janmohamed, Hannah
Pierrot, Thomas
Cully, Antoine
Type
Conference Paper
Abstract
Quality-Diversity (QD) algorithms have recently gained traction as optimisation methods due to their effectiveness at escaping local optima and capability of generating wide-ranging and high-performing solutions. Recently, Multi-Objective MAP-Elites (MOME) extended the QD paradigm to the multi-objective setting by maintaining a Pareto front in each cell of a MAP-ELITES grid. MOME achieved a global performance that competed with NSGA-11 and SPEA2, two well-established multi-objective evolutionary algorithms, while also acquiring a diverse repertoire of solutions. However, MOME is limited by non-directed genetic search mechanisms which struggle in high-dimensional search spaces. In this work, we present Multi-Objective MAP-Elites with Policy-Gradient Assistance and Crowding-based Exploration (MOME-PGX: a new QD algorithm that extends MOME to improve its data efficiency and performance. MOME-PGX uses gradient-based optimisation to efficiently drive solutions towards higher performance. It also introduces crowding-based mechanisms to create an improved exploration strategy and to encourage greater uniformity across Pareto fronts. We evaluate MOME-PGX in four simulated robot locomotion tasks and demonstrate that it converges faster and to a higher performance than all other baselines. We show that MOME-PGX is between 4.3 and 42 times more data-efficient than MOME and doubles the performance of MOME, NSGA-11 and SPEA2 in challenging environments.
Date Issued
2023-07-12
Date Acceptance
2023-03-31
Citation
GECCO '23: Proceedings of the Genetic and Evolutionary Computation Conference, 2023, pp.165-173
URI
http://hdl.handle.net/10044/1/104027
DOI
https://www.dx.doi.org/10.1145/3583131.3590470
ISBN
9798400701191
Publisher
ACM
Start Page
165
End Page
173
Journal / Book Title
GECCO '23: Proceedings of the Genetic and Evolutionary Computation Conference
Copyright Statement
© 2023 Copyright held by the owner/author(s). This work is licensed under a Creative Commons Attribution‐NonCommercial‐
ShareAlike International 4.0 License (https://creativecommons.org/licenses/by-nc-sa/4.0/)
License URL
https://creativecommons.org/licenses/by-nc-sa/4.0/
Source
GECCO 2023
Publication Status
Published
Start Date
2023-07-15
Finish Date
2023-07-19
Coverage Spatial
Lisbon, Portugal
About
Spiral Depositing with Spiral Publishing with Spiral Symplectic
Contact us
Open access team Report an issue
Other Services
Scholarly Communications Library Services
logo

Imperial College London

South Kensington Campus

London SW7 2AZ, UK

tel: +44 (0)20 7589 5111

Accessibility Modern slavery statement Cookie Policy

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Cookie settings
  • Privacy policy
  • End User Agreement
  • Send Feedback