SIAM J. Sci. Comput. | 2021

A supervised learning approach involving active subspaces for an efficient genetic algorithm in high-dimensional optimization problems

 
 
 

Abstract


In this work, we present an extension of the genetic algorithm (GA) which exploits the active subspaces (AS) property to evolve the individuals on a lower dimensional space. In many cases, GA requires in fact more function evaluations than others optimization method to converge to the optimum. Thus, complex and high-dimensional functions may result intractable with the standard algorithm. To address this issue, we propose to linearly map the input parameter space of the original function onto its AS before the evolution, performing the mutation and mate processes in a lower dimensional space. In this contribution, we describe the novel method called ASGA, presenting differences and similarities with the standard GA method. We test the proposed method over n-dimensional benchmark functions -- Rosenbrock, Ackley, Bohachevsky, Rastrigin, Schaffer N. 7, and Zakharov -- and finally we apply it to an aeronautical shape optimization problem.

Volume 43
Pages B831-B853
DOI 10.1137/20M1345219
Language English
Journal SIAM J. Sci. Comput.

Full Text