Abstract
In the field of neuroevolution (NE), evolutionary algorithms are used to update the weights, biases and topologies of artificial neural networks (ANNs). A recent theoretical work presented the first runtime analysis of NE in a simple setting, considering a single neuron and intuitive benchmark function classes. However, this work was limited by the unrealistic settings with regard to activation functions and fitness measurements.In this paper, we extend upon this first work by overcoming the two shortcomings. Firstly, we consider a more realistic setting in which the NE also evolves a third parameter, termed the bend, allowing the previous benchmark function classes to be solved efficiently even in the fixed bias case. This setting mimics rectified linear unit activation functions, which are common in real-world applications of ANNs. Secondly, we consider a dynamic fitness function evaluation paradigm where the weights and biases are updated after each new sample. Experimental results in both cases support the presented theoretical results.
Original language | English |
---|---|
Title of host publication | Proceedings of the Genetic and Evolutionary Computation Conference, GECCO '24 |
Publisher | Association for Computing Machinery |
Publication date | 2024 |
Pages | 1560-1568 |
ISBN (Electronic) | 79-8-4007-0494-9/24/07 |
DOIs | |
Publication status | Published - 2024 |
Event | 2024 Genetic and Evolutionary Computation Conference - Melbourne, Australia Duration: 14 Jul 2024 → 18 Jul 2024 |
Conference
Conference | 2024 Genetic and Evolutionary Computation Conference |
---|---|
Country/Territory | Australia |
City | Melbourne |
Period | 14/07/2024 → 18/07/2024 |
Keywords
- Dynamic fitness
- Neuroevolution
- Runtime analysis
- Theory