A simplified implementation of Neuro-evolution of Augmenting Topologies, a novel technique for neuro-evolution developed by Kenneth O. Stanley.
Another implementation of NEAT in Python already exists (by CodeReclaimers, here), which is thorough and really nice work. However, as things stand, it faces some deeper issues and thus doesn't perform in accordance with the benchmarks provided in the original paper, and is no longer maintained. This solution has cut down to the bone in an attempt to simplify both usage and the codebase, and to achieve the expected results. This implementation is not as efficient as it could be - if you want to pick up some work, there's lots to be done on the performance side of this project.
In the simplest case, all you need to begin training a neural network for any given problem is a fitness function.
1. Install the package
$ pip install sneat
2. Set up your fitness function
Your fitness function should take a genome and output a fitness score based on how well that genome solves the task. Here's a simple example, training a neural network that will output the sum of its inputs:
def fitness_function(genome):
inputs = list(np.random.randint(5, size=2))
target = sum(inputs)
# feed input to the genomes neural network, return its output
output = genome.activate(inputs)
difference = (output - target) ** 2
fitness = 1 / difference
return fitness
3. Magic
There's a bunch of hyperparameters that can (should) be configured for your given problem, but again we'll take a simple approach and just use the default hyperparameters along with the default evolution loop:
from sneat import evolve
def fitness_function(genome):
...
winner = evolve(fitness_function)
...now watch the generations unfold and enjoy! If you let it run long enough, we might get to experience our very own doomsday. The evolve
function outputs the winning genome when one of the following three conditions are fulfilled:
- The fitness threshold is reached by the best genome in the population
- The maximum number of generations are reached
- The user cancels (
CTRL+C
)
A default configuration file is supplied, but you'll probably want to change some details (such as the number of input and output nodes in the networks). You can include as few or as many configuration elements as you want; those you don't provide will fall back to the defaults.
Create a config.ini
file in your working directory with the settings you want to change. Here's the default config file for inspiration:
[NeuralNetwork]
num_inputs = 2
num_outputs = 1
input_activation = linear
output_activation = sigmoid
use_normalizer = False
[Population]
population_size = 150
compatibility_threshold = 3.0
min_species_size = 5
elite_size = 3
survival_threshold = 0.2
[MutationRates]
add_node=0.1
add_connection=0.2
toggle_connection=0.08
change_weight=0.65
change_activation=0.05
change_bias=0.05
remove_node=0.08
[Evolution]
max_generations = 100 # set to 0 to disable
max_fitness = 4 # set to 0 to disable
If you want to have more control over the whole loop (for custom reporting, for example), I'd suggest importing the Population
class and working around that. This class has .reproduce()
, which will perform selection, cross-over and mutation on all genomes based on their fitness values. Finally, it will properly speciate the new genomes and move on to the next generation.
Population.species
is a list containing all the species, which in turn offers Species.genomes
. I'll let you figure out the rest - the code is pretty straight-forward.
Some examples are included in this repo, using Gymnasium. To run them, first install the examples dependencies:
pip install sneat[examples]
You can then run one of the examples to begin training; fx:
python examples/lunar_lander/lunar_lander.py
The winning genome will be saved as winner.pkl
. If you want to see it in action, run the example with the name of the saved genome as the first argument:
python examples/lunar_lander/lunar_lander.py winner.pkl
Cart Pole always solves in less than 10 generations. Lunar Lander consistently solves in under 100 generations. Bipedal Walker takes about 1500 generations (but it seems to vary a lot with these more difficult tasks). I've never been able to solve either of the last two environments with other open-source NEAT implementations, even though the original paper indicates that tasks like these should be possible to solve - and indeed they are.