Skip to content

TBU-AILab/LLMdesignedEA-comp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Competition on LLM designed EA

We are organizing a competition centered on the innovative use of large language models (LLMs) to design evolutionary algorithms (EAs). This contest aims to explore the potential of LLMs in creating sophisticated EAs that can tackle complex optimization problems. Join us in contributing to this emerging field, showcasing how LLMs can enhance and accelerate the development of evolutionary algorithms.

Ongoing competitions

Past competitions

Deadlines

  • Submission deadline: 30th June 2025
  • Notification of acceptance: TBA
  • Early registration deadline: TBA

Benchmark definition

We will be using GNBG benchmark for box-constrained numerical global optimization A. H. Gandomi, M. N. Omidvar, R. Salgotra, and K. Deb, "A Generalized and Configurable Benchmark Generator for Continuous Unconstrained Numerical Optimization," arXiv preprint arXiv:2312.07083, 2023. There are 24 test functions of varying dimensions and varying problem landscapes.

The benchmark is provided in 3 languages

How to participate

Submit your results in the following form. You have to submit the following items:

  • Authors
  • Title of your algorithm
  • ZIP file with results in the format specified below
  • The used LLM with specified settings (e.g. OpenAI GPT 4 turbo, temperature = 0.8)
  • Algorithm code (for validity check)
  • Generating prompts (how did you arrive to the solution)

Results format

You should run 31 independent runs of each test function with a stopping criterion defined by either maximum number of function evaluations (500,000 for f1-f15 and 1,000,000 for f16-f24) or reaching the acceptance threshold set to an absolute error (difference between best-found and optimum value) = 10^-8.

The results will be reported for each function by a tuple of text files - f_x_value.txt and f_x_param.txt (where x denotes the function number). The format of the f_x_value.txt should be:

  • 31 best-found values - one for each run on a separate line
  • example in f_x_value.txt

The format of the f_x_params.txt should be:

  • 31 parameter vectors of best-found solutions (dimensions separated by a comma ,) - one for each run on a separate line
  • example for a 5D problem in f_x_params.txt

Rules

  • A LLM has to be used during the design process of the algorithm
  • You can fine-tune the algorithm parameters but the values have to be consistent across the whole benchmark
  • No modification of the benchmark is allowed
  • The benchmark has to be treated as a blackbox

Evaluation criteria

The algorithms will be ranked on the basis of their mean value on each function.

  • The mean will be computed from all 31 runs
  • The score range on each function is 0 to 1 (worst to best)
  • The scores will be proportional
  • Maximum score is therefore 24 for the algorithm that would perform the best on each benchmark function
  • Maximum considered difference between algorithm performance is 10^-8
  • Runs with parameters out of range or better than optimum value will result in score 0 on that function.

Competition results

The competition results and certificates are now available in the competition_results subfolder

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published