Question about parallel evaluation #444
Replies: 3 comments 1 reply
-
|
When The AlgoTune blog showed serial (1) was catastrophically bad vs parallel (16), but that doesn't mean more parallelism is always better. For most examples we run with For the Erdős problem with Qwen3-8B, may be the mathematical refinements need that tighter feedback loop. Try |
Beta Was this translation helpful? Give feedback.
-
|
@codelion Thanks for your kind response. By the way, under the |
Beta Was this translation helpful? Give feedback.
-
|
On individual tasks there is variance in model responses. May have to test on a number of examples to say if a model is really better than the other. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I have one question. When I run Qwen3-8B with OpenEvolve on the Erdős problem, the performance varies depending on the parallel evaluation configuration. For example, when I set the parallel evaluation to 1, it shows 0.3810, but when I set it to 16, it shows 0.495 (which is worse).
In the AlgoTune task, your team already demonstrated that parallel evaluation performs much better than sequential evaluation. In my case, however, I am trying to understand why the opposite trend appears. Do you have any comments on this?
Beta Was this translation helpful? Give feedback.
All reactions