Skip to content

Latest commit

 

History

History
117 lines (74 loc) · 3.77 KB

strategies.rst

File metadata and controls

117 lines (74 loc) · 3.77 KB

Strategies

Summit has several machine learning strategies available for optimisation, as well as some more naive ones.

All strategies have a similar API. They are instantiated by passing in a ~summit.domain.Domain. New reaction conditions are requested using the suggest_experiments method which, potentially, takes results from previous reactions.

Bayesian Optimisation

Bayesian optimisation (BO) is an efficient way to optimise a wide variety of functions, inculding chemical reactions. In BO, you begin by specifying some prior beliefs about your functions. In many cases, we start with an assumption that we know very little. Then, we create a probabilistic model that incorporates this prior belief and some data (i.e, reactions at different conditions), called a posterior. In reaction optimisation, this model will predict the value of an objective (e.g., yield) at particular reaction conditions. One key factor is that these models are probabalistic, so they do not give precise predictions but instead a distribution that is sampled.

With the updated model, we use one of two classes of techniques to select our next experiments. Some BO strategies optimise an acquisition function, which is a function that takes in the model parameters and some suggested next experiement and predicts the quality of that experiment. Alternatively, a deterministic function can be sampled from the model, which is then optimised.

Illustration of how acquisition functions eanble BO strategies to reduce uncertainty and maximise objective simulataneously. Dotted line is actual objective and solid line is posterior of surrogate model. Acquisition function is high where objective to be optimal (exploration) and where there is high uncertainty (exploitation). Adapted from Shahriari et al.

Illustration of how acquisition functions eanble BO strategies to reduce uncertainty and maximise objective simulataneously. Dotted line is actual objective and solid line is posterior of surrogate model. Acquisition function is high where objective to be optimal (exploration) and where there is high uncertainty (exploitation). Adapted from Shahriari et al.

To learn more about BO, we suggest reading the review by Shahriari et al.

The BO strategies available in Summit are:

TSEMO

summit.strategies.tsemo.TSEMO

SOBO

summit.strategies.sobo.SOBO

MTBO

summit.strategies.MTBO

ENTMOOT

summit.strategies.ENTMOOT

Reinforcement Learning

Reinforcement learning (RL) is distinct because it focuses on creating a custom policy for a particular problem instead of a model of the problem. In the case of reaction optimisation, the policy directly predicts the next experiment(s) should be given a history of past experiments. Policies are trained to maximise some sort of reward, such as achieving the maximum number of yield in as few experiments possible.

For more information about RL, see the book by Sutton and Barto or David Silver's course.

summit.strategies.deep_reaction_optimizer.DRO

Simplex

summit.strategies.neldermead.NelderMead

Random

Random

summit.strategies.random.Random

Latin Hypercube Sampling

summit.strategies.random.LHS

Other

SNOBFIT

summit.strategies.snobfit.SNOBFIT

Full Factorial

summit.strategies.factorial_doe.FullFactorial