Clarissa. Y. P. Wilding*a,
Richard. A. Bourne
a and
Nicholas. J. Warren
*b
aSchool of Chemical Process Engineering, University of Leeds, Woodhouse Lane, LS2 9JT, UK. E-mail: C.Y.Wilding@leeds.ac.uk
bSchool of Chemical, Materials and Biological Engineering, University of Sheffield, Mappin Street, Sheffield, S1 3JD, UK. E-mail: nicholas.warren@sheffield.ac.uk
First published on 26th August 2025
Discovery of sustainable, high-performing materials on timescales to meet societal needs is only going to be achieved with the assistance of artificial intelligence and machine learning. Herein, a Bayesian optimisation algorithm is trained using in silico reactions facilitated by a new mechanistic model for reversible addition fragmentation chain transfer polymerisation (RAFT). This subsequently informs experimental multi-objective self-optimisation of RAFT polymerisation using an automated polymerisation platform capable of measuring the critical algorithm objectives (monomer conversion and molecular weight distribution) online. The platform autonomously identifies the Pareto-front representing the trade-off between monomer conversion and molar mass dispersity with a reduced number of reactions compared to the equivalent fully experimental optimisation process. This model-informed AI approach provides opportunities for much more sustainable and efficient discovery of polymeric materials and provides a benchmark for other complex chemical systems.
While small-molecule chemistry is inherently complex, polymer chemistry presents different challenges; scientists must also navigate the complexities of molecular weight distribution, monomer composition, and polymer architecture (e.g. block vs. statistical copolymers).7 These additional challenges make the implementation of AI (Artificially Intelligent) enabled platforms in polymer chemistry a formidable and intricate frontier.
Even though polymers play an immeasurable role in daily life from applications in drug delivery, to additives in nearly all liquid formulations, it is doubtful that the materials, and their manufacturing routes are truly optimised.8
Gaining a precise understanding of structure–property relationships in tandem with fine control over materials structure using AI enabled reactors has the potential to truly optimise material performance and unlock applications that are currently beyond our reach. Achieving these goals first requires the ability to optimise the synthetic process, which from the chemistry perspective has been achieved through much more accessible controlled polymerisation reactions.9 The most important class of these are reversible de-activation polymerisation (RDRP) techniques, including reversible addition fragmentation chain transfer (RAFT) polymerisation, atom transfer radical polymerisation (ATRP) and nitroxide mediated polymerisation (NMP).10–12 These chemistries have incrementally evolved since conception, including reduction in catalyst concentrations,13 removal of unwanted components (e.g. metals14 or sulfur15) and use of more efficient and sustainable photochemical,16–18 electrochemical19,20 and sonochemical21 initiation methods.
The optimisation of these processes still relies significantly on dated laboratory techniques, including one-factor at a time (OFAT)22 and design of experiments (DoE) approaches to screen conditions.23–25 These techniques can lead to local optima22 and increase experimental overheads of inherently complex reactions such as polymerisation.22,26 The polymer chemistry community is beginning to exploit an array of technologies to accelerate innovation.27 This includes developments in the areas of machine learning,6,26 reactor design,25,28 online monitoring,29–31 and computer control32–35 which have the potential to facilitate the next era of polymeric materials.
Automated experimental flow chemistry platforms have recently been applied to RAFT polymerisation, including the ability to facilitate programmable targeting of molecular weight and/or conversion, and to conduct automated kinetic screens.33,36 Each of these approaches required integration of gel permeation chromatography (GPC) and/or benchtop nuclear magnetic resonance spectroscopy (NMR) as measurement tools. Warren and co-workers also exemplified the first use of these for closed loop multi-objective Bayesian optimisation (BO) of RAFT polymerisation.34 Within this work, it was shown that the Thompson sampling by efficient multi-objective optimisation algorithm (TS-EMO), could efficiently map out a trade-off between monomer conversion (α%; measured by NMR) and molar mass dispersity (Đ; measured by GPC) in a 2D parameter space where residence time and temperature were the input variables. Further examples included its use for polymerisation-induced self-assembly37 and aqueous emulsion polymerisation.38 Importantly, these examples report experimental platforms that required no prior knowledge of the chemistry, other than defining the bounds of reaction space.
In polymer chemistry, since kinetics are relatively well studied, and rate coefficients for the important steps are well reported in the literature,39–42 the availability of such information is beginning to reduce reliance on real experiments, with so-called in silico experiments providing a wealth of information which can be used for better defining parameter space. For example, Kandelhard et al.25 used commercially available modelling software (PREDICI and COMSOL) to optimise different reactors and conversion in silico which were then successfully validated using benchtop NMR. The caveat here was that it required decisions to be made by a trained human operator despite there often being an array of models to predict reaction outcomes. Several models for predicting the kinetics and molar mass dispersity of RDRP systems have been reported,43–45 including a combined kinetic and dispersity model for RAFT polymerisation. This relies on input of known propagation, chain-transfer, and termination constants, kp, ktr, and kt for specific monomers and has been made available as a software package to facilitate in silico RAFT polymerisation. Furthermore, it allows for augmenting the effects of residence time distribution for modelling RAFT polymerisation in a flow reactor. This provided the optimal approach for validation through dissipating exotherms (typically observed in batch) which can cause unwanted rate acceleration.45 The emergence of these technologies has now set the scene for a model informed approach to conduct more sustainable machine learning directed polymer synthesis.
Herein, we demonstrate the first example of a pre-informed self-optimising polymerisation platform capable of exploring 3D parameter space facilitated by in silico modelling. Reaction optimisation, specifically multi-objective problems, can be considered a ‘black-box’ function as the objective function is unknown, complex or expensive to evaluate, meaning only a surrogate model based on input variable data and output analytical data is feasible to guide the optimisation.46,47 As with our previous work,45 all aspects of a closed-loop reactor platform were interfaced with the BO algorithm TS-EMO to enable the reciprocal exchange of data and control over conditions. TS-EMO was selected for its ability to outperform several multi-objective optimisation algorithms; including, Pareto efficient global optimisation (ParEGO), non-sorting genetic algorithm ii (NSGA-ii) and expected improvement matrix efficient global optimisation (EIM-EGO).48,49 In this work, three input variables (rather than two) define 3D reaction parameter space (residence time, temperature and initiator concentration), and the option to conduct and include in silico experiments to inform regions of interest was programmed. In both the fully experimental and model informed cases, the ability to map out the trade-off between low molar mass dispersity (Đ) and high conversion was characterised.
A computer-controlled reactor platform was constructed, enabling an autonomous feedback loop by varying input variables (temperature, residence time, and initiator concentration) and analysing output data (monomer conversion and molar mass dispersity). This advanced setup included three pumps, two packed bed mixers, a temperature-controlled tubular flow reactor, a benchtop NMR for monitoring monomer conversion, a GPC setup for obtaining the molecular weight distribution. This latter process also required and an in-line switching valve for sampling the reaction solution emerging from the reactor, which was then diluted and passed though the GPC. A schematic of the platform can be observed in Fig. 1a.
This autonomy was enabled through the development of a bespoke graphical user interface which was configured to operate in two ways (Fig. 1b): the first operates in a wholly empirical manner – where the algorithm learns exclusively from experimental data (termed “fully experimental”). The second operates in a model-informed manner, where the algorithm is trained using in silico data obtained using our previously developed model for RAFT polymerisation,45 (termed “model-informed”). As demonstrated in Fig. 1b(i) and (ii), respectively. Both optimisation campaigns use training data obtained by selection of experiments in the defined parameter space using LHC sampling, forming an objective function containing 15 experiments. In the model-informed approach, a preliminary objective function is generated from simulated α and Đ data obtained from the in silico LHC, which is then used to predict 5 experiments using an initial iteration of TS-EMO50 (TS-EMO 1). As the model is not a fully analytical function of the reaction/reactor, the model-informed approach can still be considered black-box. Upon executing these 5 experiments, and the results are used to construct a new experimentally obtained objective function. Subsequent, iterations of TS-EMO (TS-EMO 2) then suggested single experiments which were evaluated experimentally and used to update the experimentally obtained objective function. This approach provides two layers of training data which considers both the chemistry, and an understanding of how the reactor configuration affects the process (e.g. RTDs and heat transfer). These approaches provide the main TSEMO campaign with the most reliable training dataset for exploration and exploitation.
The subsequent model-informed experiments involved the in silico simulations using conditions defined by LHC sampling – see black diamonds in Fig. 2b(ii). This data was obtained in approximately 15 s using a standard laboratory PC (see SI for specifications), compared to the fully experimental platform which took 561 minutes to conduct the physical set of training experiments. This simulated data was then used to generate a preliminary objective function approximation that informed the first iteration of the TSEMO algorithm which generated 5 conditions for physical experimentation on the flow platform.51–53 Once these were completed, the data was fed into a second iteration of TSEMO that would suggest single experiments that could gain hypervolume. The model informed approach mapped the Pareto front more effectively and efficiently (Fig. 2b), including determination of more non-dominated solutions at lower conversion and lower dispersity compared to the fully experimental optimisation (Fig. 2a). When there was no visible qualitative improvement in the Pareto front the optimisation was terminated. On plotting the hypervolume obtained (Fig. S11) it is clear after 15 experiments the total hypervolume exceeds that obtained during the fully experimental method (which conducted 22 experiments).
The accuracy of the model is further validated in these experiments, where the in silico data overlaps the experimental data. It can be ascertained from the plots in Fig. 2b that there is a positive correlation between the three control variables (temperature, initiator concentration and residence time) on the conversion. Furthermore, more reaction space corresponding to areas of low dispersity was investigated by the model informed platform compared to the experimental platform shown in Fig. 2a. When considering the condition selection, it is clear the model informed platform selects more conditions at lower temperatures which is likely a result of the model identifying this as a region of interest pertaining to a steady supply of radicals and reduction in terminative events.
As a result, the model-informed platform identified more points with the lowest dispersity such as point 2 in Fig. 2b(ii) where GPC indicates a Đ of 1.14 whilst NMR reports a conversion of 58%. The conditions for this experiment involved an 8 min residence time at 56 °C using 0.06 eq. of initiator (relative to CTA). The closest non-dominated solution in the fully experimental plot was found after 6 min at 58 °C using 0.08 eq. of initiator – where 64% conversion was achieved with a Đ of 1.15. There is a huge difference in the number of experiments required to find these low Đ points, the fully experimental platform found the lowest Đ in 22 experiments compared to just 2 in the model informed optimisation. This is in the first iteration of the TSEMO algorithm for the model-informed platform, which is the initial experimental exploration, so we can assume that this is not a fortuitous result as the first iteration learns directly from the model. In contrast to the fully experimental platform, the model-informed platform finds conditions that lead to lower Đ, thus widening the exploration of reaction space.
Within this optimisation algorithm, a set of hyperparameters generated comprising the mean and covariance are used to generate a surrogate mathematical model. The covariance function has an inbuilt noise parameter associated (σ2) which allows the algorithm to handle experimental data. Conveniently, σ2 can be used to determine how important each output parameter is for controlling the optimisation process, where a lower σ2 indicates more effective choice of conditions for obtaining non-dominated solutions. Consequently, it determines whether there are truly benefits in using the model informed approach over the fully experimental approach.
The significantly lower noise hyperparameters (see Table 1) associated with the model-informed optimisation (σMI2) compared to the experimental optimisation (σexp2) demonstrate that the model directed platform is much more efficient than the fully experimental approach, with identified solutions leading to the Pareto front being identified in 40% fewer experiments than the fully experimental optimisation. This is a 58% reduction in experiment time and a saving of 270 mL of reagent solution demonstrating a step-change in the efficiency of chemical synthesis optimisation (Table 2).
Hyperparameters | GP1 (dispersity) | GP2 (conversion) |
---|---|---|
σexp2 | 0.32 | 0.46 |
σMI2 | 0.04 | 0.09 |
Fully experimental optimisation | Model-informed optimisation | |
---|---|---|
Total number of experiments | 28 | 13 |
Number of training experiments | 15 | 5 |
Volume of solution (mL) | 450 | 180 |
Time taken to map the Pareto front (h) | 17 | 7 |
During the fully experimental optimisation, there is a steep increase in molar mass dispersity and an inability to attain conversion higher than 80%. The Pareto front is also relatively poorly defined, with only four points contributing to hypervolume gains (see Fig. 3a(ii)). The non-dominated solution with the lowest molar mass dispersity Đ = 1.09; point 6 in Fig. 3a(i) was obtained at a low monomer conversion of 17%. The non-dominated solution with the highest conversion is obtained after 19 min at 100 °C using 0.06 eq. of initiator to obtain a conversion of 76% but a broad molecular weight distribution (Đ = 1.47).
For the model-informed experiment, marginal improvements are observed with respect to optimal samples (either high conversion or low dispersity) but notably, the Pareto front is better defined at higher conversion. In this case, the reaction yielding the polymer with the lowest dispersity has around a 10% improvement in monomer conversion (Fig. 3b), whereas the highest conversion non-dominated solution provides a negligible improvement in molar mass dispersity (1.46 vs. 1.47).
The simulated reactions, which assume only initiation, propagation, chain-transfer, and termination, show minimal up-turn in dispersity at high conversion (Fig. 3b(ii); black diamonds). This contrasts with experimental results, where there is a much larger up-turn. This directly reflects the added complications for acrylate polymerization, including slower propagation from mid-chain radicals.54 This example highlights the importance of the three-stage optimisation process: initial in silico experiments followed by two iterations of TSEMO. The in silico experiments can simulate the “ideal” chemistry, while the first TSEMO iteration optimises for the real system, providing crucial real-world data which is influenced by side reactions in addition to reactor-specific phenomena including fluid dynamics (manifesting as a broader RTD) and non-ideal heat transfer. This data contributes to improving the relatively poor initial surrogate model providing a better basis for the second TSEMO campaign. This methodology demonstrably reduced experimental overheads by over 35% compared to the black box study.
Once again, quantitative comparison (see Table 3) for the optimisations indicates the model-informed approach is more effective, where σMI2 is smaller than σexp2. However, the difference in this case is less drastic, which is a likely result of the more complex chemistry. Nevertheless, it demonstrates that implementation of a model that predicts the best-case scenario can be used as the foundation to a thorough optimisation of the real system. This results once again in the Pareto front mapping being achieved more efficiently, using fewer experiments, less time, and materials (Table 4).
Hyperparameters | GP1 (dispersity) | GP2 (conversion) |
---|---|---|
σexp2 | 0.33 | 0.63 |
σMI2 | 0.21 | 0.31 |
Fully experimental optimisation | Model-informed optimisation | |
---|---|---|
Number of experiments | 42 | 26 |
Number of training experiments | 15 | 5 |
The volume of solution (mL) | 714 | 442 |
Time taken to map the Pareto front (h) | 36 | 22 |
This journal is © The Royal Society of Chemistry 2025 |