Open Access Article
This Open Access Article is licensed under a
Creative Commons Attribution 3.0 Unported Licence

A simple compound prioritization method for drug discovery considering multi-target binding

Alžbeta Kubincová and David L. Mobley*
Department of Pharmaceutical Sciences, University of California, Irvine, Irvine, CA 92697, USA. E-mail: dmobley@uci.edu

Received 15th October 2025 , Accepted 19th January 2026

First published on 10th February 2026


Abstract

Active learning is an emerging paradigm used to help accelerating drug discovery, but most prior applications seek solely to optimize potency, whereas multiple properties influence a compound's utility as a drug candidate. We introduce a method for multiobjective ligand optimization, which is able to efficiently handle distinct molecular properties that are expensive to compute, such as binding affinities with respect to multiple protein targets. We validate this protocol retrospectively using docking scores, showing an improved retrieval of the top 0.04–0.4% binders from the dataset with our method compared to greedy acquisition, owing to a better distribution of the compute budget between different properties. Our results also suggest that fitting individual properties separately leads to a better rank correlation of the resulting predictions. This workflow addresses the needs of pharmaceutical research for improving the efficiency of hit-to-lead and lead optimization by considering binding to multiple targets. Our code is freely available on Github: https://github.com/MobleyLab/active-learning-notebooks/blob/main/MultiobjectiveAL.ipynb.


1 Introduction

Active learning (AL) strategies for identifying molecular drug candidates gained significant attention in the pharma industry over the last decade. Instead of scoring large chemical libraries with expensive physics-based methods, a surrogate machine learning (ML) model is trained on a small subset of scores from the expensive method, and used to predict the scores of a much larger compound dataset. These predictions are used to select next batch of compounds using an acquisition function, which links the predicted scores and their uncertainties to selection priority. The selected compounds are then scored with the physics-based model, and the ML model is updated with the new scores to improve its predictive performance (Fig. 1A). This approach allows for scanning large chemical spaces using only a fraction of the computational cost associated with scoring of the whole library in a brute force manner, while also building and improving a predictive model.1–3
image file: d5dd00464k-f1.tif
Fig. 1 Single- and multiobjective optimization algorithms considered in this work. A small sample of a ligand set is scored, and a surrogate ML model (Gaussian process regression in this work) is trained on these compounds to predict the scores of a larger set, which is usually orders of magnitude faster than computing the scores. The predictions from ML are used to select a small batch of compounds for scoring, and the cycle is repeated multiple times. (A) Algorithm for optimizing the docking score (green tag) towards a single target. (B) Optimization of the ligand utility, which is encoded by an objective function composed of docking scores towards multiple receptors (green and pink tags for the two sets of scores). The workflow is technically identical to the single-objective case, with the utility function being used in place of a single property. (C) Optimization of ligand utility, while training a separate ML model for each target property. The acquisition can either be guided by the individual properties predicted for each molecule (separated acquisition) or by the overall objective function (and its uncertainty) computed from the predicted scores (joint acquisition).

The application of AL to computational drug design is still in its early stages, and most studies to date only optimize for potency, estimated from docking scores4,5 or relative/absolute binding free energies (RBFE/ABFE) from molecular dynamics (MD) simulations.6–9 However, the utility of a molecule as a drug candidate is influenced by properties other than the potency, particularly by ADME (absorption, distribution, metabolism, and excretion) properties, toxicity and selectivity. Some AL studies account for ADME properties by applying molecular property filters upon constructing the ligand dataset6,9,10 to remove compounds from the pool that are likely to have poor ADME properties. On the other hand, compounds that are potentially useful and fall slightly short of the ideal property ranges are discarded by this procedure, whereas a smooth penalty can account for trade-offs between properties.11

However, a filtering approach is rarely desirable to account for toxicity and selectivity. While the ranges for ADME properties are usually well-defined, drug safety is measured as the ratio of potency and toxicity (therapeutic index).12 In a typical drug discovery campaign, selectivity is only assessed for compounds with good potency,13 using separate free-energy calculations. The staged approach for on- and off-target free energy calculation reduces the number of expensive MD simulations, but can miss regions in chemical space, with slightly lower potency but better safety. Compounds found in lead optimization entering the preclinical phase are therefore often fraught with liabilities that are hard to mitigate. More than 90% of compounds entering clinical trials fail, with 10–15% due to poor drug-like properties and 30% due to toxicity.14 It might therefore be advantageous to explicitly account for off-target effects during optimization rather than as a filtering stage, especially in a streamlined early discovery process with AL.

To integrate multiple properties directly, researchers can either define a compound's utility using a single overall objective function combining these properties or use Pareto optimization. An objective function defines the relationship of each property to a drug's utility.15,16 The construction of an objective function requires the user to define a weighting scheme for the individual properties, which is not straightforward and involves trial and error.17 On the other hand, Pareto optimization considers properties as separate objectives to optimize. A compound is Pareto optimal if none of its properties can be improved without degrading another property. This method does not require a weighting scheme and is often preferred for its stability and consistency.18–21 On the other hand, Pareto optimal solutions may not always reflect the needs of drug discovery (e.g. minimizing off-target binding at the cost of potency), and domain-specific knowledge may need to be included in a post hoc manner.22 In general, it is however not possible to directly compare the present scalarization of objectives with Pareto optimization, since their different definitions of the optimization target imply different optimization approaches and goals. Both methods at best only approximate the true utility of a compound as a drug molecule, as utility often depends on a particular balance of properties (which may not be known a priori) and may be influenced by other factors, such as structural novelty.

These approaches for multiobjective ligand optimization are well-suited to problems where at most one property is expensive to compute. Properties such as ADME can efficiently be precomputed and combined with free-energy estimates once they are calculated for a given compound. This is not the case for binding to multiple targets, where the affinity to each target requires a separate calculation that is expensive, especially if binding free energy calculations are used. All affinities need to be calculated at once to evaluate the objective function, which is used as a reward for the surrogate ML model (Fig. 1B). However, a subset of these affinities is often sufficient to discard a compound. Poor on-target potency eliminates the need for off-target screening. Likewise, if a dual inhibitor is sought, a weak binding affinity to one of the targets means there is no need to calculate potency towards the second target since the compound can already be ruled out. Therefore, a simultaneous calculation of multiple properties only makes sense if the prior distributions of all properties (e.g. predictions from an ML model) are in a favorable range and sufficiently narrow (e.g. good potency and selectivity are predicted with high confidence). However, in case of high uncertainty, the compute budget would be better split across potency calculations for multiple compounds rather than exhausted on computing all off-target affinities for a single compound that may not be potent. This can become an issue for targets with many similar analogues (for example, the selectivity of a JNK1 inhibitor series was affirmed upon screening against a panel of 74 kinases23).

In this work, we introduce an AL method that efficiently accounts for multiple properties that are expensive to compute (Fig. 1C). The method relies on an objective function for a given problem, but acquires different batches of compounds for the computation of each property (separated acquisition) using a modified version of the Expected Improvement (EI) acquisition function24,25 (see the Theory section). The separated acquisition allows for a better distribution of the compute budget by taking into account each property's impact on the target candidate profile. The acquisition is separate from the ML model, and therefore does not stipulate any particular ML architecture as long as prediction uncertainties are available. The ML models are fitted to structural features of molecules to predict their properties. Here, we use a separate Gaussian process model to fit docking scores with respect to each target. The protocol can also be described as an independent AL cycle for each property, with the planner communicating upon acquisition of new compounds.

The goal of this work is to create a framework for the optimization of potency in the presence of off-targets or for multi-target inhibition. We develop and validate this framework using docking scores as a proxy for binding affinity, although more accurate and expensive methods will be required in real applications. The use of docking scores here, as we develop and test this framework, is motivated by docking scores' low cost. In particular, for our study here, they are available in large quantities for a suitable compound library while being derived from a consistent docking protocol. However, our framework also works in the same way using more expensive (and accurate) means of binding affinity prediction, such as free-energy calculations with MD, where these calculations dominate the computational (and monetary) cost of the workflow. Here, we do not devote significant attention to evaluating computational cost in terms of wall time. Specifically, since the framework is evaluated in a retrospective way using precomputed docking scores, wall times are not indicative of the performance in prospective applications in this case. We expect the potency calculation to be the slowest step in a prospective scenario, which is why we always compare between protocols that have access to the same number of potency samples for selecting the next batch of ligands.

Here, we demonstrate the superior hit rate of our separated acquisition strategy compared to joint acquisition on two optimization problems and provide insight into various parameters impacting hit retrieval.

2 Theory

In this section, we lay the theoretical framework for our acquisition strategy, which scores individual target properties rather than compounds as a whole. We provide a brief overview of the Expected Improvement acquisition function and adapt the function to separate the contributions of target properties.

2.1 Expected Improvement (EI) acquisition function

EI is a widely used strategy in Bayesian optimization to locate the maximum of an objective function f.24,25 Assuming that the probability density of f follows a normal distribution prior to measurement, image file: d5dd00464k-t1.tif, EI is defined as the expected increase in the value of f over a reference value f* (usually set to the current maximum) following measurement (where we assume that the measurement is exact):
 
image file: d5dd00464k-t2.tif(1)
Eqn (1) evaluates to
 
image file: d5dd00464k-t3.tif(2)
where [f with combining circumflex] and σf are the mean and standard deviation of f, and ϕ(x) and Φ(x) denote the assumed probability density function and cumulative distribution function of x, respectively (here, for a normal distribution). In a drug discovery context, f stands for the objective function of a given compound, and eqn (2) would be evaluated for every compound in the pool.

2.2 Adapted EI for scoring individual properties

The original definition of EI, as given by eqn (1), measures the overall improvement of f for a given sample. We will now adapt EI to account for the improvement in f resulting from a given property p, which contributes to f. In practice, properties are chosen to reflect the target candidate profile of a drug, such as potency, selectivity and ADME, and the objective function defines their relationship to each other. This strategy is therefore different from Pareto optimization and the related Expected Hypervolume Improvement (EHVI) acquisition function.26 EHVI does not rely on an objective function, but instead measures the possible improvement of the Pareto front as a hypervolume in multivariate space. By contrast, our method relies on an objective function, but separates the contributions of the individual properties.

Consider a piecewise linear objective function f of multiple properties pP, defined in segments sS spanning the entire range of values that f can adopt. Consequently, in each segment, f can be written as a linear function of any individual property p, if all other properties Q = {qP|qp} are kept constant:

 
f(p|Q) = αps(Qp + βps(Q), p∈[aps(Q),bps(Q)]. (3)
α and β are the linear coefficients, and a and b are the bounds of each segment. To give an example, the objective function f(A, B) = min(A, B) can be written in this notation for the property A as f(A|B) = A if A < B, B if A > B. This results in two segments (S = 2), with the coefficients αA1 = 1, βA1 = 0 ranging from aA1 = −∞ to bA1 = B for the first segment, and αA2 = 0, βA1 = B ranging from aA1 = B to bA1 = ∞ for the second segment. Since the coefficients α, β, a and b can be functions of properties other than p, the terms from individual properties do not need to be additive. This allows one to consider scenarios where properties are coupled to each other, such as for the PPAR test system considered in this study (eqn (9)).

Next, we define the EI of a property p as the expected improvement of f over f* while optimizing only that single property p, while assuming the maximum likelihood estimate (MLE) for all other properties, [Q with combining circumflex]:

 
image file: d5dd00464k-t4.tif(4)
Eqn (4) is approximate owing to the MLE assumption, which avoids integrating over the joint probability distribution ρ(P) of all properties. This means that the sum of EI(p) terms over all properties will not add to EI(f). The MLE assumption has critical consequences in practice, which we will address later. However, the assumption is also crucial for practicality, as it breaks an integral over a multidimensional probability density down to a set of one-dimensional integrals, which can be computed analytically (in contrast to the EHVI, which is typically evaluated by Monte Carlo integration).

Substituting the definition of f from eqn (3) into (4), together with image file: d5dd00464k-t5.tif, results in

 
image file: d5dd00464k-t6.tif(5)
where image file: d5dd00464k-t7.tif and image file: d5dd00464k-t8.tif are adapted from the bounds aps and bps, ensuring that f(p|[Q with combining circumflex]) − f* remains positive in the interval image file: d5dd00464k-t9.tif:
 
image file: d5dd00464k-t10.tif(6)
for bound ∈ {aps, bps} and image file: d5dd00464k-t11.tif. image file: d5dd00464k-t12.tif is the value of p for which the segment function evaluates to f*, noting that image file: d5dd00464k-t13.tif may fall outside the interval [aps, bps],
 
image file: d5dd00464k-t14.tif(7)
In the context of AL, out of the full library, a separate pool or subset of compounds is maintained for each property (i.e. docking score), and eqn (5) needs to be evaluated separately for every compound and property available for acquisition.

The Python implementation of EI(p) from eqn (5)–(7) is provided in the notebook included in the Github repository (see the Data availability). The user needs to provide the definition of f, as well as the coefficients aps, bps, αps and βps for each property and segment. The original EI formulation from eqn (1) can be recovered as a special case of eqn (5) with the parameters a = −∞, b = ∞, α = 1 and β = 0.

3 Methods

3.1 Benchmark design

We validate our active learning method using the DOCKSTRING dataset,27 considering discovery use cases where binding to more than one target influences a compound's utility. This dataset contains docking scores from 260[thin space (1/6-em)]000 molecules against 58 clinically relevant targets and was constructed for the purpose of validating ML models for potency prediction. Ligand molecules were taken from the ExCAPE database,28 which curates bioactivity assays from PubChem and ChEMBL. The selection ensures that at least 1000 active molecules were selected for each target, together with 150[thin space (1/6-em)]000 ligands with inactive labels against all targets. Experimental actives are expected to have better docking scores than inactives, although a good correlation between docking scores and experimental affinities cannot be guaranteed.27 The dataset was filtered to remove duplicate molecules (7 in total).

We encode the desirability of a ligand by an objective function to be maximized, which aggregates docking scores with respect to multiple targets. In addition to the docking scores, a penalty is added for molecules that are not drug-like, relying on the quantitative estimate of drug-likeness (QED)29 metric, which we compute using the image file: d5dd00464k-u1.tif function from RDKit. The QED accounts for eight molecular properties (molecular weight, octanol–water partition coefficient, number of hydrogen bond donors, number of hydrogen bond acceptors, molecular polar surface area, number of rotatable bonds, number of aromatic rings and number of structural alerts), converting their desirability to a continuous scale ranging between 0 and 1. We consider two design objectives as suggested by the DOCKSTRING authors:27

(1) Selective for JAK2: this task involves maximizing the binding to JAK2, while minimizing binding to its off-target LCK. In order to avoid offsetting poor JAK2 binding by unusually weak LCK binding, the contribution from LCK is capped at its median score (−8.1).

 
fJAK2 = −JAK2 + min(LCK,−8.1) − 10(1 − QED). (8)

Fig. 2A shows a graphical representation of fJAK2 as a function of JAK2 and LCK. Eqn (8) offers the advantage of simplicity, but other functional forms can be used to account for selectivity in practice. In particular, a penalty for the difference between on- and off-target binding may better represent the need for a therapeutic window, rather than anchoring off-target scores to a fixed value.


image file: d5dd00464k-f2.tif
Fig. 2 Visualization of objective functions for JAK2 and LCK, which aggregate docking scores across multiple targets. (A) fJAK2 from eqn (8) is shown as a function of JAK2 and LCK scores for QED = 1. (B) fPPAR from eqn (9) is visualized as a function of PPARA and PPARD scores (assuming PPARG = −∞ and QED = 1). (C) Contour plot of fJAK2 with a representative data point predicted with a mean and uncertainty for both JAK2 and LCK. Marginal probability distributions are shown for both properties (blue curves) together with their improvement max(f(p|[Q with combining circumflex]) − f*,0) in red. The blue and red lines represent the components of the property-specific EI integral from eqn (4). (D) Contour plot for fPPAR, in analogy to (C).

(2) Promiscuous for PPAR: this task involves finding inhibitors simultaneously binding to three PPAR receptors, namely PPARA, PPARD and PPARG. The objective function therefore maximizes the potency towards the weakest-binding receptor and is visualized in Fig. 2B:

 
fPPAR = −max(PPARA,PPARD,PPARG) − 10(1 − QED). (9)

3.2 Multiobjective AL workflow

The workflow (Fig. 1B and C) is seeded with an initial set of 1000 randomly selected ligands, for which all properties are acquired at once. The size of the initial dataset was chosen upon confirming that all Gaussian process models fitted to these data are strongly predictive (Spearman ρ = 0.6 at minimum). We chose the initial batch size in accordance with comparable studies by the DOCKSTRING authors,27 and we did not optimize batch sizes in our study. In practice, the amount of samples for training a predictive ML model varies greatly with the underlying data30 and is particularly dependent on the structural diversity within the dataset and the strength of the dataset's structure–activity relationships. In case of the JAK2 test system, we observed similar outcomes when seeding the workflow with only 100 ligands (SI Section 2).

The ligands are featurized using chiral Morgan fingerprints as implemented in RDkit with a radius of 4 and a length of 1024 bits. We chose this feature set based on its widespread use in the literature and because we are aware of molecules in DOCKSTRING that differ only in their stereochemistry. However, we did not optimize our feature selection with respect to the predictive accuracy on this dataset, in order to reflect an application scenario where insufficient data are available to carefully select a high performing feature set. A comparison of Morgan fingerprints with MACCS keys is provided in Section 1 of the SI, showing that concatenating these two representations can further improve the predictive accuracy of the ML model.

One or multiple Gaussian process (GP) models were fitted to the reward given by either the objective function or individual docking scores. The GP was chosen for its ability to natively provide prediction uncertainties and for its good performance with small amounts of training data. The model was implemented using GPyTorch with a Tanimoto similarity kernel. In each cycle, the budget for acquisition was fixed at 300 scores for the separated acquisition strategy, which translates to 150 and 100 compounds with the joint acquisition strategy for JAK2 and PPAR, respectively. In total, we acquire about 1% of all DOCKSTRING ligands for JAK2 and 0.8% for PPAR, taken together from the initial batch and the subsequent 10 optimization cycles. Five repeats are executed for each scenario with different random seeds for the initial batch selection, in order to calculate the uncertainty of overall metrics. We also show that it is possible to use this protocol with smaller batch sizes in Section 3 of the SI.

3.3 Training and acquisition strategies

The separated acquisition strategy given by eqn (5)–(7) requires individual docking scores to be available for pool compounds, rather than only the objective function f (since docking scores are treated as separate properties p in eqn (5)). This strategy therefore necessitates training individual ML models for each target property. In order to evaluate training one or multiple ML models separately from the acquisition strategy, we consider three major versions of the optimization procedure: joint training and acquisition, separated training with joint acquisition, and fully separated training and acquisition.

The joint training and acquisition strategy relies on a single ML model, which is fitted to f directly, as shown in Fig. 1B. The workflow is technically identical single-objective optimization (Fig. 1A), where the objective function is used as an optimization objective in place of a single affinity score. Docking scores acquired for the initial dataset as well as subsequent compound batches are converted to the objective f, which is used as a reward for training the ML model. Consequently, inference on the compound pool and acquisition of the new batch is only informed by f rather than individual properties. We implement two acquisition strategies: EI based on eqn (2), and a greedy strategy selecting compounds with the highest predicted score f.

Another strategy involves the training of separate ML models for each property while acquiring new compounds based on f computed from individual properties (Fig. 1C). In this case, individual ML models are fitted to each docking score separately, rather than to f, and the QED is precomputed for the whole dataset (which only takes a few minutes). Upon inference, f is computed from the predicted properties, with their uncertainties propagated onto f according to

 
image file: d5dd00464k-t15.tif(10)
Here, we also apply EI from eqn (2) as well as a greedy acquisition strategy.

Lastly, we consider a setup with separate ML models coupled with the separated acquisition strategy (Fig. 1C). In this case, the acquisition is informed by the individual docking scores rather than f, and different batches of compounds can be acquired for each target property. The acquisition is implemented as follows; For each compound–property pair that has not been measured yet, EI(p) is calculated using eqn (5)–(7). The values are sorted, and compound–property pairs with the highest EI(p) are acquired, where the batch size is defined as the number of scores, or compound–property pairs. This strategy adds flexibility to efficiently distribute the acquisition budget across different targets as a response to the objective function as well as the underlying data (cf. Fig. 2C and D). However, the acquisition strategy does not guarantee that valuable compounds will be scored against all targets, which is why we switch to a greedy strategy for the last cycle, acquiring (missing) scores of compounds with the highest predicted value of f.

In summary, five different strategies are evaluated:

EI joint: joint training with joint acquisition based on EI.

greedy joint: joint training with joint acquisition based on a greedy strategy.

EI sep. train.: separated training with joint acquisition based on EI.

greedy sep. train.: separated training with joint acquisition based on a greedy strategy, and

EI sep. acq.: separated training with separated acquisition using the modified EI algorithm.

All of the strategies outlined above are exploitative, aiming to acquire potent ligands rather than enhancing the accuracy of the underlying ML model(s). Instead of adding rounds of explorative acquisition, we rely on the models built on the initial batch to be predictive enough to assist in exploitation. In some cases, balancing exploration with exploitation can be important in active learning, especially with regard to retrieving diverse hits and improving quality of the model. However, these considerations – and the possibility of balancing exploration vs. exploitation – are beyond the scope of the present work, where the number of hits is used as the primary metric for comparing between optimization protocols.

There is one key difference in our calculation of the EI compared to its original definition of f*. The value of f* can be interpreted as an acquisition boundary, such that the improvement from possible values of f below f* is zero. Although this parameter is traditionally defined as the best (true) score among the acquired compounds, it is treated as a constant in eqn (1)–(7) and can therefore be set to any other (constant) value without loss of generality. Here, to avoid problems where acquisition might get blocked when further score improvement requires optimization of multiple properties, we apply a different strategy. Specifically, we set f* such that f* < [f with combining circumflex] for any compound in the pool, which would be acquired by a greedy strategy in the current cycle. For the joint acquisition strategy, f* is set equal to the predicted score of the top 150th and 100th compound in the pool for JAK2 and PPAR (since 150 or 100 compounds are acquired in each cycle), whereas separated acquisition relies on the top 300th score (since the scores can be distributed freely across different properties, which may result in acquiring a single property for 300 compounds). We introduce this modification because separated acquisition does not perform well with the original choice of the f* boundary. The improvement EI(p) can only be greater than zero if the value of f can increase by changing any single property at a time (i.e. an upward move is possible along the dark green lines in Fig. 2A and B). This is not always the case for properties, which only contribute to f up to a limit (LCK in eqn (8), and all of the PPAR receptors in eqn (9)). Therefore, it is possible for a compound to have a high EI(f) (resulting in acquisition by the joint strategy), but EI(p) = 0 for all properties (resulting in non-acquisition by the separated strategy). This behavior is a consequence of the MLE approximation in eqn (4). The alternative choice of anchoring f* to the predicted scores of pool compounds circumvents this issue, ensuring that all compounds that would be selected by a greedy strategy fulfill EI(p) > 0 for all properties. We will show in Section 4.4 how the choice of f* affects the retrieval of desirable compounds.

4 Results and discussion

4.1 Separating training and acquisition boosts the retrieval of top binders

The five optimization protocols are compared based on the number of hits retrieved within 10 cycles, measured by the objective functions in eqn (8) for JAK2 and eqn (9) for PPAR. Fig. 3 shows consistent trends between different protocols for JAK2 and PPAR. Here, we define an active compound as one with a rank higher than a given threshold, considering thresholds ranging from 100 to 1000. The recall is then computed for a given threshold as the fraction of active compounds retrieved within 10 cycles (i.e. retrieving 60 out of the top 100 compounds according to f results in a recall of 0.6 for the threshold 100). The thresholds considered account for the top 0.04–0.4% of the whole dataset, and all protocols show a clear advantage in recall over a random selection in this range of thresholds (random selection results in a recall of <0.01 for both systems).
image file: d5dd00464k-f3.tif
Fig. 3 Fraction of active compounds retrieved for the JAK2 (A) and PPAR (B) systems, as a function of the threshold rank for classifying a compound as active. The recall is calculated as the number of actives found within the 10 cycles, divided by the corresponding threshold. Lines represent the different strategies for training and acquisition as outlined in Section 3.3. In case of separated acquisition, the recall is based on compounds that had all of their properties acquired. These data show that training separate ML models has a distinct advantage over fitting the objective function directly (specifically, all of the curves with separate training are far higher than those with joint training), whereas no significant difference can be seen between EI and greedy acquisition. The separated acquisition strategy offers an additional advantage for both systems, especially for PPAR.

Comparing individual protocols reveals that the training as well as acquisition strategies significantly impact the number of hits retrieved, whereas no significant difference can be seen between greedy and EI acquisition. The largest improvement in hit retrieval is achieved by fitting separate models for each property, rather than fitting f directly. Separated acquisition does however offer an additional advantage in almost all cases. We will rationalize these findings in more detail in the following sections.

The docking scores of JAK2 and LCK are positively correlated, as are the scores of the three PPAR receptors.27 While the correlation of scores increases the difficulty of selectivity optimization, it simplifies the task of finding promiscuous inhibitors in most cases. A unified ML model can better account for features that enhance the binding to all PPAR receptors at the same time, in contrast to learning miniature differences between on- and off-target binding. The separated training strategies show comparable trends between JAK2 and PPAR, since each model learns distinct structure–activity relationships for each target. It is however not clear how the correlation between scores affects the separated acquisition trends, as the improvement in PPAR recall may also be influenced by the higher flexibility for distributing the acquisition budget across three, rather than only two, receptors.

4.2 Training separate models for individual properties improves model predictivity

To understand the relationship between recall and model predictivity, we compare the models from each protocol based on their Spearman rank correlation ρ between properties predicted in the final cycle with the true values for these same properties (Fig. 4A and B). While all models are strongly predictive, computing the objective function from individual docking scores predicted by separate models improves ρ by a factor of 1.5 compared to predicting f directly. On the other hand, the acquisition strategy does not influence the model predictivity in any significant manner.
image file: d5dd00464k-f4.tif
Fig. 4 Spearman's rank correlation of the docking scores predicted in the last cycle with their true values and their effect on acquisition. (A) Rank correlation of the predicted objective function fJAK2 for the five protocols, as well as the correlation of scores predicted by the target-specific models for JAK2 and LCK. (B) Rank correlation for the models fitted to PPAR scores, in analogy to (A). (C) Correlation of the objective function fJAK2-only and JAK2 predictions in a JAK2-only optimization, eqn (11). (D) Retrieval of the top 100 compounds measured by fJAK2-only. (E) Probability density of the objective fJAK2-only, calculated for the compounds retrieved within 10 cycles (excluding the initial batch), compared to the fJAK2-only distribution of all ligands in DOCKSTRING. These data show that models trained on f directly have an inferior rank correlation compared to models fitted separately to each target, which is mainly caused by the QED being fitted together with docking scores for the joint models.

Two factors contribute to the performance gap of separated and joint ML models. First, being trained on a combination of docking scores towards multiple targets, the joint models need to learn multiple structure–activity relationships at the same time as well as their relationship to each other, making this a difficult problem for training. Second, the joint models are required to learn drug-likeness (QED) simultaneously with the docking scores, whereas the QED is treated as a lookup with the separated training strategy, since the QED computation is fast and does not warrant being treated as a separate property for training and acquisition in practice. To assess the impact of including the QED into the training objective in the absence of other targets, we consider a simplified version of the JAK2 problem by eliminating LCK as an off-target:

 
fJAK2-only = −JAK2 − 10(1 − QED). (11)

The correlation of final models from optimizing fJAK2-only from eqn (11) in Fig. 4C shows a similar trend to JAK2 and PPAR regarding the separated models outperforming the joint ones by a good margin. This suggests that the QED is the primary factor impacting model predictivity. On the other hand, any of the models trained on a distinct target outperforms the corresponding joint model, suggesting that individual docking scores may be easier to fit compared to the aggregated objective. The importance of improving the underlying model for finding hits can also be understood from Fig. 4D and E. Although the joint training strategy achieves two thirds of the ranking performance compared to separated training, it only retrieves one third of the top 100 compounds in comparison (Fig. 4D), since the recall is determined by the tail of the score distribution of acquired compounds, which is very sensitive to small improvements of the model (Fig. 4E).

Despite the improved fit, we must also note that training and predicting from separate ML models increases the computational cost. The overall computational cost will only be comparable if the calculation of binding scores is orders of magnitude more expensive than the cost of training and inference. Maintaining separate models may not be desirable if the costs of training and obtaining labels are comparable, and multi-task learning approaches may be preferred.31 We did not pursue this approach here for the sake of simplicity.

4.3 Separated acquisition adapts to the objective function as well as to the underlying data

In the last Section, we have shown how the model predictivity impacts the hit rate in order to rationalize the superior recall of the separated training protocols. In this Section, we relate the improvements in hit rate from the separated acquisition to the distribution of the acquisition budget across different properties.

Fig. 5 shows the number of docking scores for each target acquired per cycle by the separated acquisition strategy for JAK2 (Fig. 5A) and PPAR (Fig. 5B), given a budget of 300 scores per cycle. Both test systems result in an unequal allocation of the score budget compared to the joint acquisition strategies, which are bound to acquire 150 scores of each kind for JAK2, and 100 for each of the three PPAR receptors. The dominant acquisition of JAK2 scores over LCK can be understood from the contributions of both properties to the objective function from eqn (8). fJAK2 scales linearly with JAK2 potency, but is insensitive to LCK beyond a maximum value, which is why EI(JAK2) is usually larger than EI(LCK) (Fig. 2C). The acquisition function responds to the design of the objective function by favoring the more impactful variable, which is JAK2. In case of PPAR, on the other hand, the objective function from eqn (9) is symmetric with respect to the three receptors, suggesting no preference for any one of them. The emphasis on PPARG acquisition seen in Fig. 5B is probably related to the distribution of the docking scores for the three receptors, which is shown in Fig. 5C. PPARG docking scores are higher (i.e. worse) on average compared to PPARA and PPARD, thus having the greatest potential for improving fPPAR. This analysis showcases the adaptability of the separated acquisition strategy towards the objective function, as well as to the underlying data. The separated acquisition protocol eliminates the need for user-defined protocols based on manual inspection of the target candidate profile and existing data, especially since this task increases in difficulty for more complex objectives and a growing number of on- and off-targets.


image file: d5dd00464k-f5.tif
Fig. 5 Number of scores acquired of each kind using the separated acquisition strategy for JAK2 (A) and PPAR (B). Protocols relying on joint acquisition translate to 150 scores per cycle of JAK2 and LCK each, and 100 scores for each of the PPAR receptors. The acquisition strategy changes in the last cycle to a greedy one for acquiring missing properties of high-scoring compounds. (C) Probability density of docking scores for the three PPAR receptors in the dataset. The preference for JAK2 acquisition over LCK can be explained by different contributions of their scores to f (Fig. 2C). For PPAR, PPARG is acquired more than the other two receptors due to weaker binding of the ligands on average (C).

4.4 Anchoring the acquisition boundary to the pool is crucial for separated acquisition

As described in Section 3.3, our definition of EI differs from the original formulation in the choice of the reference value f*. The original definition of EI relies on the current maximum of f from among the acquired data, whereas we anchor f* to the compounds, which are still in the pool, such that any compound that would be acquired by a greedy strategy fulfills [f with combining circumflex] > f*. In the following, we compare protocols for different definitions of f* in order to illustrate why we made this choice.

Fig. 6 shows the number of top 100 compounds retrieved within a given number of cycles for different choices of f*, using both the joint (Fig. 6A and C) and separated (Fig. 6B and D) acquisition strategies. It is immediately visible that the recall is not significantly altered by changing f* for joint acquisition, while having a strong impact on separated acquisition. If f* is anchored to the acquired compounds, the recall flattens out after a few cycles, once most of the compounds with f > f* have been acquired. In case of JAK2 (Fig. 6B), shifting f* to higher values (and lower ranks) leads to acquiring more JAK2 scores and fewer LCK scores, resulting in a lower recall. However, all protocols converge in the last (greedy) cycle, which acquires the missing LCK scores. This is not the case for PPAR, where the protocols diverge, and setting f* relative to the acquired compounds drastically impedes retrieval of high-scoring hits (Fig. 6D). The problem with PPAR can be understood from its objective function in eqn (9), where each of the scores can improve f* only up to a limit imposed by the two other scores. In this case, improvement would only be possible from acquiring two or more of the scores at the same time, and this kind of improvement cannot be computed within this framework, which only considers properties in isolation. In the pathological scenario where PPARA = PPARD = PPARG = f* − δ, EI = 0 for all three receptors even though the probability of improving f* is roughly 1/8 (if all uncertainties are equal). Therefore, this compound will never be acquired, even if it was the second highest scoring compound in the dataset. Setting f* relative to the pool eliminates this problem (see Section 3.3), since f* now changes at every cycle according to the predictions for the top candidates for acquisition. This choice of f* results in a higher number of hits retrieved for PPAR, compared to both the original definition of EI, as well as the joint acquisition. We have not tested ranks different from 300 for selecting f* relative to the pool, and the protocol may further be improved by adjusting this parameter.


image file: d5dd00464k-f6.tif
Fig. 6 Comparison of acquisition protocols derived from EI with different choices for f*, coupled with separated models for each property. The figures show the recall of top 100 compounds according to the corresponding objective function for JAK2 and PPAR. f* is the reference value for computing improvement, which can be either set relative to f of the acquired compounds, or relative to the predicted value of f for compounds, which are still in the pool. The original EI definition is reflected by the blue lines, and our modified EI, which was used in the rest of the study, is shown in red. (A) Joint acquisition based on eqn (2) for JAK2. (B) Separated acquisition based on eqn (5) for JAK2. (C and D) Analogous to (A) and (B) for PPAR. It can be seen that anchoring f* to the pool compounds is beneficial for separated acquisition, whereas the choice of f* does not significantly impact joint acquisition.

5 Conclusion and outlook

We introduced an AL strategy for multiobjective ligand optimization using an acquisition function, which can efficiently handle multiple properties that are expensive to compute. We retrospectively validated our method on two toy problems from the DOCKSTRING benchmark, showing that our separated acquisition protocol retrieves more hits compared to joint acquisition. The separated acquisition strategy, which is based on EI, takes advantage of imbalances in the contributions of each property to the objective function, as well as imbalanced distributions of the individual properties in the underlying data in order to rule out compounds without acquiring all of their properties. Our results also suggest that individual docking scores are better fitted with separate ML models rather than training on the objective function directly. Building separate models results in predictions that better correlate with ground truth values, as well as an improved recall of top compounds.

The method is straightforward to implement and scalable with the number of properties (receptors) and is therefore expected to perform well for more complex utility functions involving multiple on- and off-targets or properties that may not directly relate to target binding. Since this framework can take advantage of any kind of ML model, it can potentially incorporate other architectures, keeping pace with rapid advancements in this field. In particular, multi-task learning is a promising avenue for a growing number of targets to keep the costs associated with training and inference low.

Limitations of our approach include the requirement of an objective function as an input, the lack of consideration for exploration and compound diversity, and the need for prediction uncertainties. The need for an objective function is inherent to the method and cannot easily be eliminated, and the outcome might be susceptible to the weighting of the individual properties. On the other hand, exploratory aspects can easily be incorporated into the workflow, e.g. by means of clustering compounds to promote the selection of different scaffolds, which was shown to be effective in previous work.7 Additionally, although our framework is agnostic with respect to the ML architecture, most ML models do not natively predict uncertainties, which the framework relies on. This need can however be addressed with a range of model-agnostic uncertainty quantification metrics.32 These limitations will be addressed in future work.

To showcase the utility of our method in computer-aided drug design, we intend to apply it prospectively to a discovery problem, which requires balancing the binding to multiple targets, by using free energies from MD as a measure of potency. We hope that the consideration of potential liabilities early in the discovery process will increase the likelihood of the found hits to become successful drug candidates.

Our protocol can easily be combined with generative design strategies. This can be achieved by filtering the designs to a small sample enriched in binders to multiple targets, and in turn using the selected compounds to fine tune the generative model. This approach can be executed in successive stages, allowing for a better exploration of chemical space.33 The use of a scalar objective function makes our protocol compatible with a wide range of de novo design packages, even those that do not support multi-objective rewards (such as REINVENT34).

Author contributions

A. K.: conceptualization, methodology, software, validation, formal analysis, investigation, data curation, writing – original draft, writing – review & editing, visualization. D. L. M.: resources, writing – review & editing, supervision, project administration, funding acquisition.

Conflicts of interest

D. L. M serves on the scientific advisory board of OpenEye Scientific Software, Cadence Molecular Sciences and is an Open Science Fellow with Psivant Sciences.

Data availability

All of the code required to reproduce the results reported in this study can be found on GitHub: https://github.com/MobleyLab/active-learning-notebooks/blob/main/MultiobjectiveAL.ipynb as well as permanently archived on Zenodo: https://doi.org/10.5281/zenodo.17180098. Docking scores from the DOCKSTRING benchmark are available at https://figshare.com/articles/dataset/dockstring_dataset/16511577.

Supplementary information (SI) is available. See DOI: https://doi.org/10.1039/d5dd00464k.

Acknowledgements

The authors appreciate financial support from the National Institutes of Health (R35GM148236). The authors also appreciate computing support from the UCI Research Cyberinfrastructure Center and from J&J Pharmaceuticals. These findings are solely of the authors and do not necessarily represent the views of the funding agency.

Notes and references

  1. J. L. Knight, K. Leswing, P. H. Bos and L. Wang, Free Energy Methods in Drug Discovery: Current State and Future Directions, American Chemical Society, 2021, vol. 1397, pp. 205–226 Search PubMed.
  2. D. Reker, Artificial Intelligence in Drug Discovery, The Royal Society of Chemistry, 2020, pp. 301–326 Search PubMed.
  3. D. Reker and G. Schneider, Drug Discovery Today, 2015, 20, 458–465 Search PubMed.
  4. D. E. Graff, E. I. Shakhnovich and C. W. Coley, Chem. Sci., 2021, 12, 7866–7881 Search PubMed.
  5. Y. Yang, K. Yao, M. P. Repasky, K. Leswing, R. Abel, B. K. Shoichet and S. V. Jerome, J. Chem. Theory Comput., 2021, 17, 7106–7119 CrossRef CAS.
  6. J. Thompson, W. P. Walters, J. A. Feng, N. A. Pabon, H. Xu, M. Maser, B. B. Goldman, D. Moustakas, M. Schmidt and F. York, Artif. Intell. Life Sci., 2022, 2, 100050 Search PubMed.
  7. Y. Khalak, G. Tresadern, D. F. Hahn, B. L. de Groot and V. Gapsys, J. Chem. Theory Comput., 2022, 18, 6259–6270 Search PubMed.
  8. F. Gusev, E. Gutkin, M. G. Kurnikova and O. Isayev, J. Chem. Inf. Model., 2023, 63, 583–594 CrossRef CAS.
  9. J. E. Crivelli-Decker, Z. Beckwith, G. Tom, L. Le, S. Khuttan, R. Salomon-Ferrer, J. Beall, R. Gómez-Bombarelli and A. Bortolato, J. Chem. Theory Comput., 2024, 20, 7188–7198 Search PubMed.
  10. K. D. Konze, P. H. Bos, M. K. Dahlgren, K. Leswing, I. Tubert-Brohman, A. Bortolato, B. Robbason, R. Abel and S. Bhat, J. Chem. Inf. Model., 2019, 59, 3782–3793 CrossRef CAS PubMed.
  11. D. J. Cummins and M. A. Bell, J. Med. Chem., 2016, 59, 6999–7010 CrossRef CAS PubMed.
  12. P. Y. Muller and M. N. Milton, Nat. Rev. Drug Discovery, 2012, 11, 751–761 CrossRef CAS PubMed.
  13. J. Hughes, S. Rees, S. Kalindjian and K. Philpott, Br. J. Pharmacol., 2011, 162, 1239–1249 CrossRef CAS PubMed.
  14. D. Sun, W. Gao, H. Hu and S. Zhou, Acta Pharm. Sin. B, 2022, 12, 3049–3062 CrossRef CAS.
  15. R. Winter, F. Montanari, A. Steffen, H. Briem, F. Noé and D.-A. Clevert, Chem. Sci., 2019, 10, 8016–8024 Search PubMed.
  16. F. Dey and A. Caflisch, J. Chem. Inf. Model., 2008, 48, 679–690 CrossRef CAS.
  17. S. Luukkonen, H. W. van den Maagdenberg, M. T. M. Emmerich and G. J. P. van Westen, Curr. Opin. Struct. Biol., 2023, 79, 102537 CrossRef CAS PubMed.
  18. J. C. Fromer, D. E. Graff and C. W. Coley, Digital Discovery, 2024, 3, 467–481 Search PubMed.
  19. J. C. Fromer and C. W. Coley, Patterns, 2023, 4, 100678 CrossRef CAS.
  20. J. Verhellen, Chem. Sci., 2022, 13, 7526–7535 Search PubMed.
  21. T. Suzuki, D. Ma, N. Yasuo and M. Sekijima, J. Chem. Inf. Model., 2024, 64, 7291–7302 Search PubMed.
  22. M. Retchin, Y. Wang, K. Takaba and J. D. Chodera, DrugGym: A testbed for the economics of autonomous drug discovery, biorxiv, 2024, preprint,  DOI:10.1101/2024.05.28.596296v1.
  23. B. G. Szczepankiewicz, C. Kosogof, L. T. J. Nelson, G. Liu, B. Liu, H. Zhao, M. D. Serby, Z. Xin, M. Liu, R. J. Gum, D. L. Haasch, S. Wang, J. E. Clampit, E. F. Johnson, T. H. Lubben, M. A. Stashko, E. T. Olejniczak, C. Sun, S. A. Dorwin, K. Haskins, C. Abad-Zapatero, E. H. Fry, C. W. Hutchins, H. L. Sham, C. M. Rondinone and J. M. Trevillyan, J. Med. Chem., 2006, 49, 3563–3580 Search PubMed.
  24. J. Močkus, Optimization Techniques IFIP Technical Conference Novosibirsk, Berlin, Heidelberg, 1975, pp. 400–404 Search PubMed.
  25. D. R. Jones, M. Schonlau and W. J. Welch, J. Global Optim., 1998, 13, 455–492 Search PubMed.
  26. M. Emmerich, K. Giannakoglou and B. Naujoks, IEEE Trans. Evol. Comput., 2006, 10, 421–439 Search PubMed.
  27. M. García-Ortegón, G. N. C. Simm, A. J. Tripp, J. M. Hernández-Lobato, A. Bender and S. Bacallado, J. Chem. Inf. Model., 2022, 62, 3486–3502 CrossRef PubMed.
  28. J. Sun, N. Jeliazkova, V. Chupakhin, J.-F. Golib-Dzib, O. Engkvist, L. Carlsson, J. Wegner, H. Ceulemans, I. Georgiev, V. Jeliazkov, N. Kochev, T. J. Ashby and H. Chen, J. Cheminf., 2017, 9, 41 Search PubMed.
  29. G. R. Bickerton, G. V. Paolini, J. Besnard, S. Muresan and A. L. Hopkins, Nat. Chem., 2012, 4, 90–98 Search PubMed.
  30. R. Gorantla, A. Kubincová, B. Suutari, B. P. Cossins and A. S. J. S. Mey, J. Chem. Inf. Model., 2024, 64, 1955–1965 CrossRef CAS.
  31. S. Allenspach, J. A. Hiss and G. Schneider, Nat. Mach. Intell., 2024, 6, 124–137 Search PubMed.
  32. T. Yin, G. Panapitiya, E. D. Coda and E. G. Saldanha, J. Cheminf., 2023, 15, 105 Search PubMed.
  33. P. H. Bos, E. M. Houang, F. Ranalli, A. E. Leffler, N. A. Boyles, V. A. Eyrich, Y. Luria, D. Katz, H. Tang, R. Abel and S. Bhat, J. Chem. Inf. Model., 2022, 62, 1905–1915 CrossRef CAS.
  34. H. H. Loeffler, J. He, A. Tibo, J. P. Janet, A. Voronov, L. H. Mervin and O. Engkvist, J. Cheminf., 2024, 16, 20 Search PubMed.

This journal is © The Royal Society of Chemistry 2026
Click here to see how this site uses Cookies. View our privacy policy here.