Structure- and Ligand-Based 3D QSAR Protocol. Our 3D-QSAR analysis represents a valuable medicinal chemistry tool, and the ever-increasing information from structural biology will surely present valuable feedback to the assumptions that form the basis of 3D-QSAR methods. Email*: Username*: Password*: Confirm Password*: First Name*: Address*. CRM Customer Service Customer Engagement. Tutorial for the > OPEN 3D QSAR software as the information given in the workflow on the > official website is not enough. In silico ADME/T and 3D QSAR analysis of KDR inhibitors S. Zahid Hosen 1*, Raju Dash, Mahmuda Khatun1. SCF-MO method at PM3 level of theory and various SlogP descriptors from MOE software. In silicoK ADME/T analysis was performed by using different software’s Discovery Studio 2.5, TOPKAT and QikProp 4.3.
Abstract
Quantitative Structure-Activity Relationship modeling is one of the major computational tools employed in medicinal chemistry. However, throughout its entire history it has drawn both praise and criticism concerning its reliability, limitations, successes, and failures. In this paper, we discuss: (i) the development and evolution of QSAR; (ii) the current trends, unsolved problems, and pressing challenges; and (iii) several novel and emerging applications of QSAR modeling. Throughout this discussion, we provide guidelines for QSAR development, validation, and application, which are summarized in best practices for building rigorously validated and externally predictive QSAR models. We hope that this Perspective will help communications between computational and experimental chemists towards collaborative development and use of QSAR models. We also believe that the guidelines presented here will help journal editors and reviewers apply more stringent scientific standards to manuscripts reporting new QSAR studies, as well as encourage the use of high quality, validated QSARs for regulatory decision making.
Where have you been?
Where are you going to?
I wanna know what's new
I wanna go with you.
Chris Rea, The Blue Café, 1998
Introduction
More than fifty years have passed since the field of Quantitative Structure-Activity Relationships (QSAR) modeling was founded by Corwin Hansch.1 Initially conceptualized as the logical extension of physical organic chemistry, QSAR modeling has grown, diversified, and evolved from application to small series of congeneric compounds using relatively simple regression methods to the analysis of very large datasets comprising thousands of diverse molecular structures using a wide variety of statistical and machine learning techniques. More than fifty years of continuous improvements, interdisciplinary breakthroughs, and community-driven developments were needed to make QSAR one of the commonly employed approaches to modeling the physical and biological properties of chemicals in use today. In fact, the analysis of published literature indicates that the continuing growth of chemical data and databases especially in the public domain has stimulated the concurrent growth in QSAR publications (Figure 1).
The growth of QSAR modeling is caused by the growth of experimental data. Chart is generated by Google Ngram Viewer (http://books.google.com/ngrams); Y-axis – percentage among all books in the Google Ngram database, X-axis – years.
QSAR modeling is widely practiced in academy, industry, and government institutions around the world. Recent observations suggest that following years of strong dominance by the structure-based methods, the value of statistically-based QSAR approaches in helping to guide lead optimization is starting to be appreciatively reconsidered by leaders of several larger CADD groups. QSAR models find broad application for assessing potential impacts of chemicals, materials, and nanomaterials on human health and ecological systems. An area of active QSAR expansion is in the use of predictive models for regulatory purposes by government agencies, where a still growing number of specialized regulatory tools and databases are being developed and validated.
Obviously, QSAR modeling is a computational field but its major beneficiaries and the ultimate judges are medicinal chemists. Whenever computational scientists begin to address problems within a primarily experimental scientific domain there is always a challenge of finding the proper interface and balance between computational and the respective experimental domain expertise. Of course, it is computational scientists who are expected to create computational tools that would be valuable for experimental researchers. One may then pose a question as to once the tools are created who should use them: still the 'developers' who have learned enough of the domain expertise to make the application of the tools meaningful or the 'users' who have learned enough of computational science to use the tools properly. We believe that generally speaking there is no definitive answer to this question as the answer depends on the level of expertise and experience of the researcher. However, we would like to point out that both fields are highly sophisticated and it will require a significant effort to accumulate high level of knowledge in both areas. In fact, very few scientists have achieved prominence both as computational and experimental medicinal chemists. We tend to think that the highest level of success will be enabled through the collaboration between computational and experimental scientists who have deep knowledge of their respective fields but have also made efforts to understand and develop working knowledge of the complimentary field.
This paper is written by a group of cheminformatics experts with a deep knowledge of the theory of QSAR modeling as well as extended experience in various types of QSAR applications, especially in designing compounds with the desired biological activities. Our objective is to provide an overview of the state-of-the-art QSAR methods and applications to a diverse community of readers, including by default mostly experimental medicinal chemists but also scientists practicing QSAR modeling. Therefore, we spend fair amount of effort to describe major concepts and methodologies in the field (as they have emerged throughout the entire history of QSAR) but then emphasize the applications of QSAR in medicinal chemistry. We describe several studies (published in J. Med. Chem., Nature Chem. Biol., and other high-profile journals) where cheminformaticians and medicinal chemists worked together to discover novel molecules with unique biological activities; this was achieved by developing QSAR models and employing them for virtual screening followed by experimental validation. In our opinion, these case studies were successful because of the distribution of labor where computational experts ensured the highest quality of models and experimental chemists guided by the predictions were in a position to skillfully synthesize and test compounds predicted as hits. We posit that such examples are of high value for all readers of this Journal because they prove the capabilities of computational techniques to guide chemical synthesis and biological testing when the number of possible experiments significantly exceeds technical capabilities of researchers (i.e., it is impossible to synthesize and/or test all compounds). We further suggest that both developers and users of QSAR modeling should employ or be familiar with the best practices for data curation, analysis, and modeling: the former group should follow them and the latter group should be aware of them to evaluate if they can rely on a particular model in planning their experiments. Therefore, we have endeavored to reach out to the broadest readership of this Journal and help readers to understand and appreciate the state-of the-art practices in the QSAR field, as well as its successes, and, most importantly, its limitations.
The amount of chemogenomics data generated by experimentalists is exploding. Databases incorporating millions of compounds with associated bioactivities are available in the public domain, whereas HTS platforms are becoming more and more common in academic structures. We believe that computational approaches are critical for accessing, querying, mining, modeling, screening such enormous ensemble of chemical data, and not only cheminformaticians can use those approaches but medicinal chemists as well. In such context, our review is an overall, recapitulative summary of what QSAR modeling was at its origin, what QSAR modeling is today, and what QSAR modeling is likely to be in the next coming years. We believe it is of high interest not only for cheminformaticians and medicinal chemists but for both computational and experimentalist communities in general.
As in any evolving field, QSAR has experienced successes, suffered failures, and responded to emerging trends. This paper aims to discuss: (i) the historical development of QSAR (Part 1) including the founding pioneers, initial concepts, and important milestones in the evolution of the field; (ii) current trends, unsolved problems, and pressing challenges (Part 2); and (iii) several novel and emerging applications of QSAR science (Part 3). Throughout this discussion, in an effort to build on past lessons learned, we provide some guidelines for use and application, and recommend best practices for developing validated and externally predictive QSAR models.
Obviously, it is not possible to address all aspects of this rich and expanding field and acknowledge all contributions made over the years by many of our outstanding colleagues. Thus, this paper should not be perceived as a comprehensive monograph covering the entire discipline of QSAR modeling. Rather, our international group of co-authors working in industry, government agencies, and academia has made an attempt to share our expertise and collective wisdom concerning some most important, in our opinion, general aspects and best practices of building, validating, and employing QSAR models using examples mostly drawn from our own research. We hope that the readers will both gain an appreciation for the challenges of developing truly rigorous and useful QSAR models (i.e., 'will wanna know what’s new'), as well as share the excitement of the authors concerning new opportunities offered by this evergrowing research area (i.e., 'will wanna go' into the field!).
1. History and evolution of QSAR
1.1. In principio erat verbum, et verbum erat QSAR
Origin of QSAR
Many mark the founding of modern QSAR practice to the 1962 publication of Hansch et al.1 This paper represented the culmination of a fifteen year struggle to understand the basis of the structure-activity relationships (SARs) of plant growth regulators – most of that time spent in the pursuit of a suitable Hammett relationship, the reigning methodology for explaining substituent effects on chemical reactivity. Unable to obtain a robust relationship, Hansch followed up on the arguments of Veldstra3 and investigated the effect of lipophilicity on biological potency.4 He wisely ignored Veldstra’s complex methodology and turned to octanol-water partition coefficients5 to serve as a surrogate measure of lipophilicity.
In the meantime, Fujita was employing quantum-chemical calculations to account for activity variations in plant growth regulators.6 On the recommendation of Fukui, Fujita accepted a post-doctoral fellowship at Pomona in mid-1961 and started to experimentally measure octanol-water partition coefficients (logP). Hansch and Fujita soon realized that logP was an additive property, i.e., that the partial contribution of a substituent to the logP of one molecule is often the same as the contribution of that substituent to the logP of another molecule. They used the term π for this substituent effect on hydrophobicity.7
Of course, Hansch and Fujita did not work in a scientific vacuum: in the 1950s, the power of the Hammett equation to account for reactivity differences dominated the explanation of substituent effects. A 1953 article by Jaffe included several hundred of such equations.8 In the late 1950s, Taft extended the concept of linear free energy relationships to propose and fit an equation that included not only electronic effects of substituents but also steric effects.9
In contrast, biochemical pharmacologists at the time were focusing on the effect of partition coefficients of molecules on drug absorption. This approach can be traced back to Overton11 and Meyer12 some fifty years earlier and by Collander in the 1930s.13 Notably, Fieser, an eminent organic chemist of the mid-1900s, showed graphically the relationship between the antimalarial potency of naphthoquinones and their ether-water distribution coefficients. He also observed a constant optimum lipophilicity for different series of molecules.
Although Kauzmann’s 1959 review article prompted biochemists to endorse the central role of hydrophobicity in determining protein structure, with early work further emphasizing the role of partitioning to the biological target, hydrophobicity as a governing factor in the biological potency of small molecules only gradually entered the vocabulary of QSAR. Having found that the relationship between logP and biological potency was no clearer than was that between Hammett’s sigma (σ) and potency, Hansch and Fujita included both terms in a new equation.4 The publications that followed successfully demonstrated a computational approach to modeling quantitative effects of substituents on potency.16,17 Part of the attraction of the work is that the substituent effects were based on model equilibria, partition coefficients, and pKa that are easy to understand. In addition, values for these substituent effects were found to be largely transferable from one series of molecules to another. Some of the attention to the publication was that the fit involved the use of a computer, the power of which was only then becoming appreciated. Less attention in early work was paid to the power of statistics to distinguish between possible explanations.
Evolution of QSAR
QSAR concepts have long been used in the design of medicinal chemistry series. Craig suggested that the properties of possible substituents be plotted versus each other and that substituents be chosen to sample the full range of plotted values. Topliss invented schemes for the stepwise exploration of a series according to the physical chemical property that governs the increase or decrease in potency of each new compound. Hansch and Unger used cluster analysis to group possible aromatic substituents, suggesting that a good series contain one molecule from each cluster.
In 1956, Fujita and Hansch published tables of π values generated from careful measurements of logP. These were useful for calculating the relative logP of members of a series. However, comparing the optimum logP value within a series would still require the logP of the corresponding parent molecules. To address this need, Hansch and Leo used an extensive database of experimental logP values for hundreds of diverse parent molecules to parameterize a fragment-additive approach for predicting logP that was automated in the CLOGP program. Over the years many other methods for computing logP have been devised but the fragment-additive approach remains one of the most commonly employed.
The parabolic relationship between potency and logP did not fit all datasets. Hence, Kubinyi suggested a bi-linear equation that allows for different slopes at low and high logP values. At approximately the same time, Martin and coworkers observed the same bi-linear property of model-based equations of ionizable compounds, where both the logP and pKa vary within the series. A key feature of these equations is that the relationship may be independent of logP at high logP values.
As the Hansch-type approach to QSAR was being established, more fundamental quantum-mechanical calculations were becoming increasingly feasible, providing an alternative means for exploring electronic and steric determinants of activity among closely related chemicals. In one of the earliest examples, the Pullmans showed how the carcinogenic potential of aromatic hydrocarbons is related to their electronic structure in predicting the potential for bay-region metabolic activation to DNA-reactive diol-epoxide intermediates. Further work in this area was limited by the computational demands of quantum chemical calculations, the greater expertise required, and the targeted nature of these methods of modeling local stereo-electronic features. Hence, the methods did not easily lend themselves to more globally applicable, higher throughput QSAR methods for evaluating diverse chemical structures.
In the early 1980’s, Klopman developed an approach to break a molecule into constituent 2D fragments, to auto-generate such fragments for large numbers (hundreds to thousands) of molecules in a training set, and to attempt to correlate the frequency of each of these fragments with biological activity.24 The approach was a breakthrough at the time in that it created an efficient computational method for representing and correlating easily interpretable structural features for a large number of chemicals; hence, it began to tackle the challenge of creating so-called global QSAR models for prediction of biological activity.
The treatment of steric effects of substituents had been another long-standing problem in traditional QSAR. Early investigations used Taft Es values. Hansch and Kutter as well as Charton26 showed that Es parameters of symmetric substituents were related to the radius of the substituent. Verloop and colleagues27 took this further and calculated five shape descriptors for substituents. Hansch also frequently used molar refractivity of the substituent as a measure of its bulk. It was not until the development of CoMFA (Comparative Molecular Field Analysis) and other 3D approaches, however, that electrostatic potential interaction energies across a series of related, superimposed structures were effectively taken into consideration. CoMFA was the first successful demonstration of a 3D QSAR technique for correlating molecular field approaches with biological activities; it was also the first commercial product that employed the partial least square (PLS) method.29
As datasets became larger and more structurally diverse, descriptors that were designed to be applied within a common reaction mechanism framework were no longer sufficient. The simplest solution was to use indicator variables to distinguish one series (i.e., assumed mechanism) from another. However, a more general solution was to generate molecular descriptors that could serve the same purpose as indicator variables by delineating series of molecules whose activity was more likely to be governed by common mechanisms: for instance, the CASE fragments31 and the widely used molecular connectivity indices. They do not discard the notion of using descriptors to directly model activity within a series, but rather attempt to capture more general determinants of activity variations both across and within series based on the principle that the entire chemical structure of a molecule dictates its properties. The work of Rosenkranz and Klopman in extending the application of CASE (now CASETOX), and later MultiCASE, to a wide diversity of toxicity modeling challenges exemplifies the power of the substructure-based approach to model previously unexamined activities and to elucidate the structural basis of activity, both across and within congeneric series of chemicals. More recently, in drawing on the more general aspects of this approach, new use has been made of descriptors that originate from the field of substructure searching; for example MDL MACCS keys, and circular fingerprints.
The traditional application of QSAR to a series of congeners requires that each molecule in the dataset has measurable biological activity, i.e., a quantitative non-zero potency value. Discriminant analysis or a logistic regression method can be applied more generally to modeling binary or categorical responses (e.g., active/inactive, mutagenic/non-mutagenic). Recursive partitioning enables the consideration of many more descriptors and larger sets of molecules and, as a result, has become the basis for newer classification methods, such as random forest.36 The support vector machines and Bayesian classifiers have also been successfully applied.
The field has also progressed in the development of objective methods for assessing models' reliability and prediction confidence. Statistical procedures were adopted early on to avoid chance correlations, e.g., using random variable simulations, as explained in the seminal paper by Topliss and Edwards. Although much of this early work concentrated on the fit of data to an equation, researchers now evaluate models by their ability to accurately forecast activity for test set molecules that were not used in model development.
Quantum chemistry remains a powerful tool for exploring fundamental reactivity determinants in QSAR, as well as for calculation of ab initio properties (e.g., dipole moment) or whole-molecule reactivity indicators, such as EHOMO and ELUMO. Its application at higher levels of ab initio theory however is typically limited to isolated systems (gas phase) and relatively few molecules. Quantum chemistry and its semi-empirical and molecular mechanics implementations are currently used in 3D QSAR approaches, such as in silico virtual ligand screening and profiling in drug discovery. In addition, such methods can be used to examine the relative stability of conformers, which, in turn, can influence 3D-dependent properties employed in QSAR. Mekenyan and coworkers have incorporated conformation-dependent property distributions into QSAR approaches to model toxicological endpoints, demonstrating their importance and relevance to QSAR modeling.
In more than 50 years of active development, the field of QSAR modeling has grown tremendously with respect to the diversity of both methodologies and applications. The subsequent sections of this review provide more details concerning some recent evolutions and current trends in the QSAR field, future directions, and some innovative applications of models in multiple domains as diverse as the prediction of bioprofiles of nanomaterials or the calculation of chemical and physical properties of molecular mixtures.
1.2. Molecular descriptors
Chemical descriptors are at the core of QSAR modeling and so many different types of chemical descriptors reflecting various levels of chemical structure representation have been proposed so far. These levels range from molecular formula (so-called 1D) to the most popular among chemists two-dimensional structural formula (2D) to three-dimensional, conformation-dependent (3D) and even higher levels taking into account mutual orientation and time-dependent dynamics of molecules (4D and higher; for discussion see study of Polanski). A comprehensive collection of molecular descriptors, along with their definitions, mathematical formulas, examples, and references, was published in 2009.43 We briefly discuss here the most popular 2D descriptors to provide the necessary context for QSAR as a tool to predict bioactivity of compounds from their structure followed by the discussion of most popular and evolving 3D QSAR approaches.
2D descriptors for QSAR modeling (2D QSAR)
The two-dimensional representation of a molecule, commonly referred to as topological representation, defines the connectivity of atoms in the molecule in terms of the presence and nature of chemical bonds. Such two-dimensional representation enables the definition of so-called molecular 2D-descriptors. The main advantages of these QSAR parameters are that they (i) contain simple and useful information about the molecular structure, (ii) are invariant to molecule roto-translation, and (iii) can be calculated avoiding structure optimization. In general, 2D descriptors do not uniquely characterize molecular topology, and, as a consequence, they do not always allow the reconstruction of the molecule. Therefore, suitably defined ordered sequences of 2D descriptors can be used to characterize molecules with higher discrimination.
A molecular graph is a topological representation of a chemical compound; it is usually denoted as G = (V, E), where V is a set of vertices which correspond to the molecule atoms and E is a set of elements representing the binary relationship between pairs of vertices; unordered vertex pairs are called edges, which correspond to bonds between atoms. When a molecular graph is obtained excluding all hydrogen atoms, it is called an H-depleted molecular graph, whereas a molecular graph where also hydrogens are included is called H-filled molecular graph (or, simply, molecular graph).
Graph-theoretical matrices are the most common mathematical tool to encode structural information provided by molecular graphs, a huge number of which were proposed in the last decades. Graph-theoretical matrices can be both vertex matrices, if both rows and columns refer to graph vertices (atoms) and matrix elements encode some property of pairs of vertices, or edge matrices, if both rows and columns refer to graph edges (bonds) and matrix elements encode some property of pairs of edges. Off-diagonal entries of the matrix encode different information about the pairs of vertices such as their connectivity (adjacency matrix), topological distances (distance matrix), and sums of the weights of the atoms along the connecting paths (weighted matrices). The matrix diagonal entries can be equal to zero or encode chemical information about the vertices (augmented matrices); besides the most common atomic properties, also Local Vertex Invariants, which are numerical quantities derived from the molecular topology and used to characterize properties of molecule atoms (e.g., vertex degree, vertex distance sum, atom eccentricity), are frequently encountered as the atomic weightings.
Graph invariants are mathematical quantities derived from a graph representation of the molecule and representing graph-theoretical properties that are preserved by isomorphism, i.e., properties with identical values for isomorphic graphs. A graph invariant may be a characteristic polynomial, a sequence of numbers, or a single numerical index obtained by the application of algebraic operators to graph-theoretical matrices and whose values are independent of vertex numbering or labeling. In the last few years, several formulas and algorithms dealing with molecular graph information have been proposed and applied to different molecular matrices and various weighting schemes, leading to several new classes of related graph invariants.
Single indices derived from a molecular graph are usually called topological indices. These are numerical quantifiers of molecular topology that are mathematically derived in a direct and unambiguous manner from the structural graph of a molecule. They can be sensitive to one or more structural features of the molecule such as size, shape, symmetry, branching, and cyclicity and can also encode chemical information concerning atom type and bond multiplicity. In fact, topological indices are usually divided into two categories: topostructural and topochemical ones. Topostructural indices encode only information about the adjacency and distances between atoms in the molecular structure; topochemical indices quantify information about topology but also specific chemical properties of atoms such as their chemical identity and hybridization state. Topological indices are mainly based on distances between atoms calculated by the number of intervening bonds and are thus considered through-bond indices; they differ from geometrical descriptors which are, instead, considered through-space indices because they are based on inter-atomic geometric distances.
Another important class of graph invariants is represented by the so-called autocorrelation indices. They were first introduced by Moreau-Broto45 in 1980 to define a relationship between atoms as a function of their spatial separation; a review of the autocorrelation descriptors is given by Consonni and Todeschini.46 The most common autocorrelation descriptors can be obtained by taking the molecule atoms as the set of discrete points in space and an atomic property as the function evaluated at those points. The autocorrelation descriptor is then the integration of the products of the function calculated at atom×and atom x+k, where k is the lag, i.e., the topological distance. This descriptor expresses how numerical values of the function at intervals equal to the lag are correlated.
Randić suggested47 a list of attributes for topological indices that should: 1) have structural interpretation, 2) have good correlation with at least one property, 3) preferably discriminate among isomers, 4) can be applied to local structure, 5) be preferably independent, 6) be simple, 7) not be based on experimental properties, 8) not be trivially related to other descriptors, 9) be possible to construct efficiently, 10) use familiar structural concepts, 11) have the correct size dependence, 12) change gradually with gradual change in structures. Most of the topological descriptors have the above characteristics, which is why they have been prolifically applied in characterizing the structural similarity/dissimilarity of molecules and in QSAR/QSPR modeling.
3D descriptors for QSAR modeling (3D QSAR)
Taken literally, many QSAR expressions suggest that biological selectivity results from each target forming highly specific interactions such as hydrogen bonds with a ligand. However it also seems that ligands binding preferences emerge primarily from non-covalent field effects exerted in the spatial vicinity of those ligands. Thus, systematic sampling of those field differences should yield molecular descriptors particularly well-suited for QSAR. This was the vision that motivated the creation of 3D QSAR, in its original and still predominant CoMFA formulation.
The most important challenge related for CoMFA implementation was the conflict between the thousands of molecular descriptors needed to sample a ligand field and the few biological responses, almost a heretical situation when contrasted with the familiar precepts of good QSAR practice, as detailed elsewhere in this review. Thus the critical event in CoMFA’s development was a private discussion at the 1982 QSAR Gordon Conference, where Wold first expounded the PLS approach to Cramer.29 More than ten thousand references to '3D QSAR' and/or 'CoMFA', which a Google Scholar search today produces, including consistent reports of successful potency predictions suggests that PLS has indeed been useful in describing biological differences as effects of ligand field differences., However, the contrasts between 3D QSAR methodology and those QSAR precepts do remain: in addition to that sacrilegious descriptor-to-structure count ratio, a CoMFA model requires high collinearity of ligand field variations for its robustness, whereas auto-scaling 3D QSAR’s molecular descriptors can prevent successful model generation.
The biggest challenge in performing CoMFA is related to the alignment protocol of training- and test-set ligands, as well as the selection of both conformation and orientation of each ligand to be included in the QSAR model. In practice, this task can often become a slow, tedious, and somewhat ad hoc pursuit of higher statistical criteria (e.g., q2 value). However, considering many uncertainties (including the biological potencies being fit), the coarseness of the ligand field sampling, the structural representativeness of any training set selection, as well as unavoidable subjectivity of individual ligand alignments, such emphasis on q2 as a metric of comparative model quality (as opposed to a threshold of model acceptability) is inappropriate. It should instead be appreciated that every statistically acceptable QSAR model, resulting from the systematic variation of the training set composition (and alignment), represents a valid alternative interpretation, and that exploring any differences among such models may improve the understanding of the causative effects of field differences. Of course, the potential benefit of such exploration also depends upon the cost of generating each individual model, favoring automatic and robust alignment methodologies.
The requirements of 3D QSAR modeling to operate on 3D descriptors and PLS methodology contributed to the availability of its methodological extensions dependent on specialized software vendors. Thus, CoMSIA, which extends ligand field varieties from CoMFA’s steric and electrostatic to hydrogen-bonding and hydrophobic effects, is available from Tripos. Use of pharmacophoric constraints to facilitate 3D QSAR considering multiple conformations, based on the original molecular field generating software GRID, is provided by a family of programs from Molecular Discovery., Cresset software employs extrema in ligand fields as guides in ligand alignment. Schrodinger package offers PHASE as its 3D QSAR capability. Two approaches that use a complementary biological target field to refine ligand-based 3D QSAR models are COMBINE and AFMoC programs; two other pioneering 3D QSAR methodologies, HASSLE and MSTD also deserve mentioning. Finally, though more 'pseudo-receptor' –like rather than 3D QSAR approach (the latest incarnation of the pioneering COMPASS 3D QSAR methodology) QMOD, shows noteworthy promise.
Notably, none of these innovations, however useful, have greatly alleviated 3D QSAR’s ligand-alignment bottleneck. Indeed, as our understanding of the physics of ligand-target interaction improves, the criteria for physicochemically meritorious alignments become less certain. Even a target-bound ligand structure, which 3D QSAR cannot compete with as a source for structural inspiration, is undermined as the 'gold standard' for 3D QSAR by an increasing awareness that the dynamic ligand interactions that may be critical for selective in vivo activity can be invisible in the non-physiologically static crystalline environment.
In considering and developing any methodology we should not lose sight of fundamental basis of any QSAR investigation: the only possible cause of any difference in biological activity of two structures is their structural difference – regardless of how complex may be the physicochemical and biological interactions that subsequently connect that structural difference to an observed biological effect. And the basis for 3D QSAR is that this causative difference in ligand structure is best expressed, primarily and initially, as a causative difference in ligand fields. These considerations suggest a different goal in ligand alignment for the purpose of 3D QSAR: to minimize the relative importance of grid locations distant from any ligand structural difference and therefore non-causative, by aligning training and test set ligands in a way that maximizes the steric overlap of their structurally identical moieties.
Accordingly, two new methods have emerged for the 3D QSAR alignments, the topomer protocol and still evolving 'template' protocol. Both approaches are extremely facile, in effect converting 3D QSAR from one of the most tedious and therefore costly CADD approaches into one of the easiest, on the verge of becoming almost completely automatable. More importantly, the reported predictive performances of the topomer protocol in discovery projects were uniformly encouraging. The topomer similarity has consistently forecasted biological similarity, for example in published 'lead-hopping' and off-target applications. And the standard error of pIC50 taken from 144 predictions followed by synthesis and testing in four different discovery organizations, was reported with extraordinarily low 0.6 value (or, expressed as an error in predicted potency ratios, 4x).–
The practice of 3D QSAR is inherently limited to local models (as herein defined elsewhere). However, it can be expected, that with the latest explosive expansion of public databases such as ChEMBL and PubChem and with further evolution of alignment protocols, that limitation will slowly recede. One encouraging indicator of that is the use of 3D QSAR (topomer CoMFA) models derived entirely from PubChem SAR data to successfully overcome the cytochrome P450 (CYP) liability in drug design.
1.3. Challenges in QSAR modeling
Since the seminal paper by Hansch and co-workers,1 great strides have been made in successful application of QSAR modeling as well as in the development of QSAR methodology itself, reflected in numerous articles published in the last decade.–73 Such efforts in developing new methods have provided guidance on recommended procedures to be followed for optimal results. Unfortunately, despite these efforts, prediction errors due to poor application of statistical methods and recommended guidelines perpetuate in the development and use of QSARs. In that regard, Dearden et al. in 2009 reported and discussed a total of 21 types of such problems; these are briefly reiterated here (for supporting references to the examples below, see).
Failure to take account of data heterogeneity
Sometimes in vivo data from two or more different species (or obtained with different protocols) are used in the same QSAR model. It may be that it is necessary to obtain an adequately-sized dataset, but it potentially compromises the integrity of the model and should be avoided if at all possible. It may seem that such considerations do not apply to the QSPR (Quantitative Structure-Property Relationship) modeling of physical-chemical properties, but this is not necessarily the case. For example, aqueous solubility can be determined in pure water, or as un-dissociated species (intrinsic solubility), or at specified pH. Another example - the flash point values, which are dependent on the size of the sample, the heating rate, the use of open- or closed sample cup, the presence or absence of stirring, and the energy and type of ignition source (e.g., spark or flame).
Use of inappropriate endpoint units
It is often not recognized that for most QSAR purposes related to molecular activities, the endpoint format should be expressed in molar units (e.g., mol.kg−1), not weight units (e.g., mg.kg−1); otherwise strict molecule-to-molecule activity comparisons are not possible. If, however, data are reported as, say, percentage response at a fixed dose of 10 mg.kg−1, then such a conversion is not possible, and the data may not be appropriate for QSAR interpretation.
Use of confounded descriptors
If two descriptors in a given QSAR solution are highly collinear, they contribute the same information twice, thus confounding the statistical association and making it more difficult to deduce a mechanistic interpretation of the QSAR. Thus, Dearden et al. showed that, although two highly correlated topological descriptors correlated well separately, and with positive coefficients, the toxicity of alkyl ethers to mouse, when the two descriptors were used together, one had a negative coefficient. There is no definitive value above which collinearity is unacceptable, but Dearden et al. recommend that r2 values of >0.8 be carefully scrutinized.
Use of non-interpretable descriptors
One of main values of a QSAR model is possible insight into mechanism of action under study, although it must always be recognized that a correlation does not provide any guarantee of causality. Currently, there are thousands of molecular descriptors available for QSAR purposes, and it is difficult to discern a clear physical-chemical interpretation for many of them. Sometimes descriptors involved in reported QSAR models are not clearly defined or identified, and often no mechanistic interpretation is given. Modelers should employ descriptors that are reasonably interpretable, particularly when the aim of a study is to yield insight into mechanisms of action or where such association can improve model’s acceptance and use.
Errors in descriptor values
With few exceptions (e.g., atom counts and molecular weight, provided that they are correctly calculated), descriptor values, whether measured or calculated, can contain errors. For example, for a simple organic chemical, 4-nitrophenol, seven published measured logP values range from 0.76 to 2.08, while seven logP values calculated from available software (VCCLAB, www.vcclab.org) range from 1.35 to 1.93, with the generally accepted 'best' measured value being 1.91. In cases where experimental values are debatable, scientific practice requires reporting error bars or uncertainties in modeled values where these are known or can be estimated. In other cases, blatant mistakes in structure representations and resultant computed properties provide yet additional sources of model errors.
Poor transferability of QSARs
It is necessary that any published QSAR model can be independently validated and used by others for predictive purposes. Transferability of QSAR is the equivalent of 'lateral validation', a concept pioneered by Hansch et al.74 Later, Hartung et al. suggested five criteria for transferability, the prime one being that descriptor values can be reproduced. Often, however, those criteria are not satisfied either due to inadequate documentation or lack of readily available descriptor-generating software.
Inadequate or undefined applicability domain
The applicability domain (AD) of a QSAR model has been defined as 'the response and chemical structure space in which the model makes predictions with a given reliability'. Essentially this means that a test chemical must be reasonably similar to some training-set chemicals, or valid prediction cannot be expected. However, very few QSAR publications actually provide sufficient information (e.g., training set structures and descriptor values) for the assessment of model applicability domain.
Unacknowledged omission of data points
Many QSAR data sources originate from the literature, and often only a selected number of data points are used in the model. However, it is sometimes the case that data are pruned without explanation and this can lead to the suspicion that pruning (or removal of outliers) was done to improve the model. Of course, outliers can occur in any data set, and it is acceptable to remove them, provided that a good explanation, independent of the modeled result, can be given for their removal.70
Use of inadequate data
Data inadequacy includes matters already mentioned, such as heterogeneity, use of inappropriate units, lack of information on the AD, and unacknowledged omission of data points. It can also include the use of incorrect, misspelled, or insufficiently defined chemical names, incorrect CAS numbers, and incorrect structures. For instance, a QSAR study of skin absorption77 listed chloroxylenol and 4-chlorocresol among the chemicals examined; chloroxylenol has 18 structural isomers and 4-chlorocresol has two. Young et al.78 found that incorrect structures in six public and private databases ranged from 0.1% to 3.4%.
Replication of chemicals in a data set
It is unfortunately quite common for chemicals to be replicated in the training and the test sets.79 Such co-occurrence can falsely improve the apparent predictive power of the QSAR solution. Replication can occur because of different names, CAS numbers, or structure codes for the same chemical, or because of different activity or property values for the same chemical. Oftentimes, however, replication occurs upon indiscriminate 'desalting' of a structure file prior to descriptor generation, after which the parent and salt (with the same or different activities) map to the same structure. It is therefore essential that datasets are carefully checked and curated (including merging or removal of duplicates) before use. More details on chemical data curation are provided in Section 2.1.
Narrow range of endpoint values
The range of endpoint values of a QSAR training set should be significantly greater than the experimental error in the values. The experimental error among in vivo data can often exceed half a log unit; as a rule of thumb, Gedeck et al. recommend an endpoint value range of at least 1.0 log unit to obtain a good QSAR model. Of course, it is not always possible to achieve such a wide range of endpoint values, either through paucity of data or because of the nature of the endpoint (e.g., melting points of a given class of chemicals). In such cases, closer consideration of the data and model performance statistics, including external validation, are required.
Over-fitting of data
The well-known 'Topliss and Costello rule' states that, to minimize the risk of chance correlations, the ratio of training set chemicals to descriptors should be at least 5:1 when using simple linear regression methods. This rule, as well as the standard requirements for basic statistical practices, is still widely broken. A glaring example is provided by the modeling of the aquatic toxicities of 12 aliphatic alcohols with 9 descriptors.81 With the use of non-linear multivariate techniques, such as artificial neural networks, over-training and over-fitting can be a problem if care is not taken, although rigorous statistical tests are available allowing for the proper processing of cases where the number of descriptors far exceeds the number of chemicals.
Use of excessive numbers of descriptors in a QSAR
Even if the 'Topliss and Costello rule' is not broken, use of a large numbers of descriptors in a QSAR can make interpretation and explanation the model more difficult. Generally speaking, the principle of 'Occam's Razor' is completely applicable to QSAR, i.e., one seeks a QSAR with the smallest number of descriptors that yield a reasonable model. Often, a small number of simplest molecular descriptors affords a model that outperforms significantly more complex ones. For instance, Oprea82 reported that the length of the molecule gave the best correlation with fiber affinity; this simple model even outperformed CoMFA.
Inadequate or missing statistic measures
The statistical measures are used to indicate the goodness-of-fit and predictivity of a QSAR, and so are vital for assessing its validity. However, even now QSARs are reported without any statistics (e.g., see the study of Ghasemi and Saaidpour), and many still appear with inadequate statistics, which makes the assessment of model’s validity difficult. As well as the correlation (R) and determination (R2) coefficients, and standard error of the estimate(s), it is useful to have the adjusted determination coefficient (R2adj), which allows for comparison between QSARs with different numbers of descriptors and can indicate when a given QSAR model incorporates too many descriptors. In addition, the internally cross-validated R2 (Q2) and the Fisher statistic or variance ratio provide an indication of a chance correlation. In addition, the regression coefficients of each of the descriptors, although rarely reported, are valuable for indicating whether a particular descriptor contributes significantly to a linear regression. The probability that the descriptor is there by chance, should generally be less than 0.05 (i.e., 5%) to be considered statistically significant; otherwise this descriptor should be discarded.
Incorrect calculation
Editors and reviewers usually assume that all calculations reported in a manuscript have been made correctly, and it is often impossible to check otherwise. But nonetheless incorrect published QSARs have been identified to date, and it is difficult to assess how widespread this problem actually is. Dearden et al. reported two such instances, and there are probably many more.
Lack of descriptor auto-scaling
Descriptor values often cover widely different numerical ranges, which necessitates the use of auto-scaling methods. When no scaling is employed, it is difficult to determine the relative contribution of each descriptor to the QSAR and those descriptors with large numerical values can dominate the model compromising its statistical validity. Many published models do not employ auto-scaling, but its use is highly recommended.
Misuse or misrepresentation of statistics
Even if QSAR practitioners are not statisticians, the basic rules of good statistical practice should be used by all and should be enforced by reviewers, editors, and publishers. The study of aquatic toxicity of alcohols81 has already been mentioned, where unjustified incorporation of additional descriptors did result in significant model improvement. Yaffe et al. used fuzzy ARTMAP statistics to obtain a standard error of prediction of 0.08 log unit in their QSPR study of aqueous solubility. However, the experimental error in aqueous solubility measurements is estimated to be 0.6 log unit, and hence, the reported QSAR with a prediction error lower than the error of measurement is most likely a result of over-fitting.
No consideration of distribution of residuals
QSAR model predictions can contain two types of errors - random and systematic. Random error is an indication of the irreproducibility in the response data and/or the descriptor values, and can be reduced by careful selection of these properties. Systematic errors usually result from the bias in measurement or calculation, and can be identified by a simple plot of residuals against measured response values. If systematic error is absent, the residuals should exhibit random distribution around zero line. If the plot shows a marked bias to one side of the zero line, or shows a regular variation of residuals with measured response values, the systematic error is present, and should be eliminated if possible. It would be useful to have residual plots reported or included in QSAR publications.
Inadequate training and/or test set selection
For best results, training set data should be well-distributed over the full range of endpoint values. This is often not possible, for various reasons, but very poor distribution, such as two clusters of chemicals, or one or two chemicals far removed from the others, will exert strong model leverage and must be avoided. Adequate distribution of properties and endpoint values within the test set is also crucial. Test set chemicals must be reasonably similar to some of the training set chemicals, and yet too great similarity can give an overly optimistic indication of a QSAR’s predictive capability.85
Inadequate QSAR model validation
It is now widely accepted that to rigorously assess the predictivity of a QSAR model, some external validation is required, i.e., to use the QSAR model to predict endpoint values for an external test set of chemicals that were not included in model training. Tropsha and Golbraikh recommended that the process of training and test set selection and external validation should be carried out a number of times so as to identify the ranges of external predictivity of a model. Perhaps even more stringent approach to external validation should be based on a 'time-split' selection as advocated in a recent study by Sheridan.
Lack of mechanistic interpretation
It is not always possible to provide a mechanistic interpretation of a QSAR model. Furthermore, it should be borne in mind that the existence of even a very good correlation does not imply causality. Nevertheless, mechanistic interpretations are often helpful, for example in guiding future synthesis of drug candidates. An report of the Organization for Economic Cooperation and Development (OECD)89 recommended that the following questions to be asked about possible mechanistic basis of a QSAR model: (i) Do the descriptors have any physicochemical interpretation that is consistent with a known mechanism? (ii) Can any literature references be cited in support of the purported mechanistic basis of the developed QSAR? If the responses to both questions are positive, one may have some confidence in the proposed mechanism of action.
To summarize, a rich history of QSAR calls for the proper use of well-established statistical practices and 'best practice' rules unifying the standards of data processing and model interpretation, and aiming to avoid the above-described common mistakes and missteps.
2. Current trends in QSAR methodology
2.1 Chemical data curation
One of fundamental assumptions of any QSAR or cheminformatics study is the correctness of input data generated by experimental scientists. As obvious as it seems, the presence of incorrect or imprecise data in modeling sets is increasingly considered a major concern for building computational models, particularly where the activity signal is sparse or potency variation limited, and a QSAR pattern is not easily discerned. A recent study90 demonstrated that on average, there are two chemical annotation errors per each medicinal chemistry publication, with an overall error rate for the compounds indexed in the popular commercial WOMBAT database91 being as high as 8%. In another recent study,78 authors investigated several public and commercial databases and reported error rates in chemical structure annotation ranging from 0.1 to 3.4%. Two main types of errors in input data can be considered: (i) directly related to chemical structures; and (ii) related to associated experimental measurements.
Recent publications78, clearly pointed out the importance of chemical data curation in the context of QSAR modeling. They suggest that having erroneous structures represented by erroneous descriptors could have a detrimental effect on models' performance. Thus, the authors demonstrated that rigorous manual curation of structural data, and elimination of questionable data points, often leads to substantial increase in model predictivity. This conclusion becomes especially important in light of the studies of Olah et al.90 and Tiikkainen and Franke, emphasizing the significant error rate in medicinal chemistry articles and bioactivity databases respectively.
Surprisingly, there are very few published reports describing how the primary data quality influences the performances of QSAR models. Beyond the calls on the importance of data (mostly chemical) curation discussed by Williams and Ekins, only the studies conducted by Young et al.,78 Southan et al., and Fourches et al. described the compendium of issues directly related to chemical data curation. Fourches et al. developed a guideline of best practices for preparing chemical data prior to any other stage of the predictive QSAR modeling workflow (see Figure 2). Organized into a solid functional process, different curation steps (see Figure 2B) allow both the identification and correction of structural errors, sometimes at the expense of removing incomplete or confusing data records. They include the removal of inorganics, organometallics, counterions, and mixtures that most QSAR descriptor generation programs are ill-equipped to handle or that lead to confounding duplicates when simplified (e.g., desalted). Additional curation elements include structural cleaning (e.g., detection of valence violations), ring aromatization, normalization of specific chemotypes, and standardization of tautomeric forms. Post processing entails deletion of duplicates resulting from curation, standardization and normalization, and manual checking of complex cases.
Workflow for predictive QSAR modeling (A) incorporating a critical step of data curation (within the dotted rectangle) that relies on its own special workflow (B).
Removal of mixtures
Treatment of mixtures is not a simple computational issue and various situations are encountered in the workflow: (i) the mixture contains a large organic compound and several smaller moieties, either organic or inorganic (e.g., complexes or non-stoichiometric hydration) – if it is reasonable to assume that the experimentally determined biological activity associated with the record is directly and only caused by the largest molecule (and is not affected by the smaller components), the largest compound can be kept and the smaller moieties should be deleted; (ii) several similar organic compounds with close molecular weights, such as in the case of mixtures of stereo or geometric isomers: usually, the deletion of the entire record is recommended (unless the active ingredient or isomer is known and can be selected manually), because it is not possible to determine which component has to be retained for modeling using simple rules and automated software; manual intervention is usually required and a representative isomer may be chosen; and (iii) a formulation or impure compound where the active or major ingredient is known, but the 'inert' or impurity ingredients are unknown – these cases usually call for a judgment call.
Normalization of specific chemotypes
Very often the same functional group may be represented by different structural patterns in the same dataset. For example, nitro groups have multiple mesomers and, thus, can be represented using two double bonds between nitrogen and oxygens in their neutral forms (i.e., a hypervalent N state), or a single bond linking the nitrogen and the protonated oxygen, or linking both nitrogen and oxygen atoms that are oppositely charged. For QSAR modelers, these situations may lead to inconsistent modeling outcomes depending on how descriptor-generation programs process these cases. Similarly, although ring aromatization and the normalization of carboxyl, nitro, and sulfonyl groups are relatively obvious, more complex cases like anionic heterocycles, quaternary ammonium ions, poly-zwitterions, tautomers, etc., require a deeper analysis and multiple normalization steps.
Removal of duplicates
Rigorous statistical analysis of any dataset assumes that each compound is unique. However, structural duplicates are often present, especially in large datasets. As a result, QSAR models built on such collections may have artificially skewed predictivity. Hence, duplicates must be pre-processed and removed prior to any modeling efforts. Once duplicates are identified, the analysis of their associated properties is mandatory, requiring some manual curation. For a given pair of duplicates, if their experimental properties are identical, one should be straightforwardly erased. However, if their experimental properties are different, there are several scenarios to consider: (i) the property value may be wrong for one compound due to, e.g., a human error when the database was built; (ii) both experimental properties are correct but the previous curation tasks (for example, the removal of counterions in salts) have modified the substance records to create such duplicates; and (iii) there are experimental replicates in the dataset with different reported property values. All three cases will require some additional scrutiny and evaluation to determine the best course of action. In the case where one value is known of suspected of being in error, rejecting the entry is the obvious course; where desalting led to a duplicate, the property associated with the original salt form (as opposed to the unmodified parent) should be deleted; and in the case of experimental replicates, results must be appropriately averaged or aggregated to produce a single result.
Final manual checking
The last step of data curation entails the manual inspection of complex molecular structures. Common errors identified during the manual cleaning procedure may have different origins: (i) the structure is wrong – a rapid check of both the compound’s IUPAC name (if available) and its structure is essential to identify possible errors concerning the scaffold and positions of substituents (e.g., due to manual errors or program bugs78 in the conversion of SMILES into 2D structures); (ii) the normalization of bonds is incomplete – common mistakes are related to the presence of different representations of the same functional groups, i.e., despite the normalization procedure, some very specific cases can still be present and thus, the corresponding chemotypes must be corrected manually; (iii) some duplicates may still be present despite the use of automated software to remove them, for instance, some tautomers can still be found; and (iv) other errors – wrong charges, presence of explicit hydrogens in a hydrogen depleted structure, incorrect bonds, etc.
Some general rules following the curation workflow were also formulated: (i) it is risky to calculate chemical descriptors directly from SMILES, whereas it is preferable to compute descriptors (integral, fragments, etc.) from curated 2D (or 3D if necessary) chemical structures where all chemotypes are strictly normalized; (ii) structural comparison across available databases may facilitate the detection of incorrect structures; (iii) even small differences in functional group representations can lead to significant errors in models; (iv) locating and processing of structural duplicates is one of the mandatory steps in QSAR analysis, although such searches based on chemical name or CAS number only are incomplete and inefficient; and (v) nothing can replace hands-on participation in the process since some errors obvious for chemists are still not obvious for computers.
Given the clear importance of data curation prior to QSAR modeling, we recommend an additional principle for QSAR model validation, stating that 'to ensure the consideration of QSAR models for regulatory purposes, the chemical datasets used to train and validate these models must be thoroughly curated with respect to both chemical structure and associated target property values.'
2.2. QSAR in toxicity prediction
Special challenges
Applications of structure-activity relationships (SAR) to modeling and predicting toxicity endpoints are not fundamentally different from those used in other fields and employ almost every existing SAR approach, ranging from Structural Alerts, to SAR heuristics (expert judgment), to QSAR for congeneric and non-congeneric sets, to combinations of models (consensus).–
Toxicity prediction, however, also poses special challenges. Applications of QSAR modeling to fields such as drug discovery are usually focused on sifting through large numbers of potential drug candidates for compounds that are active at a well characterized enzyme target, where some knowledge of the target interaction and the chemical space of known ligands constrains the search. In the field of toxicology, QSAR methods are typically applied towards the more elusive goal of predicting potential toxicity outcomes for in vitro cell cultures or in vivo animal test systems, where the toxicity endpoint (e.g., mutagenicity, developmental toxicity, cancer) tends to be less well understood, and is likely to encompass multiple mechanisms and pathways to adverse outcome, i.e., where no single interaction mechanism is known or anticipated. Other significant challenges pertain to the chemical knowledge base used for model building (i.e., the training set) and the chemical space to which models will be applied (i.e., the prediction space) and for what aim (mechanism elucidation, screening, prioritization, safety assessment, etc.).
In contrast to the design of drugs and pesticides, where a chemical-activity space of interest is usually populated to serve as a training set for model building, a researcher in toxicity modeling is most often constrained to work with whatever limited data are publicly available, with the goal of predicting whether a chemical is potentially harmful. In particular, within a regulatory or safety assessment workflow, where exposure of humans or ecosystems to each of hundreds to thousands of diverse chemical compounds is a distinct possibility, there is not only greater weight placed on individual QSAR predictions, but regulatory action most often requires a greater body of supporting evidence accompanying each prediction.,
It is generally accepted that QSAR success in modeling toxicity is more likely when one or more of the following conditions are met: (i) compounds within the training set are structurally similar (i.e., congeneric), implying that a single target-mediated mechanism, even if unknown, is more likely; (ii) the toxicity endpoint being modeled is either non-target specific (e.g., narcosis in aquatic toxicity due to membrane concentration effects), or subject to relatively well-understood chemical reactivity principles (e.g., electrophilic theory of carcinogenicity); (iii) the toxicity endpoint is linked to a well-defined molecular target (e.g., estrogen receptor) or phenotype (e.g., cleft palate malformation, or liver tumors in rats); or (iv) toxicity data are available for a sufficiently large number of diverse chemicals to capture all or most of the possible structure-activity associations, representing multiple possible adverse outcome pathways within the same dataset (e.g., genotoxicity). Hence, a defining challenge for QSAR applied to toxicology is that of balancing the highest possible endpoint resolution with the need for sufficient statistical representation, with the latter closely tied to the number of chemical-activity pairs in the training set. To meet this challenge, toxicity endpoints have sometimes been aggregated to what toxicologists may consider biologically meaningless extremes (e.g., lumping all possible developmental malformations and outcomes, such as cleft palate and fetal death, to one endpoint, 'developmental toxicity') to yield the largest possible chemical training set. On the other hand, data for organ or species-specific toxicity phenotypes (e.g., mouse liver tumors, rat cleft palate, etc.) tend to be available for far fewer compounds, resulting in insufficiently robust QSAR models.
In practice, the use of prior knowledge of biological or chemical mechanisms in guiding and constraining a QSAR modeling study, or the use of in vitro test data in conjunction with structural features and properties have proven to be critical for overcoming the perennial challenge of 'not enough data'. Examples include skin sensitization QSAR models built on clear mechanistic and chemical reactivity principles, and a recent proposal of Benigni to strategically combine the use of well-established structural alerts with the results of in vitro mutagenesis and cell transformation assays for the prediction of genotoxic and non-genotoxic chemical carcinogens.
QSAR and computational toxicology
A number of advances and new initiatives in the growing field of 'computational toxicology' have the potential to move QSAR approaches beyond current limitations, as well as to extend models into areas of toxicology previously considered as intractable (e.g., reproductive toxicity). Notable progress in computer technologies, computational chemistry and cheminformatics, as well as increasingly sophisticated statistical and machine-learning approaches, have fueled much of the methodological advancements in this field. Equally, if not more important, however, have been major initiatives on the toxicity data side of the equation, both in the better capture, representation, and utilization of existing toxicity data, and in the generation of new data.– There have been great strides in the chemical structure annotation, curation, and ontological organization of in vitro and in vivo public toxicity datasets to serve as) or in combination with chemical structure descriptors to model selected in vivo endpoints represented within ToxRefDB., These purely statistical efforts applied to a complex toxicity data set have been, for the most part, unsuccessful. The latter is not surprising after taking into account the structural diversity and limited size of this Phase I data set, the noise within the qHTS data, the relatively low hit rate across qHTS assays (generally 10% or less), and the relatively low incidence of in vivo positives for the modeled phenotypic endpoints. However, when rational data constraints (e.g., rejecting incomplete experimental protocols), prior biological knowledge (e.g., literature-reported results implicating particular gene targets), and pathway-based hypotheses were employed, several studies have successfully demonstrated significant in vitro to in vivo correlations, most remarkably for reproductive toxicity., In other studies, Tropsha and co-workers have demonstrated that qHTS data, when used as biological descriptors (without regard to biological relevance to the modeled endpoint), added significant information content beyond what could be achieved with chemical descriptors alone and improved overall model performance for predicting in vivo toxicity.,
What seems clear from these results is that QSAR approaches can potentially benefit from the new information content contained within qHTS data, information that extends beyond purely chemical structure analogy and into the biological realm, but that some prior knowledge, biological (or chemical) mechanism considerations and hypotheses are needed to guide QSAR modeling efforts into productive areas.
In addition, although qHTS in vitro to in vivo models have met with some initial success, they have thus far failed to integrate QSAR approaches that could potentially guide the development and improve models' performance. Greater opportunities will present themselves with the expansion to nearly 1,000 chemicals in ToxCast Phase II and with qHTS screening of the larger Tox21 library, consisting of more than 8,000 diverse structures across 50–100 selected assays being run at the NCGC. These new qHTS data and computational toxicology initiatives represent an area of open possibility and challenge for QSAR, to better integrate with biologically based models and to extend its reach in chemical space and in modeling toxicity at more refined levels of biological organization.
A 'QSAR21' approach using MoA QSAR
Large in vitro and in vivo data sets, such as being generated within Tox21, probe diverse biological pathways to reveal assay-endpoint signatures. However, when focusing exclusively on the biological aspects, computational toxicology modelers are in danger of making the same mistakes that chemists made in the early days of QSAR: focusing too narrowly on the chemical side, while reducing complex biological phenomena into overly-simplistic numeric values. In the early phases of Tox21 and ToxCast data analysis, due in part to the small size of the chemical landscape considered, biological models mostly focused on linking biology to biology in relating in vitro to in vivo outcomes, with some limited success. Whilst this approach brings a way to formulate mechanistic pathways for a large, diverse inventory of chemicals, it fails to utilize the value of the underlying chemical information in helping to discern mechanistically driven, meaningful patterns in the data. It also confines models to the experimental data realm only, where HTS data are required inputs. In the end, it is only by creating linkages across the full progression of 'genotype ↔ chemotype ↔ phenotype' that mechanistic approaches will produce actionable knowledge in modeling chemically-induced toxicity on the basis of chemical structure inputs.
The MoA QSAR approach has been recently applied to build on the mode-of-action (MoA) concept of classifying chemicals to establish a collection of biologically similar compounds for a given phenotype training set based on in vivo toxicity data. In choosing predictors, the MoA QSAR also employs biological assay results as descriptors in addition to chemical structure-based features and properties. The biological descriptors can include qHTS results where assay targets (genotypes) proposed to be relevant to toxicity mechanisms are grouped by co-occurrence in known or hypothesized molecular pathways. The presumption is that the new qHTS assay data are effectively populating the vast data space of toxicity pathways, and as this landscape becomes more mature, it becomes possible to infer more robust and biologically based connections from chemical structure to toxicity endpoints. With the chemicalactivity landscape bounded by these mechanistic principles, chemical elements may be more easily discernible and a modeler has greater freedom to employ unbiased statistical approaches to reveal chemical features and determinants of activity. A molecular initiating event can trigger numerous cellular responses, with key events leading up to the organ responses. The pathway information, a representation that incorporates chemical structure, physicochemical properties, and biological information all together. A chemotype thus serves to link a chemical structure to a toxicity pathway. A chemotype, at minimum, is a structural alert for a given toxicity endpoint, but augmented with chemical reactivity within an MoA context. Thus the chemotype inherits biological information and can be used to group chemical structures based on biological and chemical similarities. The chemotypes carrying the MoA information guide the process of constructing training sets by providing mechanistic interpretations. MoA QSAR uses this link between biological event and chemical group to identify more mechanistically biased training sets that ultimately relate to phenotypic effects.
Results from models are then combined to obtain one prediction outcome by employing quantitative decision methods, including naive Baysean or Dempster-Shaffer Theory approaches. The prediction outcome obtained by such combination approaches is designed to give more robust and improved predictivity while maintaining model interpretability. The MoA QSAR approach combined with a decision theory based on a Bayesian treatment has been successfully applied to modeling of bacterial mutagenicity, clastogenicity, tumorigenicity, developmental fetal toxicity and specific malformation endpoints, in vivo skin irritation, and skin sensitization in safety assessment in a regulatory setting.,124
Within the regulatory workflow for assessing potential chemical hazards, an important requirement of information is that it can support the decisions that a regulator needs to make by a clear rationale within a reasonable time frame. Transparency does not just apply to the ability to access and scrutinize underlying information sources and model details, but also to clear communication of the basis for the rationale in both biological and chemical terms. The MoA QSAR/Decision Theory approach meets the needs of toxicologists and regulators due to its transparent reasoning anchored in relatively simple conveyances of molecular structure linked to biological mechanisms. The 'QSAR21' paradigm enabled by the MoA QSAR/Decision Theory approach, offers a means to bridge chemistry and biology along a mechanistic framework, promising more accurate and usable models for regulatory applications. To fully capitalize on these advances, however, will require QSAR practitioners to gain more intimate knowledge of, and engagement with the biological data, both at the in vitro and in vivo level.
Present and future role of QSAR in toxicology
Despite the great promise of computational toxicology approaches, there continue to be areas of chemistry and chemical risk assessment in which relevant test compounds are either unavailable (such as in early phases of chemical design or pre-manufacture review), or qHTS test results are unattainable with current technologies (e.g., volatiles, reactives, insolubles, metabolites). Such problematic areas have been and will continue to be heavily reliant upon QSAR. With heightened pressures to regulate new and existing commercial and environmental chemicals, and decreasing resources for testing, QSAR methods are being increasingly used in screening, testing prioritization, pollution prevention initiatives, green chemistry, hazard identification, and risk assessment. To be fully accepted by end-users (toxicologists, regulators, industry), however, these QSARs must meet a range of needs, including relevance to regulatory schemes, transparency, biological plausibility, and understandability by non-developers.
2.3. QSAR prediction of metabolism
Historical prospective
The physician Kahn illustrated in a popular scientific book series the view of a human as a powerful machine using metaphors from industrial society.126 Kahn showed that during a typical 70 year life span, a human consumes up to 1,400-fold mass of bodyweight in foodstuffs, with both nutritional processing and clearance of toxic substances governed largely by endogenous metabolism. Hydrophilic substances undergo limited biotransformation and can be excreted unchanged. Lipophilic compounds are extensively metabolized but poorly excreted. In the course of evolution, enzymes developed that preferentially act on lipophilic xenobiotics and transform them to more hydrophilic, easily excretable metabolites. Unfortunately, very lipophilic compounds such as insecticides and other persistent organic pollutants (e.g., DDT, chlordane, polychlorobiphenyls, etc.) are less easily metabolized and eliminated, thus leading to bioaccumulation.
Driving forces for the progress in metabolism research during the past five decades are largely due to the tremendous progress in analytical instrumentation and the increasing awareness of the impact of metabolism on unwanted drug effects. Pharmacokinetic consequences may be observed because of the following factors: (i) a drug might induce one or multiple enzymes in metabolism, resulting in a time-dependent therapeutic response over days or weeks; (ii) a drug or metabolite can inhibit a metabolic pathway, resulting in complex kinetics; (iii) the physicochemical properties of the drug metabolites might differ significantly from the parent drug, e.g., higher polarity may result in faster urinary excretion, whereas high lipophilicity may lead to retention in tissue and bioaccumulation. A major issue in pharmacotherapy is that of severe adverse drug reactions (ADRs) and whether they are predictable, avoidable, and iatrogenic. Frequently, ADRs are related to metabolism and, therefore, special focus is placed on drug metabolism during the drug discovery and development process, as well as on pharmacovigilance. Drug-drug interactions due to metabolic inhibition or competition for storage binding sites may result in pharmacological potentiation, whereas metabolic induction of drugdrug interactions may result in a decreased clinical response. Polymorphism of some drug-metabolizing enzymes may be responsible for a low metabolic capacity. Hence, phenotyping or genotyping of patients is increasingly considered an appropriate means to improve the patient’s safety in pharmacotherapy.
In a recent publication, Testa et al. reviewed the reactions and enzymes involved in the metabolism of drugs. Their analysis of the metabolic reactions of over 1,000 different substrates in three selected journals during the years 2004–2009 underlines the importance of cytochrome P450-catalyzed oxidations and UDP-glucuronosyl-catalyzed glucuronidations in drug metabolism. Nevertheless, the study demonstrates the role of other oxidoreductases, esterases, and transferases that significantly contribute to all drug metabolism reactions. Whereas almost 58% of metabolites in the first generation are produced by CYPs, their relative frequency decreases to 32% in the second generation, and 21% in the third generation of metabolites.
Metabolism prediction models and resources
The prediction of metabolites has to address at least three different kinds of selectivity questions. As the metabolic reaction can be catalyzed by different enzymes, the corresponding metabolism prediction models have to address enzyme selectivity first and foremost. Thus, in the case of cytochrome P450 the affinity toward different isoforms has to be modeled. Furthermore, CYPs mediate different reaction types and, therefore, the prediction of chemoselectivity is also mandatory. Finally, a particular reaction type might be applicable at multiple sites of a substrate. Therefore, the prediction of the regioselectivity of a reaction type is also required.
The prediction of the metabolism of a chemical on theoretical basis from first principles is not (yet) possible. The influence of solvation and flexibility of protein side chains are very complex phenomena to be directly calculated. Furthermore, individuals of a particular species might have differences in their metabolism due to enzyme polymorphism. Hence, a comprehensive knowledge base is required that can be used for inductive approximations.
A few metabolic reactions databases are currently available. The most widely used is the Metabolite database128 that contains more than 70,000 reactions for different species (humans, rats, etc.) and is particularly reflective of the metabolism of drugs and drug-like compounds. Complementary to the Metabolite database (which originally was distributed by MDL), Accelrys offers a database named Metabolism.129Metabolism includes data from the primary literature and its initial scope was focused on metabolic pathways for agrochemicals. The third commonly used database of metabolic reactions is called ADME DB.130 ADME DB is available as an online service only; its content is mainly focused on metabolism associated with cytochrome P450, with the corresponding data extracted from the primary literature.,
There are numerous studies focused on modeling cytochrome P450 substrates and inhibitors using QSAR methods, pharmacophore-based approaches, docking and molecular dynamics simulations (see a recent review by Kirchmair). Chohan et al. reviewed 61 QSAR studies that have been used to elucidate the molecular features that influence the binding and metabolism of a compound by the major phase 1 and phase 2 metabolizing enzymes; Cytochrome P450 (CYP) and UDP-glucuronosyltransferase (UGT), respectively. Braga and Andrade discussed a perspective of the utility of QSAR and QM/MM approaches on drug metabolism prediction including their present limitations and future perspectives in medicinal chemistry. Several software packages are also available for predicting metabolites, including the commercial products, METEOR, META, MetaSite–, and TIMES, as well as the open-source Metaprint2D, and SMARTCyp packages, among others.
Current challenges
Coverage and comprehensiveness of metabolic data represents a critical issue with respect to QSAR modeling; several reasons could be outlined:
Quantitative data on binding constants, kinetic parameters of metabolic reactions and product distribution metrics have to be measured under comparable experimental conditions to be suitable to derive Quantitative Structure-Metabolism Relationship models.
Experimental elucidation of chemical metabolites is a time- and cost-intensive process.
Experimental systems used to investigate metabolism in vitro are not necessarily comparable because the cells originate from different species.
Polymorphism of critical metabolic enzymes (such as cytochrome P450 2C19, etc.) might not be considered in older publications.
Clinical data (if available) must properly take into account differences in sex, age, diseases, medication, etc., of the study subjects.
As a consequence, the available metabolic datasets are typically heterogeneous. Thus, standardization of in vitro metabolism data would enable a paradigm shift from approaches that do not lend themselves as easily to interpretability. However, the importance of interpretability to practical acceptance of QSAR solutions is well-established and is reflected in one of the OECD principles: 'To facilitate the consideration of a QSAR model for regulatory purposes, it should be associated with … a mechanistic interpretation, if possible'.147 For purposes of this discussion, the term 'interpretability' refers to the ability of a user to understand and rationalize both the underlying model determinants of activity, i.e., what the model deems to be the primary predictors of activity, as well as the individual model predictions, in terms of chemical structure, reaction mechanisms, or known or plausible biological mechanisms of activity.
A predictive and interpretable model has a clear advantage over an equally predictive but non-interpretable one. The former allows for the targeted design of compounds with desired properties, lends itself to mechanistic interpretation and hypothesis generation, and can contribute to further understanding of mechanism(s) of action. When predictions from such a QSAR model can be independently supported, and are chemically and biologically plausible, they also can carry significantly greater weight in, for instance, a safety assessment workflow. The importance and benefits of building interpretable models have been demonstrated in several studies,– and some examples will be characterized herein. Below, we outline different constituent blocks of a general QSAR model workflow with respect to model interpretability.
Role of molecular descriptors
QSAR model interpretability strongly depends on the nature of chemical descriptors used. The use of well understood physical-chemical characteristics of a molecule as descriptors (van der Waals volumes and surfaces, lipophilicity, H-bond related parameters, etc.) can aid in model interpretation. Metrics related to the electronic configuration of a compound (partial charges, dipole moment, orbital energies, etc.) are also suitable for structural interpretation when placed in the context of reactivity hypotheses. In addition, various molecular fingerprints,156 and fragment descriptors such as MNA, G-QSAR, 157 ISIDA, 158 and Simplex (SiRMS) have a direct connection to molecular structure and, thus, have the potential to lend interpretability to a model.
QSAR models built with topological indexes reviewed above are often more difficult to relate to easily understood chemical concepts. Autocorrelation descriptors pose a similar problem in that they encode an indirect relationship between molecular structure and descriptor values, such as BCUT, WHIM, GETAWAY, RDF, etc., although with some post processing, they can be used to convey insight into the role of branching, degree of structural nonhomogeneity and cyclicity of a molecule on activity.161
Modeling techniques
Some QSAR modeling techniques, such as linear regression and decision tree approaches are very straightforward to interpret given their intuitively understandable architecture; when combined with interpretable QSAR descriptors, these approaches allow for the development of ready-to-use structural rules and alerts. Other types of models, however, require additional operations to provide chemical insights.29 Generally, all approaches enabling interpretation of QSAR solutions can be divided into model-specific and model-independent.
Model-specific interpretation methods
The estimation of weights of individual descriptors in a linear model is routinely made by considering the corresponding regression coefficients., More sophisticated methods are also applied, e.g., to interpreting PLS models, in this case relying on the analysis of a descriptor’s contribution to the variation of the investigated property.29,165 Similarly, linear support vector machine (SVM) models are amenable to this approach for providing interpretability of results.
Some efforts have been made to interpret artificial neural network (ANN) models using their weights and biases. Along a similar line, Kuz'min et al. proposed an approach for interpretation of Random Forest models using the differences between mean activity values of compounds in the corresponding parent and child tree nodes.
Model-independent interpretation approaches
Similar to the concept of assessing leverage, the relative importance (or significance) of a given descriptor, even in a complex model, can be assessed by comparing the reduction in the overall predictive power of a model when one descriptor versus another is removed or altered. Originally suggested by Breiman169 for interpreting Random Forest (RF) models, this approach was also adopted by Guha et al. to evaluate neural networks-based solutions. Another approach analyzes local gradients or partial derivatives of descriptors as reflecting their contributions to the variation of the modeled property. These methods have been used to aid interpretation of PLS, RF, ANN, and SVM models, among others.–
A universal approach to model interpretation
To avoid many problems outlined in the previous sections, Polishchuk et al. proposed an approach to the interpretation of QSAR models that does not depend on the nature of the descriptors, utilized mathematical approach, and a type of endpoints (continuous or binary). The gist of this methodology is very simple (Figure 3): activities of a target molecule (PQSAR(AB) in Figure 3a or PQSAR(ABC) in Figure 3b) and its virtual analog (PQSAR(A) in Figure 2a or PQSAR(A…C), derived from this initial molecule by eliminating a molecular fragment with a pre-defined structure, are estimated using the QSAR model. The difference in the calculated activities of a target molecule and its virtual analog is considered as an influence (contribution) of eliminated fragment (P'(B) in Figure 3). This simple approach is equally applicable for estimating the contributions of both terminal and central parts of a molecule. Notably, the contributions to activity are more sensitive to chemical descriptors, which are used to represent the molecule with and without the fragment, than to the modeling technique.
Estimation of terminal (a) and central (b) fragment's contributions to activity. PQSAR – contribution estimated by developed QSAR model; P' – contribution estimated by the universal interpretation approach.
Concluding this section, we shall stress that, using the universal approach, any QSAR model, despite of the complexity of the modeling technique or nature of the descriptors used, can be formally interpreted in terms of significant chemical features that can be easily understood by medicinal chemists. Hence, all QSAR models should be primarily evaluated on the basis of well-established methods for assessing external predictive ability. Stated otherwise, a QSAR model with easily interpretable descriptors and poorer performance statistics should not be preferred over a more predictive model that is less easily interpreted.
2.5. Multi-task modeling
Often, chemicals have multiple biological activities (cf. polypharmacology) that may be interrelated (not to mention multiple physical properties that are frequently the subject of prediction by QSAR approaches). Typically, however, QSAR models are developed for each target property individually, without utilizing knowledge that can be extracted from QSAR models for other activities of the same compounds. Individual QSAR models of this sort should not be viewed as separate entities, but rather as nodes in a network of interrelated models. This concept is accounted for in an inductive knowledge transfer approach realized in multi-task learning (MTL) and Feature Net (FN) methods. MTL176 treats several tasks in parallel and uses a shared representation of data. This can be carried out using machine learning methods yielding models with several outputs, such as neural networks, PLS, or SVM with special kernels.177 FN treats different tasks sequentially when predictions made by previously developed models are used as extra descriptors for the main task.
An inductive knowledge transfer approach could significantly improve predictive performance of individual QSAR models (or, Single Task Learning (STL) models) built on small and structurally diverse datasets. Since the cost of obtaining new data is rather high, especially for in vivo experiments, the integration of available experimental data on related activities could serve as an alternative to costly and time-consuming experiments. The FN technique is widely used in QSAR studies whenever a successfully employed descriptor is reused. Thus, models employing parameters used by other QSAR models solutions (e.g., logP, pKa, Ehomo, Elumo) as descriptors could formally be considered as FN. More generally, FN could be realized in the form of a multilayer network of models, in which the outputs of the models of previous layers are the inputs for the models of the next layers.178
Unlike FN, only a few applications of MTL in QSAR studies have been reported in the literature. For instance, higher performance of MTL and FN approaches over the conventional STL models was demonstrated by Varnek et al. in QSAR modeling of tissue-air partition coefficients (logK) using backpropagation neural network with several output neurons. The initial datasets encompassed 11 different types of logK data points for humans (H) and rats (R). Only four datasets were of reasonable size (about 100 compounds), whereas the others included from 27 to 38 compounds. The output layer of the network contained either one (for STL and FN) or 11 (for MTL) neurons. In STL and MTL studies, only fragment descriptors were used as inputs, whereas in all FN calculations, all input neurons were fed by 10 predicted properties. The results of that work demonstrated that conventional STL allowed reasonable prediction of only 4 activities for which larger (about 100 compounds) data sets were available, whereas MTL and FN calculations generated models with acceptable quality for 9 activities.
Of note, MTL should not be confused with multi-target learning, in which the performance of models predicting binding of ligands to different biological targets is boosted by incorporating certain information concerning the proteins’ structure., Such information can be represented either directly as a set of target-specific descriptors or indirectly through special kernels for protein targets. In the latter case, descriptors for ligands and targets are combined in the feature spaces induced by kernels. In some publications the multi-target learning is erroneously associated with MTL,–181 although in reality this is an STL model for complex chemical objects, i.e., protein-ligand pairs. In our view, the term MTL should designate only the tasks based on common internal representation, which cannot be reduced to STL.
The growth of available experimental data and predictive structure-activity models will stimulate further development of the inductive learning transfer approach. We believe that in the future, isolated and unrelated QSAR models will be connected to the network of interrelated models solutions organized in the form of a 'chemical brain' accommodating considerable volumes of experimental data and knowledge, which will significantly improve the quality of prediction of various chemical and biological properties. This expected development will further advance the integration of QSAR as part of Systems Chemical Biology.
2.6. Experimental validation of QSAR models
The discovery of novel bioactive chemical entities is the primary goal of computational drug discovery, and the development of validated and predictive QSAR models is critical to achieve this goal. Moreover, experimental validation is the only indicator of actual utility of QSAR modeling. The general scheme of QSAR-guided research project inclusive of experimental validation is shown in Figure 4. Virtual screening (VS) approaches are the inherent part of this workflow. They are used to identify chemical hits predicted to be active against selected target(s) from large chemical libraries (see Figure 5). With the continuous emergence of novel biological targets of therapeutic potential from drug discovery teams in industry and academia, effective and accurate VS technologies continue to be in high demand. Meanwhile, this demand, especially in terms of both computational efficiency and accuracy, is amplified by the very rapid growth of chemical compound collections that are available for virtual screening even in the public domain. For instance, both PubChem and ChemSpider, the two major collections of chemical structures on the web, currently include over 30 million compounds each, and ZINC, a database frequently used for virtual screening applications, incorporates a total of approximately 21 million compounds. Fara et al.187 discussed the integration of virtual (QSAR-based) and actual (physical) screening. However, as shown in Figure 5, modern VS workflow incorporates several critical filtering steps to eliminate compounds that are unlikely to be active: (1) sets of empirical rules (e.g., Lipinski’s, QED drug-likeness), (2) chemical similarity cutoff commonly computed using molecular fingerprints, (3) QSAR-based filter(s) (e.g., only retain compounds predicted to be active or having predicted pKi ≥ 8), and (4) chemical feasibility and/or purchasability.
Overall study design of a QSAR-guided drug discovery project.
General workflow for screening chemical libraries using empirical and QSAR-based filtering.
Several important studies have been published recently in which QSAR-based predictions have been experimentally confirmed. These studies illustrate how useful and reliable computer-assisted approaches can be in assisting medicinal chemists to design novel compounds with controlled bioprofiles.
Keiser et al. developed the Similarity Ensemble Approach (SEA) to compare targets utilizing the overall similarity between their known ligands. The authors applied this approach to prognosticate unknown drug-target interactions for demonstrating the potential use of computational approaches to study and predict drugs’ polypharmacology. More than 3,600 FDA approved and investigational drugs were analyzed by SEA and thousands of unknown associations were discovered. Out of 30 experimentally tested associations, 23 precedently unknown drug-target interactions were confirmed (five of them characterized by a potency less than 100 nM).
Lounkine et al. conducted a large-scale prediction and testing of drug potency on different side-effect targets. The authors extracted and curated experimental data for 285,000 ligands and 1,500 biological targets from the ChEMBL database. Then, they used the SEA similarity search approach to predict the activity of 656 marketed drugs on 73 unintended 'side-effect' targets. Out of 1,644 significant drug-target associations predicted by SEA, there were 893 that were unknown and never reported before. The authors then conducted experimental tests for confirming these predictions. Out of 694 experimental tests, 478 drug-target associations (68.9%) were disproved and 65 found to be ambiguous. However, significant potencies (less than 30 µM) were confirmed for the 151 remaining drug-target associations, especially for 48 compounds with sub-micromolar activities. Interestingly, the authors linked those targets with known side effects and successfully established previously unknown links between drugs and several side effects. This study demonstrated that QSAR-like predictions can successfully prognosticate ligand-target interactions, which can then be confirmed experimentally. One can argue that the use of additional filters such as applicability domain and prediction confidence scores could potentially have avoided the large number of unconfirmed ligand-target associations.
Besnard et al. utilized cheminformatics methods to explore and design compounds with unique polypharmacology (see Figure 6). As both clinical efficacy and overall safety of a drug is determined by its activity profile towards many biological targets, there is a huge need for approaches capable of predicting and designing drugs with a specific multi-target behavior. The authors developed new methods that (i) generates one (or several) generation(s) of chemical analogues of a given parent drug with known properties, and (ii) predicts their polypharmacology using an ensemble of ligand-based QSAR models. Then, most interesting compounds with the preferred polypharmacology profiles are synthesized and confirmed experimentally. The authors explored the case study of an approved acetylcholinesterase inhibitor drug and its in silico generated analogues, all tested for their specific or promiscuous polypharmacology towards G-protein-coupled receptors. More than 800 ligand-target predictions of prospectively designed ligands were tested experimentally and 75% were confirmed.
Adaptive drug design for computer-aided generation of compounds with controlled polypharmacology (see earlier work).
Other examples of experimentally validated QSAR-based predictions have been published by the Tropsha group at the UNC. Recently, Hajjo et al. have developed and validated binary classification QSAR models capable of predicting potential serotonin 5-HT2B actives that are known to cause valvular heart disease. The models were employed to screen the World Drug Index library and 122 compounds were prognosticated to be 5-HT2B actives. Ten of them were tested experimentally and nine were confirmed to be active. These QSAR models can thus be employed for predicting 5HT2B-related valvulopathy. A similar strategy has been followed to design novel antimalarial compounds by modeling a dataset of 3,133 compounds defined as either active or inactive towards P. falciparum. The virtual screening of the ChemBridge library using the QSAR models led to the identification of 176 putative antimalarial compounds that were submitted for experimental validation along with 42 putative inactives as negative controls. The authors reported that 25 computational hits (14.2%) were confirmed to have antimalarial activities, whereas all 41 putative inactives were also confirmed as inactives. Importantly, confirmed hits featured novel chemical scaffolds that could be promising for developing novel antimalarial agents. In another case, the Tropsha group generated QSAR models for 5-HT6 receptors and utilized them for identifying novel actives in combination with predictions from the Connectivity Map (http://www.broad.mit.edu/cmap/). Thirteen common hits were tested experimentally and ten were confirmed as actives.
QSAR studies–, of various antiviral activities represent another good example of targeted design of new compounds with desired properties using cheminformatics tools. Fifteen novel antiviral agents against influenza H3N2, herpes HSV-1, rhinovirus HRV-2, and coxsackievirus B3 were computationally designed as a result of interpretation of QSAR models. Then, their high activity and selectivity was confirmed by subsequent synthesis and experimental testing. These studies were summarized in review.
In summary, a growing number of published QSAR studies include the experimental validation of predicted hits and this critical step should become a standard component of any QSAR investigation. Scientific journals should raise the bar for accepting and publishing papers that employ QSAR techniques. Importantly, Journal of Medicinal Chemistry already supports this trend by not accepting for publication any papers describing experimentally untested QSAR models and predictions.
3. Novel applications of QSAR and future trends
3.1. QSAR modeling of peptides
Antimicrobial peptides (AMPs) represent a diverse class of natural peptides that form part of the innate immune system of mammals, insects, amphibians, and plants.– In the face of increasing antibiotic resistance by pathogens, AMPs have drawn significant attention as a prospective class of antimicrobial therapeutics– as they hold several notable advantages, including broad range of activity, low toxicity, and minimal development of resistance in target organisms., More than fourteen peptides are currently in development or clinical trials, with two having demonstrated efficacy in Phase III clinical trials.
Despite the fact that a broad spectrum of antimicrobial peptides have been reported and discussed, their structure-activity relationships are not well understood, largely because of substantial diversity in AMP structures and their non-specific mechanism of action. The general view on characteristic features of the AMPs is typically focused on their cationic character, relatively high hydrophobicity/amphipathicity and short length., Thus, the previous attempts of 'in silico' modeling of peptide-based antibiotics were largely based on sequence-based approximation.
Sequence-based AMP modeling
The majority of the previous sequence-based modeling efforts was of qualitative nature and relied on available AMP sequences to (a) discover previously unknown natural peptides, and (b) to modify sequences of known AMPs to improve their therapeutic properties.
The AMP optimization methods relying on various sequence templates were recently reviewed by Fjell et al. In short, such approaches imply systematic change of one or a few amino acids in the sequences of prominent AMPs such as cecropin, magainin, protegrin, lactoferricin, and bactenicin (and, recently, indolicidin and brevinin) to enhance their antimicrobial activity or reduce toxicity. Rather commonly, template-based studies attempted to introduce unnatural amino acids into AMPs to increase stability.,
The sequence-based efforts did not result in drastic improvement of AMP properties., However, such template-based studies (when one changes a part of the molecule, while keeping the rest intact, and records the overall response of the system) has brought the spirit of early-day QSAR into the field of AMP research.
Residue-based QSAR modeling of AMPs
The previous QSAR work on antimicrobial peptides was mainly focused on derivatives of three natural substances: bactenecin, protegrin and lactoferricin, and utilized a residue-based level of modeling. Thus, back in 1987 a set of based on the physical-chemical properties of the peptide z-descriptors, was proposed and used to investigate peptide variants of lactoferricin. Several more recent studies have examined activities of lactoferricin derivatives against bacteria– and herpes simplex virus in response to specific amino acids changes. Thus, Strom et al. modeled a set of 20 peptides with such descriptors as alpha-helicity, HPLC retention time, net charge, molecular surface, and symmetry of charge and hydrophobicity distribution. Later, using an expanded set of peptides, good predictive accuracy was achieved using z-values on a larger set of peptide analogues where only a few amino acid substitutions were made., However, predictions were much less accurate for the cases when more than one or two substitutions were made. In a recent study by Sánchez-Gómez, residue-based QSAR descriptors were used to model membrane permeability of 8-to-12 amino acid long lactoferrin derivatives.
It should be noted that earlier QSAR attempts at AMP modeling suffered from similar shortcomings. Thus, modest numbers and sequence variation in AMP training sets did not typically allow large sets of QSAR descriptors to be employed (with a notable exception of an early work by Mee et al.) and, hence, limited the use of more rigorous, nonlinear modeling techniques. Combined with traditional residue-based level of consideration of the AMPs, these limitations typically did not enable the development of improved therapeutics that outperformed peptide variants in the training set. This situation has recently changed with the advances in high-throughput technologies of peptide synthesis, screening and analysis.,–
Atom-based QSAR modeling of AMPs with the Use of Machine-Learning Approaches
In an important original study, Cherkasov showed that AMP activity can be effectively quantified using machine learning QSAR and atomic levels of structural considerations. When large-scale training data were generated for the AMPs, this assumption led to a series of studies that resulted in the development of synthetic peptides with significantly improved antibacterial activity and lowered toxicity.,– This was achieved by employing artificial neural networks to build computational models of peptide activity based on the data from >1400 sequences, only biased in the content of certain amino acid types believed to be important for antibacterial activity. As a basis for enumeration of peptide structures, a set of 44 descriptors were employed, including 3D QSAR parameters that utilize atomic-scale molecular information, the so-named 'inductive' QSAR descriptors.235, Briefly, the 'inductive' descriptors describe whole molecules based on the calculated effects of the atomic constituents of a molecule. Previously, these parameters allowed quantification of diverse sets of organic and organo-element molecules and free radicals.235–
A total of 26 'inductive' descriptors of electronegativity, hardness, charge, substituent, and steric effects were used. An additional eighteen conventional molecular descriptors were also added, including numbers of hydrogen acceptors and donors, surface area, total and partial charges, and molecular weight. The developed QSAR models proved remarkably accurate in prediction of antimicrobial activity of nine amino-acid long training set AMPs, and allowed the creation of novel and improved peptide variants. In particular, 100,000 virtual peptides were created and QSAR models (pre-trained on experimental results for >1,400 AMPs) were used to predict their hypothetical antimicrobial effects. Based on the QSAR predictions for 100,000 sequence variants, 25 peptides were identified from each of the predicted activity quartiles (roughly corresponding to high-, medium-, low-, and inactive predictions by the QSAR). The selected 100 peptide candidates were then synthesized and assayed against 12 of the most dangerous 'superbugs' including multidrug resistant strains of Pseudomonas aeruginosa, Pseudomonas maltophilia, Staphylococcus aureus, Escherichia coli, and Enterobacter cloacae among others.
Remarkably, of 50 peptides that were predicted as most active, 49 peptides (or 98%) were actually found to be more active than the control antimicrobial peptide Bac2a, while the 2nd, 3rd and 4th quartiles were respectively 88%, 4% and 0% similar to or better than the control peptide Bac2a. Moreover, the best predicted peptides, when tested experimentally, not only demonstrated sub-micromolar in vitro activity against major, life-threatening human pathogens, but also showed significant activity in animal models.
These results have unambiguously demonstrated that QSAR methodology is applicable to AMP data, and that the atomic level of consideration of AMPs combined with machine learning techniques results in practical models that deliver the most active peptides identified to date. Not surprisingly, these findings initiated a broader interest of the QSAR community in the field of antimicrobial peptides. A number of recent studies have successfully utilized QSAR for analysis and discovery of AMPs using both atom- and sequence-based descriptors., Such an approach is allowing rapid advancement of the field of AMPs and has resulted in the development of peptide candidates with improved therapeutic properties.
3.2. QSAR modeling of chemical mixtures
It is a common knowledge that chemical mixtures have a very broad application in all fields of experimental science, as well as broad use in commercial, industrial, and pharmaceutical products. It is also feasible to forecast that the use of mixed formulations, reagents, and industrial releases into the environment will substantially increase in the future, especially in the medical field. At the same time, although modern QSAR is successful in dealing with individual compounds, there are no mature QSAR methodologies that could be directly used to model properties of mixtures, reflecting the lack of robust, well-benchmarked data pertaining to such properties. To date, only a few published QSAR studies of mixtures could be considered reliable., An interested reader can find detailed descriptions of studies devoted to mixtures and QSAR modeling of their properties elsewhere.,245 Herein, we will highlight those studies, that exemplify lack of awareness of some researchers of the best practices of QSAR modeling, which should apply equally to modeling of both individual compounds and chemical mixtures.
Current challenges in QSAR modeling of chemical mixtures
The lack of reliable data poses one of the biggest challenges for the development of QSARs for chemical mixtures. Some information can be found in the ChEMBL database,248 which contains fragmental data on 356 organic mixtures. The NCI database249 stores endpoints of anticancer activity for some mixtures, as does the DTP AIDS Antiviral Screen database250 for anti-HIV activity.
Some limited and sparse data can also be found in the literature. For instance, toxicity of about 100 mixtures toward Photobacterium phosphoreum species can be extracted from papers by Lin et al., Three datasets of reasonable size (271–411 binary mixtures) related to various properties of liquids were published by Ajmani et al.,253, Vapor-liquid equilibrium data for 101 pure compounds and 261 binary mixtures were compiled by Oprisiu et al. using the Korean Thermophysical Properties Databank as a source.255 550 mixtures composed of 33 individual compounds were investigated by Small et al. to discover strong anti-inflammatory combinations. It is expected, however, that the ongoing rapid growth of publicly available databases will begin to address this problem, and QSAR modeling of mixtures will advance, as it did for individual compounds some years ago.257
Current approaches to QSAR modeling of mixtures
QSAR modeling of organic mixtures requires the use of appropriate descriptors. All studies published to date on the subject can be divided into several groups depending on the descriptor type used: (i) descriptors based on the mixture partition coefficient or biological descriptors;(ii) additive molecular descriptors (weighted sum of descriptors of individual components), e.g., see work;(iii) integral non-additive descriptors of mixtures (mixture components are taken into account in a different manner from the additive scheme), e.g., see study; and (iv) fragment non-additive descriptors (structural parts of different mixture components simultaneously taken into account in the same descriptor).262
We consider non-additive fragment descriptors to be the most promising for QSAR modeling of mixtures. For instance, the SiRMS approach is suitable for QSAR analysis of binary mixtures of any composition. The method represents a mixture by two molecules considered simultaneously, where bounded simplexes (tetratomic fragments) describe single components of the mixture individually, while unbounded simplexes describe the constituent parts of the mixture as a whole (see Figure 7). In this approximation, it is necessary to indicate whether the parts of unbounded simplexes belong to the same or different molecules. In the latter case, such unbounded simplexes will not reflect the structure of a single molecule, but will characterize a pair of different molecules.
Generation of simplex descriptors for mixtures.
A special mark is used during descriptor generation to distinguish such 'mixture' simplexes from ordinary ones. The mixture composition is taken into account, i.e., descriptors of constituent parts (e.g., compounds A and B) are weighted according to their molar fraction and mixture descriptors are multiplied on molar fraction of deficient component (see Figure 7). If in the same task both mixtures and pure compound have been considered, pure compound is regarded as a mixture with composition A1B0. In this case, only descriptors of pure compound A will be generated with the weight equal to 1. Thus, the structure of every mixture is characterized by both descriptors of the mixture as a whole and descriptors of its individual constituents.
Analysis of the existing descriptors of mixtures demonstrates that additive descriptors of mixtures, where the latter are characterized by mole-weighted average descriptors of the constituents, have the following disadvantages: (i) they wholly rely on the expectation that conventional descriptors can be significant in the explanation of a property of mixture; and (ii) the consideration of (inter)action effects is impossible and, thus, only simple tasks with additive or very close to additive effects can be investigated. Advantages of the additive approach are: (i) the process of descriptor generation is simple and intuitively understandable; and (ii) this approach is not property-oriented, i.e., additive descriptors could be applied (bearing in mind the drawbacks of this methodology) to any investigated activity or property, and, sometimes, this method has shown good results, as in the study of Ajmani et al.
More complicated mixture descriptors developed by Ajmani et al.,253 are capable of encoding important non-covalent intermolecular interactions, which is their significant advantage. However, they are system- and property-specific, i.e., applicable only to particular systems (such as binary mixtures where the components are dissolved in each other) that were described by Ajmani et al.,253 Another serious disadvantage of this approach is that two mixtures of different composition could be described by identical descriptors, thus the descriptions are not sufficiently unique.
Simplex and ISIDA mixture descriptors are free of the aforementioned drawbacks; they can be applied to any property of interest, and they are capable of capturing interaction or joint effect of components. However, in the current version, these approaches can be applied only to binary mixtures. From a methodological point of view, these two methods appear better than others, but major improvements are still required.
QSAR modeling of mixtures
In our view, no single QSAR study or methodology published to date can yet be recommended as a reliable tool or a 'golden standard' for QSAR analysis of mixtures. In addition to the mentioned limitations in datasets sizes and drawbacks of mixture descriptors, many published reports contain significant methodological errors.
As in traditional QSAR, rigorous external validation is required for modeling mixtures. However, proper external validation is less straightforward for QSAR models of mixtures, especially when the same compounds with different ratios are present throughout the training set. The conventional external cross-validation procedure is also not always acceptable for those cases – if both training and external sets include data points for the same mixture, the model’s true predictive performance will not be estimated properly. There is a consensus opinion that novel, more rigorous external validation protocols are required in the field. Depending on the initial data and potential application of the developed models, three different strategies of external validation could be used: (i) 'points out' – prediction of the investigated property for any composition of mixture from the modeling set, (ii) 'mixtures out' – filling of missing cells in the initial data (mixtures) matrix, i.e., prediction of the investigated property for mixtures with unknown activity created by combining pure compounds from the modeling set, and (iii) 'compounds out' – prediction of the investigated property for mixtures formed by novel pure compound(s) absent in the modeling set (the most rigorous method of external validation in QSAR modeling of mixtures). Furthermore, careful collection and understanding of initial data, thorough its curation, rigorous internal and external validation, and application of developed models for virtual screening of large databases (which are mostly absent in current mixture studies) will significantly improve the quality of QSAR models of mixtures.
The field of QSAR modeling of mixtures is very new and is under active development. Given the importance and the increasing need for such models, efforts directed to the development of new methods, and the improvement of existing QSAR approaches for mixtures are welcomed and encouraged.
3.3. QNAR – quantitative nanostructure-activity relationships
Combinatorial chemistry and HTS technologies have been recently extended towards designing novel manufactured nanoparticles (MNPs)., With more than 1,000 manufacturer-identified, nanotechnology-based consumer products available on the market, nanotechnology is drawing worldwide attention for its numerous applications in various industrial areas, such as material science, medical research, cosmetics, or even clothing. Importantly, a significant portion of these efforts is directed towards the development of 'green' products intended to achieve efficient and less polluting energy sources. In this context, QSAR science has a role to play by: (i) facilitating the access, storage, search, and integration of all experimental results currently distributed in literature, databases, and other sources;(ii) achieving externally predictive QSAR models to compute MNPs’ properties based on their structural characteristics; and (iii) boosting the development and testing processes by identifying the most promising nanoparticles that require focused experimental investigations. The latter point is especially true due to the concerns about the safety of certain MNPs and the development of nanomedicine., A growing compendium of experimental results shows that certain MNPs intended for industrial applications could cause toxic effects in humans.– Thus, computational tools capable of evaluating MNPs for their potential health risk are needed.
From a chemical perspective, MNPs are very different from small molecules in ways that make their modeling more challenging. Firstly, MNPs are characterized by high physical/structural complexity and diversity as they represent assemblies of inorganic and/or organic elements, sometimes mixed or coated. Moreover, the exact molecular stoichiometry may vary from one particle to another even in the same cohort, and there is a vast variety of particle categories with numerous potential applications, having ranges of desired and undesired physical, chemical, and biological properties.
These factors help to explain why there are no systematic quantitative studies of MNPs in the literature. Thomas et al. recently published a useful ontology for MNPs applicable in the field of cancer research, which allows for the integration of experimental results for nanoparticles. Similarly, the NanoTAB initiative deals with the development of a common database and file exchange format for nanotechnology information.
There are also relatively few literature reports on computational modeling of MNPs, especially in regard to nanotoxicology. Liu et al. used molecular dynamics simulations to reveal the overall changes in the structure of cellular membranes caused by the insertion of carbon nanotubes and to estimate affinity of drug-like molecules to the nanotubes. Puzyn et al. recently introduced a term 'nano-QSAR' and published a study on a small set of MNPs with metal-oxide cores. Published studies and developing trends in QSAR modeling of nanomaterials have been summarized in few reviews.,
In a recent proof-of-concept study, Fourches et al. introduced the terminology of Quantitative Nanostructure-Activity Relationships (QNARs) that employs classical machine-learning methods for establishing links between chemical descriptors and various measured activities of MNPs. Using an ensemble of 51 diverse NPs tested in vitro against four cell lines in four different assays at four different concentrations (resulting in 51x64 biological data points), they were able to identify three clusters of NPs for which QNAR models were obtained using four experimental descriptors: size, two measures of relaxivity, and zeta potential. The developed models resulted in an external accuracy of prediction as high as 73%. In another study, Fourches et al. modeled 109 cross-linked iron oxide NPs decorated with small organic ligands to predict their uptakes by PaCa2 pancreatic cancer cells. Models afforded a reasonable predictive power of R2 = 0.72 using a 5-fold external cross-validation procedure. More recently, the same group succeeded in modeling a set of 84 decorated carbon nanotubes tested in six different in vitro assays. Predictive QNAR models were obtained to accurately predict protein binding profiles and toxicological properties of these nanotubes. QNAR models were then utilized to screen a database of 200,000 ligands potentially attachable to carbon nanotubes in order to design new nanotubes with less protein binding affinities and less acute toxicity.
As nanomaterials continue to proliferate in many areas, computational methodologies such as QNAR modeling are expected to provide critical support to experimental studies to identify safe nanoparticles with desired properties. However, it is important to emphasize that such procedures require relatively large amounts of reliable and consistent experimental data where MNPs can be characterized by a set of physical chemical properties and tested in well-defined assays.
4. QSAR and regulatory decision support
Historical context
QSAR has had a long-standing history of use within the areas of environmental research and regulation, particularly for food use, cosmetics, and industrial chemicals where regulations are often limited, while property and toxicity data are sparse or unavailable. For more than 30 years, EPA has made an extensive use of QSAR models for hazard identification among new industrial chemicals (chemicals subject to pre-manufacturing / premarketing notification) under the Toxic Substances Control Act (TSCA), particularly in the area of ecotoxicology. These approaches include the use of expert systems, QSAR modeling, nearest analogue and chemical class analyses, and prediction of mechanisms of toxicity among others. To enable these efforts, EPA has supported the development and public release of various QSAR and decision support tools, including the Estimation Program Interface (EPI) Suite,281 which computes a variety of physical-chemical properties (based primarily on fragment-based QSAR approaches), and the Ecological Structure Activity Relationships (ECOSAR) software for predicting aquatic toxicity of compounds. The latter tool has also been demonstrated to have some useful applicability to pharmaceuticals.
Worldwide, QSAR methods have been used for identification of potential health hazards, screening and prioritization by various government agencies including Health Canada, the FDA, and the European Union (EU) authorities. In Canada, under the New Substances Provisions of the Canadian Environmental Protection Act (1999), regulators use QSAR predictions for assessing and prioritizing the Canadian inventory of existing substances (Domestic Substances List, DSL). Within the EU, the Danish EPA utilizes QSAR-generated endpoints for ecological and health hazard assessments, developing an advisory list for self-classification of dangerous substances where experimental test results are incomplete or unavailable.
Within the EU, the New Chemicals Policy of the European Commission (REACH: Registration, Evaluation and Authorisation and Restriction of Chemicals) has proposed a new system for managing chemical information in a single regulatory framework. According to that initiative, nearly 30,000 substances will be processed on a phased basis over a period of 11 years (2007–2018). An important part of this policy is the fostering of research on the development and validation of alternative (non-animal) methods, including QSAR models and in vitro tests. Importantly, within the REACH initiative, it was considered important to develop an internationally recognized set of principles for QSAR validation, to increase the confidence in QSAR predictions, and to provide regulatory bodies with a scientific basis for making decisions on the acceptability of QSAR estimates.
OECD principles
Some principles for assessing the validity of QSARs were proposed in 2002, as the 'Setúbal Principles', at the international workshop in Setubal, Portugal. Two years later in 2004 those were modified by the OECD Work Programme on QSARs, as the 'OECD principles for the Validation, for Regulatory Purposes, of QSAR Models'.89 The corresponding OECD principles are as follows:
'To facilitate the consideration of a QSAR model for regulatory purposes, it should be associated with: (1) a defined endpoint; (2) an unambiguous algorithm; (3) a defined domain of applicability; (4) appropriate measures of goodness-of-fit, robustness and predictivity; (5) a mechanistic interpretation, if possible.'
There are long-standing and ongoing debates in the scientific community regarding the so-called 'mechanistic versus statistic' approaches to QSAR modeling. The first approach considers Principle 5 as the most important and focuses mainly providing a mechanistic basis for a model and its predictions, whereas the second approach follows the order of OECD Principles in assessing different model characteristics, with the complete model validation process (internal and external), requested by Principle 4, carried out before the interpretation of descriptors for their mechanistic meaning, if possible (Principle 5).
Also it was recommended that the results of (Q)SARs may be used instead of testing when the following conditions are met: (1) results are derived from a (Q)SAR model whose scientific validity has been established; (2) the substance falls within the applicability domain of the (Q)SAR model; (3) results are adequate for the purpose of classification and labeling and/or risk assessment, and, (4) adequate and reliable documentation of the applied method is provided.
Available resources
The need for 'adequate and reliable' documentation requires standardized QSAR reporting formats, such as proposed in 'A (Q)SAR Model Reporting Format' (QMRF). The QMRF is a communication tool for reporting and assessing QSAR predictions that is meant to provide industry and regulators with reliable information. Access to this form is provided through the JRC (Joint Research Commission) QSAR Model Database, available through a web-interface.287 A similar project led by the OECD has developed the 'QSAR Toolbox', a software tool to facilitate grouping and category building and to assist in the development of read-across justifications and their transparent documentation, supporting the use of QSAR models in different regulatory frameworks.113 This approach is based on chemical categorical grouping, emerging from 'read-across' or trend analyses. It should be noted, however, that the 'read across' technique implies that endpoint information for an untested chemical can be predicted from the existing results for a compound deemed 'similar' in some way (e.g., activity, property, or structure). This clearly is a very simplistic approach that is not always able to accurately predict real trends in the data. This method is also known to be prone to failure when the notorious 'activity cliff' occurs. Clearly, further improvements in regulatory QSAR models are urgently needed.
To improve and promote the use of non-test methods for REACH, the European Commission and the European Chemicals Agency (ECHA), in cooperation with the stakeholders, has developed guidance and practical guides on 'QSARs and grouping of chemicals', 'How to report QSAR', 'How to report read-across and categories', and recently on 'How to avoid unnecessary testing on animals'. ECHA also coordinates, together with OECD, the continuous development of the 'QSAR Toolbox'.113
It is important to underscore that QSAR models supporting hazard or risk assessment must be relevant, reliable and sufficient for their intended purpose. It is one of the most important goals for a scientific community to develop a common agenda and approaches in the area of regulatory QSAR.
5. Conclusions: Best practices and the future of QSAR modeling
The intention of bringing together the present collection of perspectives and reports on QSAR modeling, each largely and separately authored by an active practitioner in the field, is to provide both a historical perspective, to convey where we have been and how we have arrived at the present, as well as to give the reader a flavor for the broadly diverse nature and applicability of QSAR practice, spanning a wide range of scientific disciplines and practical applications. A reader who has ventured this far will have detected a few seemingly disparate viewpoints, as well as some general, common themes that run throughout the various sections. We conclude by briefly elaborating further on a few of these areas of active (and healthy) disagreement within the QSAR community, as well as revisiting some larger common themes, touched on separately by many of the coauthors, for which there is broad consensus moving forward.
Best practices in QSAR modeling
General guidelines pertaining to QSAR modeling practice, governing elements of model construction, reporting, validation, and use have emerged from years of scientific practice and experience. Several sub-sections of this report have directly addressed these challenges, from a traditional practitioner (Section 1.3) and methodological modeling workflow standpoint (Sections 2.1 and 2.4), as well as from the standpoint of establishing guidelines for QSAR model acceptance for use in safety assessment in a regulatory setting (Section 3.4). Clearly, best practices are essential to ensuring the overall integrity and validity of any QSAR modeling study and, if not adhered to, can negatively impact the entire field. As with any maturing scientific discipline, the development of harmonized rules, standards and common practices are exceedingly useful. Given the multidisciplinary and diverse nature of the QSAR enterprise, however, these rules and practices must also be sufficiently encompassing and flexible to accommodate the wide range of problems to which QSAR is applied. Discussion of best practices in this review has primarily centered on four elements of a data workflow: (i) data collection and curation; (ii) model building; (iii) rigorous external validation using compounds that were not part of the modeling set; and (iv) model use. The latter can include prospective application to virtual screening and targeted molecular design of novel compounds or mixtures with desired properties, as well as application to toxicity screening and safety assessment within a regulatory setting. The approaches employed in a data workflow, possible pitfalls of modeling, and some strategies for avoiding common errors in QSAR model development were described in several recent reviews, and are highlighted in Sections 1.3 and 2.1 of this review.
Almost every cheminformatics lab has its own protocols for developing reliable QSAR models (e.g., discussed in studies)., The workflow described by Tropsha is recommended because of particular attention paid to rigorous internal and external cross-validation, estimation of AD, and the Y-scrambling test as necessary steps of model building. Consensus QSAR modeling is another highly recommended part of this workflow given increasing evidence that the quality of predictions and AD of consensus models are usually higher than for individual QSAR models. However, Thomas et al. have shown that, it is often impossible to build predictive models even when the most sophisticated algorithms and rigorous modeling workflows are employed. In response to this finding, Golbraikh et al. recently introduced the concept of 'dataset modelability', i.e., an a priori estimate of the feasibility to obtain externally predictive QSAR models for a dataset of bioactive compounds. This concept has emerged from analyzing the effect of 'activity cliffs', i.e., very similar compounds with very different activities, on the overall performance of QSAR models. In his seminal publication, Maggiora suggested that the presence of 'activity cliffs' in a dataset is significantly challenging for QSAR modeling. Later, SALI and ISAC scores were developed for identifying activity cliffs based on ligand- and structure-based approaches, respectively. Excellent perspectives recently published in the Journal of Medicinal Chemistry, are pointing out many issues posed by activity cliffs for cheminformatics investigations and cover related topic of molecular similarity. 'MODelability Index' (MODI) was proposed not only as a quantitative tool to quickly estimate whether predictive QSAR model(s) can be obtained for a given binary dataset but also as an attempt to answer the following questions: (i) how the number of activity cliffs in a given dataset correlates with the overall prediction performance of QSAR models for this dataset; (ii) whether such correlation is conserved across different datasets; (iii) whether one could use the fraction of activity cliffs in a datasets to assess the overall possibility of success or failure for QSAR modeling; (iv) why some datasets are modelable whereas others are not; and (v) how (and whether it is possible at all) to find the subset of compounds in overall non-modelable dataset, for which local QSAR models can be obtained.
Model validation and use
Rigorous external validation must be considered an integral part of model development. We separate these stages only to emphasize that a model should be externally validated using molecules which had no involvement in either model development or selection. The simplest approach is n-fold external cross-validation, where the entire dataset is randomly divided into n parts (folds) and each part is used as an external set for the model developed, and internally validated with the remaining compounds. The situation when new experimental data for new compounds become available after the model was built (exemplified by study) is even more preferable; it additionally strengthens external validation.
The discovery of novel bioactive chemical entities is the primary goal of computational drug discovery, and the development of validated and predictive QSAR models is critical to achieving this goal. Moreover, the study illustrated that robust QSAR models could be used for initial experimental data curation, i.e., to question true positives as well as to recover false negatives resulting from the high throughput screening campaigns.
Once a QSAR model is statistically validated, applicability domain (AD) estimation remains one of the most difficult and pressing challenges for QSAR modelers. Assessing the reliability of a QSAR for prospective predictions of properties or activities of new chemicals is a crucial adjunct to any model. The challenges and current approaches to defining the AD within a modeling workflow have been well described in several publications., We wish to add that since the AD derives from the model and descriptors to which it is applied, it carries the inherent limitations of that model. However, the same principles of validation applied to assessing a QSAR model can be iteratively applied to any proposed definition of the AD to determine its utility, demonstrating for external predictions that the model actually performs better within the AD than outside the AD. As with QSAR model validation, the ultimate proof of utility is judged by external performance.
Model Interpretability
The issue of QSAR model interpretability is explicitly considered in Section 2.5, from a methodological vantage point, but is discussed in other contexts throughout this review, most particularly in Sections 1.3, 2.2, 2.3, and 3.4. This issue is intimately tied to fundamental aspects of QSAR modeling, including selection of model descriptors and methods, and in some settings largely determines whether a model is ultimately deemed valid and useful. In Section 2.5, the goal of model interpretability is considered alongside the goal of achieving the best model performance and external predictivity by objective statistical measures, with a valid argument made that the latter should be the first and primary objective of any modeling enterprise. In the discussion of the OECD Principles, the point is also made that model (or mechanistic) interpretability is desirable, but is a secondary goal to demonstrated model validity.
Elsewhere in the text, particularly in the context of the regulatory applications of QSAR in toxicity modeling and safety assessment, the point is made that the ability to provide mechanistic support for model predictions is highly desirable and can add biological and chemical plausibility and weight-of-evidence to individual predictions. One could also argue that introducing prior knowledge into the initial selection of model descriptors or classifiers, which by definition lends interpretability, is particularly important in overcoming the perennial problem of insufficient data for robust modeling of complex toxicity in vivo endpoints, weak model statistics, and large uncertainties in applying models to new chemicals (see Section 2.2). Hence, interpretability and validated predictive power of QSAR models used for regulatory decision support can decrease the uncertainties of the toxicity endpoints predictions (see Section 2.2 and 3.4).
Promoting best practices
Although good practices of QSAR modeling have been well described in the literature ,298, many models published even in the recent literature fail to adhere to these practices. Bajorath attributes this phenomenon to the relative ease of computational modeling and the possibility to carry out advanced calculations without critical assessment and understanding of their scientific foundations and limitations. An additional factor is the highly multidisciplinary nature of the QSAR modeling enterprise: biologically oriented journals tend to place greater emphasis on the endpoint aspects of modeling studies and journal editors and reviewers tend to be less able to judge the quality of the QSAR modeling study from a methodological standpoint; by the same token, editors and reviewers of QSAR or cheminformatics oriented journals are less equipped to judge the biological appropriateness of the QSAR modeling study, in terms of the endpoint and data chosen for study.
In an effort to improve the quality of publications in the QSAR modeling field, the Journal of Chemical Information and Modeling published an editorial highlighting the requirements for QSAR papers that authors should follow to publish their results in the journal.298 Accompanied by more rigorous editorial and peer review standards, such guidelines have significantly decreased the number of low-quality QSAR publications in top cheminformatics journals. Additional efforts, including publication of the OECD principles, and reviews published by various authors, have all contributed to reductions in the publication of low-quality QSAR papers. There is a need to develop a minimal set of standards that any QSAR study should follow to be accepted by leading peer-reviewed journals. The proliferation of high-quality QSAR models could be helped by placing online such models developed by following best practices; recent appearance of portals such as OCHEM (http://ochem.eu), ChemBench (http://chembench.mml.unc.edu), end emerging NIH BARD project (https://bard.nih.gov/) are steps in this direction.
The continuing importance of QSAR
As with any scientific discipline, there have been some voices in the community questioning the viability and practical utility of QSAR modeling due to instances of poor model external performance, lax scientific practices, and the advent of newer biologically-based models built using HTS data. Paraphrasing the famous Mark Twain’s quote, we are confident that the 'reports of [QSAR] death are an exaggeration'. In this paper we have openly discussed challenges faced by QSAR modeling and offered guidelines for developing rigorous and properly validated QSAR models that, if followed, afford multiple and diverse successful applications of QSAR, as discussed herein. The enormous, continuing growth of data in various molecular sciences, from medicinal chemistry to nearly any 'omics' discipline, growing application of QSARs in regulatory decision making, emerging applications in materials (including nanomaterials) informatics suggest a growing importance of the QSAR approach to molecular data modeling. The developing trends on minimizing animal use in biomedical research place additional focus on QSAR as a source of alternative predictors of in vivo effects in both animals and humans. We hope that this paper will help both computational and experimental chemists to develop reliable QSAR models and to use these models to optimally exploit the experimental data to guide future studies. In addition, we hope that the guidelines presented here will help journal editors and reviewers apply more stringent scientific standards to manuscripts reporting new QSAR studies, as well as encourage the use of high quality, validated QSARs to provide reliable support for experimental study design and regulatory decision making.
We conclude with a final nod to Hansch, who founded the modern practice of QSAR over 50 years ago. In the latter half of his long and productive career, he championed the idea of 'comparative QSAR' and cofounded a company Biobyte (with A. Leo) to implement the approach of comparing related QSAR models to glean new insights into common mechanistic drivers for activity. Over a 30 year period, Hansch and coworkers compiled over 17,000 QSARs from the literature, with approximately half pertaining to biological systems and the other half to mechanistic organic chemistry, and published several studies illustrating the power of this approach to generate new insights (see review and references therein). Decades before the term 'cheminformatics' entered the lexicon of QSAR modelers, and likely without the knowledge or use of sophisticated machine-learning approaches, such as described here, Hansch foresaw the power of large, searchable databases and put 'inductive knowledge transfer' into practice. The field of QSAR is infinitely indebted to Hansch for extending careful investigations grounded in physical organic chemistry principles to the applied discipline of QSAR widely practiced today, and to the many investigators who have followed in his footsteps. We believe that he would be gratified to see that QSAR continues as a vibrant scientific enterprise and is advancing and contributing to many scientific disciplines along the paths he originally laid forth.
Acknowledgement
This review combined a series of separately written, invited contributions from the various coauthors (some sections with multiple coauthors). Primary attributions for the various contributed sections are as follows: 1.1 – Y. Martin; 1.2 – V. Consonni, R. Todeschini, and R. Cramer; 1.3 – J. Dearden and M. Cronin; 2.1 – D. Fourches, E. Muratov, and A. Tropsha; 2.2 – A. Benigni, C. Yang, J. Rathman, and A. Richard; 2.3 – L. Terfloth, J. Gasteiger; 2.4 – V. Kuz'min and E. Muratov; 2.5 – I. Baskin and A. Varnek; 2.6 – D. Fourches and A. Tropsha; 3.1 – A. Cherkasov; 3.2 – E. Muratov and V. Kuz'min; 3.3 – D. Fourches; 4. – P. Gramatica; and 5 – A. Tropsha. Figures were done by D. Fourches and E. Muratov. Substantial assistance in editing of all contributions was provided by A. Cherkasov and A. Richard, and final editing was accomplished by A. Tropsha, who also takes primary responsibility for the final content. The senior author also appreciates the support provided to his lab by NIH (grants GM096967 and GM066940) and EPA (grant RD 83499901). The first author acknowledges the support from the Natural Sciences and Engineering Research Council of Canada (Discovery Grants Program).
The views expressed in this article are those of the authors and do not necessarily reflect the views or policies of the U.S. Environmental Protection Agency and other research institutions. Mentioning of trade names or commercial products does not constitute endorsement or recommendation for use.
Abbreviations used
AD | applicability domain |
ADR | adverse drug reaction |
AMP | Antimicrobial Peptide |
ANN | Artificial Neural Network statistical approach |
CoMFA | Comparative Molecular Field Analysis |
CYP | cytochrome P450 |
FN | Feature Net |
(M)NP | (Manufactured) Nanoparticles |
MODI | MODelability Index |
MTL | Multi-Task Learning |
OECD | Organization for Economic Cooperation and Development |
PLS | Partial Least Square statistical approach |
Q2 | cross-validated determination coefficient |
qHTS | quantitative High-Throughput Screening |
QNAR | Quantitative Nanostructure-Activity Relationship |
(Q)SAR | (Quantitative) Structure-Activity Relationship |
(Q)SPR | (Quantitative) Structure-Property Relationship |
R | correlation coefficient |
R2 | determination coefficient |
REACH | Registration, Evaluation and Authorisation and Restriction of Chemicals |
RF | Random Forest statistical approach |
SEA | Similarity Ensemble Approach |
STL | Single Task Learning |
SVM | Support Vector Machine statistical approach |
UGT | UDP-glucuronosyltransferase |
VS | Virtual screening |
Biographies
Artem Cherkasov is an Associate Professor of the Faculty of Medicine, University of British Columbia, Vancouver, Canada. His research interests include computer-aided drug design, structure-activity modeling, drug reprofiling, and development of novel cheminformatics and bioinformatics tools. Dr. Cherkasov has authored and co-authored more than 200 articles in peer reviewed journals and conference proceedings and 5 book chapters. During his tenure at UBC, Dr. Cherkasov has been an applicant or co-applicant on a number of successful grants totaling over 40 million dollars, and participated in starting several spinoff companies. His research has been featured in various media sources including 'Financial Times', 'New Scientist', 'Toronto Star', 'Georgia Straight' newspapers, 'Montecristo' magazine and other media sources.
Eugene N. Muratov is a research assistant professor in the Eshelman School of Pharmacy, University of North Carolina, Chapel Hill and a senior researcher in the Department of Cheminformatics and Molecular Structure in the A. V. Bogatsky Physical-Chemical Institute of National Academy of Sciences (NAS) of Ukraine. He received his MS in technology of organic substances from Odessa National Polytechnic University in 2000 and PhD in organic chemistry in 2004 from the A. V. Bogatsky Physical-Chemical Institute. Before obtaining research assistant professor position in UNC in 2012, he spent several years as a researcher in Odessa, Jackson, Strasbourg, and Chapel Hill. His research interests are in the areas of cheminformatics (especially QSAR), computer-assisted drug design, antiviral research, and medicinal chemistry.
Denis Fourches, PhD is a Research Assistant Professor at the UNC Eshelman School of Pharmacy, University of North Carolina at Chapel Hill, USA. Prof. Fourches received his PhD in Chemistry in 2007 from the University of Strasbourg, France. He came to UNC in 2008 as a postdoctoral fellow in Alexander Tropsha's group. In 2010, he became a research assistant professor at UNC, and in 2011, he was promoted as the Associate Director of the Laboratory for Molecular Modeling. His research interests are in the areas of cheminformatics, computer-assisted molecular design, computational (nano)toxicology, material informatics, and structural bioinformatics.
Alexandre Varnek got his PhD in physical chemistry from the Institute of Inorganic and General Chemistry of the Russian Academy of Sciences, Moscow. In 1988–1995, he was Associate Professor in theoretical chemistry at the Moscow Mendeleyev University of Chemical Technology. In 1995, Alexandre joined the University of Strasbourg, France, where he holds the position of a Professor in theoretical chemistry, Head of the Laboratory on Cheminformatics and the Director of the master program in cheminformatics. His research interests focus on the development of new approaches and tools for virtual screening and on 'in silico' design of new compounds and chemical reactions.
Igor Baskin received his PhD (kandidat nauk) (1990) and habilitation (doctor nauk) (2010) from Lomonosov Moscow State University, Russia. After holding several positions at the Semenov Institute of Chemical Physics and Zelinsky Institute of Organic Chemistry of the Russian Academy of Sciences, Moscow, he joined in 2001 Lomonosov Moscow State University, where since 2005 he holds the position of a Leading Research Associate. He is regularly engaged as Visiting Scientist and Invited Professor at the University of Strasbourg, France. He has published more than 140 articles related to SAR/QSAR/QSPR methodology, medicinal chemistry, and molecular modeling. Igor Baskin is a member of the International Academy of Mathematical Chemistry since 2009. His current work focuses on the application of advances machine learning approaches in cheminformatics.
Mark Cronin is Professor of Predictive Toxicology at Liverpool John Moores University, England. He has over 25 years expertise in the development of quantitative structure-activity relationships and the application of cheminformatics to provide solutions in product development and risk assessment. Specifically, he has developed models to predict the human health effects and environmental toxicity of industrial chemicals, pharmaceuticals and cosmetics ingredients etc. Current work aims to develop strategies to predict toxicity from a firm mechanistic basis. This will allow for the grouping of chemicals based on so-called Adverse Outcome Pathways and robust read-across to fill data gaps. He has published over 190 papers and 4 books and worked on numerous collaborative projects in the area of in silico ADMET.
John Dearden is Emeritus Professor of Medicinal Chemistry in the School of Pharmacy and Biomolecular Sciences at Liverpool John Moores University, UK. His research includes the prediction of drug activity and toxicity, the prediction of environmental toxicity and fate, and the prediction of ADME properties, using the QSAR (quantitative structure-activity relationship) approach. He has published about 260 scientific papers and book chapters, mostly in the field of QSAR. He was elected to honorary membership of the Royal Pharmaceutical Society of Great Britain in 1993 in recognition of his QSAR research in pharmaceutical sciences. He was the recipient of the 2004 International QSAR Award in recognition of significant contributions to the advancement of QSARs in environmental sciences.
Paola Gramatica is full professor of Environmental Chemistry at the Department of Theoretical and Applied Sciences in Insubria University (Varese). She is leader of the QSAR Research Unit in Environmental Chemistry and Ecotoxicology. Since 1994, PG researches are on QSAR modeling and screening/ranking of organic compounds of environmental concern, mainly for PBT properties. The main interest is on validation of QSAR models; she was member of the OECD Task Force of QSAR experts for QSAR model validation for application in the REACH regulation and was actively involved in the definition of the OECD Principles on this topic. More recently, she was leader of the WP on QSAR model development for emerging pollutants in the EU-FP7 Project CADASTER. Web site: http://www.qsar.it
Yvonne Connolly Martin obtained a BA from Carleton College and a PhD from Northwestern University. Her whole career (retired 2006) was at Abbott Laboratories. She is the author of 'Quantitative Drug Design' book (1978 and 2010) and 100+ articles including highly cited ones on pharmacophore detection, 3D database searching, and molecular similarity. She is editor of several books, Perspectives Editor of the Journal of Computer-Assisted Drug Design and a titular member of the IUPAC Division VII. She served on several NIH study sections, various editorial boards, and as chair of 1977 Gordon Conference on QSAR, and as chair of the QSAR Society. Honors include AAAS fellow; life-time member of the Molecular Graphics and Modelling Society; and the Herman Skolnik award of the ACS CINF Division.
Roberto Todeschini is full professor (since 2000) of chemometrics at the Department of Environmental Sciences of the University of Milano-Bicocca (Milano, Italy), where he constituted the Milano Chemometrics and QSAR Research Group. His main research activities concern chemometrics, Quantitative Structure-Activity Relationships (QSAR), molecular descriptors, multicriteria decision making and software development. President of the International Academy of Mathematical Chemistry and president of the Italian Chemometric Society, he is author of more than 180 publications on international journals and of the books 'The Data Analysis Handbook', by I.E. Frank and R. Todeschini; Elsevier, 1994, 'Handbook of Molecular Descriptors', by R. Todeschini and V. Consonni; Wiley-VCH, 2000, and 'Molecular Descriptors for Cheminformatics', by R. Todeschini and V. Consonni; Wiley-VCH, 2009.
Viviana Consonni received her PhD in chemical sciences from the University of Milano in 2000 and is now full researcher of chemometrics and cheminformatics at the Department of Earth and Environmental Sciences of the University of Milano-Bicocca (Milano, Italy). She is a member of the Milano Chemometrics and QSAR Research Group and has 15 years’ experience in multivariate analysis, QSAR, molecular descriptors, and software development. She is co-author of more than 40 publications in peer-reviewed journals and the books “Handbook of Molecular Descriptors” (2000) and “Molecular Descriptors for Chemoinformatics” (2009) by R. Todeschini and V. Consonni, Wiley-VCH. In 2006, she obtained the International Academy of Mathematical Chemistry Award for distinguished young investigators and, in June 2009, was elected as the youngest Member of the Academy.
Victor E. Kuz'min is a full professor and vice-director of the A.V. Bogatsky Physical-Chemical Institute of National Academy of Sciences of Ukraine. He received his PhD in organic chemistry in 1980 and defended his Doctor of Sciences (habilitation) thesis in 2004 at the A.V. Bogatsky Physical-Chemical Institute, Odessa, Ukraine. He has been the head of the Laboratory of Theoretical Chemistry since 1980. He was the scientific advisor for more than 10 PhD students. His research interests are in the areas of theoretical chemistry, cheminformatics, computer- assisted drug design, computational toxicology, and medicinal chemistry. His research is supported by multiple grants from the STCU, INTAS, and other funds.
Richard D. Cramer has been Sr. VP for Science at Tripos, and now Certara, since 1983. From 1971 to 1983, he founded and led SmithKline’s CADD activity. He received an AB in Chemistry & Physics from Harvard in 1963 and a PhD in Physical Organic Chemistry from MIT in 1967. The other four years were as bench chemist at Polaroid and as post-doc with E.J. Corey at Harvard. Dick is the inventor of 3D-QSAR, as CoMFA; yet he believes his more recent topomer innovations will ultimately have greater significance for drug discovery. Dick is also a pioneer SABRmetrician (baseball “QSAR”) and was a founder and the chief technical officer of STATS, Inc, as detailed within the books “Moneyball” and “Numbers Game”.
Romualdo Benigni received his education in chemistry at the University of Rome “La Sapienza”. He then joined the Istituto Superiore di Sanita’ (Italian National Institute of Health) in 1977, and where he remained except for two sabbaticals (New York University, 1988; Jawaharlal Nehru University in New Delhi, 2000). He worked experimentally in the field of molecular biology and environmental chemical mutagenesis. In the 1980's, he turned his attention to the statistical analysis and modeling of toxicological data, and to the study of the relationships between the structure of organic compounds and their toxicological properties (mainly mutagenesis and carcinogenesis). He has published over 180 journal articles and book chapters, applying a wide variety of quantitative analysis techniques, including QSAR, to the examination of chemical toxicity information.
Chihae Yang is the Managing Director of Molecular Networks, GmbH and the Chief Scientific Officer of Altamira LLC. She is also a visiting scientist at US FDA CFSAN where she has been a fellow in the computational toxicology program (2008 – 2011). She previously was Chief Scientific Officer of Leadscope, Inc. She joined Leadscope in 2000 from her position as a tenured chemistry professor at Otterbein University and an adjunct professor at the Ohio State University. At Leadscope she developed Leadscope’s predictive toxicology platform and QSAR methodology. She also developed ToxML, a public toxicity database standard, and developed a ToxML database entry tool in collaboration with the US FDA. Her interests include methods to assist decisions in early discovery as well as in safety assessment.
Jim Rathman is a Professor of Chemical and Biomolecular Engineering at The Ohio State University. He is also a co-founder and Managing Director of Altamira, LLC, a company based in Columbus Ohio that provides consulting, database development, and software tools for molecular informatics applications, with a focus on computational risk assessment of complex chemical systems. His research areas of interest include chemical informatics, statistical data analysis and experimental design, and computational modeling and simulation. Jim previously spent seven years in industrial research and development with Conoco Inc. and The Clorox Company.
Lothar Terfloth, Ph.D., Senior Research Scientist, joined Molecular Networks GmbH in 2007 after his position as a research scientist in Prof. Dr. Gasteiger’s group at the Computer-Chemistry-Center University Erlangen-Nuremberg. He has more than ten years of experience in the field of cheminformatics, the application of artificial neural networks and machine learning methods in drug discovery. His current research interest is directed towards the field of computational toxicology and metabolism prediction. Dr. Terfloth received his diploma in chemistry and his Ph.D. from the University of Munster working on the investigation of chiral organolithium compounds, their characterization by NMR spectroscopy as well as ab initio and DFT calculations.
Prof. Dr. Johann Gasteiger studied chemistry and has had positions at the University of Munich, Germany, University of California, Berkeley, USA, and Technical University of Munich. In 1994 he moved to the University of Erlangen-Nuremberg, Germany where he cofounded the 'Computer-Chemie-Centrum'. He is one of the initiators of Cheminformatics in Germany and has produced more than 250 scientific publications in this field. His work has been recognized by several awards, including the Gmelin–Beilstein Medal of the German Chemical Society, awards from the ACS Division of Chemical Information and of the ACS Division of Computers in Chemistry as well as of the Chemical Structure Association. In 1997 he founded Molecular Networks GmbH, a company developing and distributing software for chemical applications.
Ann Richard, PhD, is a Research Chemist within the Environmental Protection Agency’s Office of Research & Development. She was a charter member and has worked for the past 7 years in EPA’s National Center for Computational Toxicology (NCCT). Her research activities have ranged from the application of computational chemistry and SAR methods to problems in environmental toxicology, with a focus on predictive toxicology to, more recently, development of cheminformatics capabilities and knowledge-informed, feature-based approaches in support of predictive toxicology. Within NCCT, she is lead for the DSSTox project and chemical library information management for the ToxCast and Tox21 projects, working to provide a foundation for improved toxico-chemoinformatics and SAR capabilities in predictive toxicology.
Alexander Tropsha, PhD. is K.H. Lee Distinguished Professor and Associate Dean for Research at the UNC Eshelman School of Pharmacy, UNC-Chapel Hill. Prof. Tropsha received his PhD in Chemical Enzymology in 1986 from Moscow State University, Russia. He came to UNC-Chapel Hill in 1989 as a postdoctoral fellow, and eventually became a full professor in 2003. His research interests are in the areas of Computer-Assisted Drug Design, Computational Toxicology, Cheminformatics, and Structural Bioinformatics. He has authored or co-authored more than 170 peer-reviewed research papers, reviews and book chapters and co-edited two monographs. His research is supported by multiple grants from the NIH, NSF, EPA, ONR, and industry. He is a member of editorial boards of several scientific journals and scientific advisory panels.
References
3d Qsar Software As A Services
3d Qsar Software As A Service Center
Software As A Service Wikipedia
- Kim, K.H., Greco, G. and Novellino, E., A critical review of recent CoMFA applications, In Kubinyi, H., Folkers, G., and Martin, Y.C., (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 257–316.Google Scholar
- Dunn III, W.J. and Hopfinger, A.J., 3D QSAR of flexible molecules using tensor representation, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 167–182.Google Scholar
- Hahn, M. and Rogers, D., Receptor surface models, in Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 117–134.Google Scholar
- Heritage, T.W., Ferguson, A.M., Turner, D.B. and Willett, P., EVA — a novel theoretical descriptor for QSAR studies, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 2, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 381–398.Google Scholar
- Klebe, G., Comparative molecular similarity indices analysis — CoMSIA, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 87–104.Google Scholar
- Walters, D.E., Genetically evolved receptor models (GERM) as a ID QSAR tool, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 159–166.Google Scholar
- Wade, R.C., Ortiz, A.R. and Gago, F., Comparative binding energy analysis. In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 2, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 19–34.Google Scholar
- Holloway, M.K., A priori prediction of ligand affinity by energy minimization, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 2, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 63–84.Google Scholar
- Todeschini, R. and Gramatica, P., New 3D molecular descriptors: The WHIM theory and QSAR applications. In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 2, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 355–380.Google Scholar
- Silverman, B.D., Platt, D.E., Pitman, M. and Rigoutsos, I., Comparative molecular moment analysis (COMMA), in Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 183–196.Google Scholar
- Jain, A.N., Koile, K. and Chapman, D., Compass: Predicting biological activities from molecular surface properties — performance comparisons on a steroid benchmark, J. Med. Chem., 37 (1994) 2315–2327.Google Scholar
- Martin, Y.C., Kim, K.-H. and Lin, C.T., Comparative molecular field analysis: CoMFA, In Charton, M. (Ed.) Advances in quantitative structure property relationships, JAI Press, Greenwich, CT, 1996, pp. 1–52.Google Scholar
- Greco, G., Novellino, E. and Martin, Y.C., Approaches to 3D-QSAR, In Martin, Y.C. and Willett, P. (Eds.) Designing bioactive molecules: Three-dimensional techniques and applications, America Chemical Society, Washington, DC, 1997 (in press).Google Scholar
- Ajay, and Murcko, M.A., Computational methods to predict binding free-energy in ligand—receptor complexes, J. Med. Chem., 38 (1995) 4953–4967.Google Scholar
- Kollman, P.A., Advances and continuing challenges in achieving realistic and predictive simulations of the properties of organic and biological molecules, Acc. Chem. Res., 29 (1996) 461–469.CrossRefGoogle Scholar
- Bush, B.L. and Nachbar Jr., R.B., Sample-distance partial least-squares — PLS optimized for many variables, with application to CoMFA, J. Comput.-Aided Mol. Design, 7 (1993) 587–619.CrossRefGoogle Scholar
- Burger, A., Medical chemistry — the first century, Med. Chem. Res., 4 (1994) 3–15.Google Scholar
- Willett, P., Similarity and clustering techniques in chemical information systems, Research Studies Press, Letchworth, 1987.Google Scholar
- Hodgkin, E.E. and Richards, W.G., Molecular similarity based on electrostatic potential and electric field. Int. J. Quantum Chem., 14(1987) 105–110.Google Scholar
- Kier, L.B., Molecular orbital theory in drug research. Academic Press, New York, 1971, p. 258.Google Scholar
- Martin, Y.C., Pharmacophore mapping. In Martin, Y.C. and Willett, P. (Eds.) Designing bioactive molecules: Three-dimensional techniques and applications, American Chemical Society, Washington, DC, 1997 (in press).Google Scholar
- Free, S.M. and Wilson, J., A mathematical contribution to structure-activity studies. J. Med. Chem., 7 (1964) 395–399.CrossRefGoogle Scholar
- Pauling, L., Campbell, D.H. and Pressman, D., The nature of the forces between antigen and antibody and of the precipitation reaction. Physiol. Rev., 23 (1943) 203–219.Google Scholar
- Allen, F.H., Kennard, O. and Taylor, R., Systematic analysis of structural data as a research tool in organic chemistry, Acc. Chem. Res., 16 (1983) 146–153.CrossRefGoogle Scholar
- Bürgi, H.-B. and Dunitz, J.D., Structure Correlation, 1st Ed., VCH Verlagsgesellschaft mbH, Weinheim, Germany, 1994, Vols. 1 and 2, pp. 900.Google Scholar
- Allen, F.H., Bird, C.M., Rowland, R.S., Harris, S.E. and Schwalbe, C.H., Correlation of the hydrogenbond acceptor properties of nitrogen with the geometry of the Nsp (2)-Nsp(3) transition in R(1)(X=)C-NR(2)R(3) substructures—Reaction pathway for the profanation of nitrogen, Acta Crystallogr., Sec. B, 51 (1995) 1068–108.Google Scholar
- Mills, J. and Dean, P.M., 3-Dimensional hydrogen-bond geometry and probability information from a crystal survey, J. Comput.-Aided Mol. Design, 10 (1996) 607–622.CrossRefGoogle Scholar
- Åqvist, J., Medina, C. and Samulesson, J.-E., A new method for predicting binding affinity in computeraided drug design, Protein Eng., 7 (1994) 385–391.Google Scholar
- Dirac, P.A.M., Proc. R. Soc. London, Ser. A, 123 (1929) 714.Google Scholar
- Dewar, M.J.S., Zoebish, E.G., Healy, E.F. and Stewart, J.J.P., AMI: A new general purpose quantum mechanical molecular model, J. Am. Chem. Soc., 107 (1985) 3902–3909.Google Scholar
- Clark, T., A handbook of computational chemistry: A practical guide to chemical structure and energy calculations, Wiley, New York, 1985, pp. 332.Google Scholar
- Stewart, J.P., Semiempirical molecular orbital methods, In Lipkowitz, K.B. and Boyd, D.B. (Eds.) Reviews in computational chemistry, VCH, Weiheim, Germany, 1990, pp. 45–81.Google Scholar
- Kroemer, R.T., Hecht, P. and Liedl, K.R., Different electrostatic descriptors in comparative molecularfield analysis: A comparison of molecular electrostatic and Coulomb potentials, J. Comput. Chem., 17 (1996) 1296–1308.3.0.CO%3B2-H'>CrossRefGoogle Scholar
- Cramer, C.J. and Truhlar, D.G., AM1-SM2 and PM3-SM3 parameterized SCF salvation models for free energies in aqueous solution, J. Comput.-Aided Mol. Design, 6 (1992) 629–666.CrossRefGoogle Scholar
- Klamt, A. and Schuurmann, G., COSMO: A new approach to dielectric screening in solvents with explicit expressions for the screening energy and its gradient, J. Chem. Soc., Perkin Trans. 2, (1993) 799–805.Google Scholar
- Giesen, D.J., Chambers, C.C., Cramer, C.J. and Truhlar, D.G., Salvation model for chloroform based on class-IV atomic charges, J. Phys. Chem. B, 101 (1997) 2061–2069.Google Scholar
- Richardson, W.H., Peng, C., Bashford, D., Noodleman, L. and Case, D.A., Incorporating solvation effects into density-functional theory: Calculation of absolute acidities, Int. J. Quantum Chem., 61 (1997) 207–217.3.0.CO%3B2-#'>CrossRefGoogle Scholar
- Hammett, L., Physical organic chemistry, McGraw-Hill, New York, 1970.Google Scholar
- Hansch, C. and Fujita, T., Rho Sigma pi analysis: A method for the correlation of biological activity and chemical structure, J. Am. Chem. soc., 86 (1964) 1616–1626.Google Scholar
- Hansch, C. and Leo, A., Exploring QSAR: Fundamentals and applications in chemistry and biology, American Chemical Society, Washington, DC, 1995, pp. 557.Google Scholar
- Hansch, C., Leo, A. and Hoekman, D., Exploring QSAR: Hydrophobic, electronic, and steric constants, American Chemical Society, Washington, DC, 1995, pp. 348.Google Scholar
- Burkert, U. and Allinger, N.L., Molecular mechanics, American Chemical Society, Washington, DC, 1982, pp. 339.Google Scholar
- Marshall, G.R., Barry, C.D., Bosshard, H.E., Dammkoehler, R.A. and Dunn, D.A., The conformation parameter in drug design: The active analog approach. In Olson, E.C. and Christoffersen, R.E. (Eds.) Computer-assisted drug design, American Chemical Society, Washington, DC, 1979, pp. 205–226.Google Scholar
- Langridge, R., Ferrin, T.E., Kuntz, I.D. and Connolly, M.L., Real-time color graphics in studies of molecular interactions, Science, 211 (1981) 661–667.Google Scholar
- Blaney, J.M., Jorgensen, E.G., Connolly, M.L., Ferrin, T.E., Langridge, R., Oatley, S.J., Burridge, J.M. and Blake, C.C.F., Computer graphics in drug design: Molecular modeling of thyroid hormoneprealbumin interactions, J. Med. Chem., 25 (1982) 785–790.CrossRefGoogle Scholar
- Weiner, P.K., Langridge, R., Blaney, J.M., Schaefer, R. and Kollman, P.A., Electrostatic potential molecular-surfaces, Proc. Natl. Acad. Sci. U.S.A., 79 (1982) 3754–3758.Google Scholar
- Martin, Y.C., Quantitative drug design, Dekker, New York, 1978, pp. 425.Google Scholar
- Fujita, T., The role of QSAR in drug design. In Jolles, G. and Wolldridge, K.R.H. (Eds.) Drug design: Fact or fantasy?. Academic Press, London, 1984, pp. 19–33.Google Scholar
- Boyd, D.B., Successes of computer-assisted molecular design, In Lipkowitz, K.B. and Boyd, D.B. (Eds.) Reviews in computational chemistry. VCH, New York, 1990, pp. 355–371.Google Scholar
- Hansch, C., and Fujita, T., (Ed.), Classical and three-dimensional QSAR in agrochemistry, American Chemical Society, Washington, DC, 1995, 342 pp.Google Scholar
- Weiniger, D., A Note on the sense and nonsense of searching 3-D databases for pharmaceutical leads, Network Science, (1995). www.awod.com/netsci/Science/Cheminform/feature 04.html.Google Scholar
- Brown, R.D. and Martin, Y.C., Use of structure-activity data to compare structure-based clustering methods and descriptors for use in compound selection, J. Chem. Inf. Comput. Sci., 36 (1996) 572–584.CrossRefGoogle Scholar
- Brown, R.D. and Martin, Y.C., The information content of 2D and 3D structural descriptors relevant to ligand-receptor binding, J. Chem. Inf. Comput. Sci., 37 (1997) 1–9.CrossRefGoogle Scholar
- Brown, R.D., Danaher, E., Lico, I. and Martin, Y.C., unpublished observations.Google Scholar
- Kirn, K.H. and Martin, Y.C., Evaluation of electrostatic and steric descriptors of 3D-QSAR: The H+ and CH3 probes using comparative molecular field analysis (CoMFA) and the modified partial least squares method, In Silipo, C. and Vittoria, A. (Eds.) QSAR: Rational approaches to the design of bioactive compounds, Elsevier Science Publishers, Amsterdam, The Netherlands, 1991, pp. 151–54.Google Scholar
- Kamlet, M., Doherty, R., Fiserova-Bergerova, V., Carr, P., Abraham, M. and Taft, R., Solubility properties in biological media: 9. Prediction of solubility and partition of organic nonelectrolytes in blood and tissues from solvatochronic parameters., J. Pharm. Sci., 76 (1987) 14–17.Google Scholar
- Klopman, G., Artificial intelligence approach to structure-activity studies: Computer automated structure evaluation of biological activity of organic molecules, J. Am. Chem. Soc., 106 (1984) 7315–7321.CrossRefGoogle Scholar
- Hall, L.H. and Kier, L.B., The molecular connectivity chi indexes and kappa shape indexes in structure-property modeling, In Lipkowitz, K.B. and Boyd, D.B. (Eds.) Reviews in computational chemistry, VCH, New York, 1991, pp. 367–422.Google Scholar
- Van de Waterbeemd, H., Clementi, S., Costantino, G., Carrupt, P.-A. and Testa, B., CoMFA-derived substituent descriptors for structure-property correlations, In Kubinyi, H. (Ed.) 3D QSAR in drug design: Theory, methods and applications, ESCOM, Leiden, The Netherlands, 1993, pp. 697–707.Google Scholar
- van de Waterbeemd, H. (Ed.), Chemometric methods in molecular design, VCH, Weinheim, Germany, 1995, 359 pp.Google Scholar
- Hansch, C., Unger, S.H. and Forsythe, A.B., Strategy in drug design: Cluster analysis as an aid in the selection of substituents, J. Med. Chem., 16 (1973) 1212–1222.Google Scholar
- Wootton, R., Cranfield, R., Sheppey, G.C. and Goodford, P.J., Physicophemical-activity relationships in practice: 2. Rational selection of benzenoid substituents, J. Med. Chem., 18 (1975) 607–613.CrossRefGoogle Scholar
- Martin, Y.C. and Panas, H.N., Mathematical considerations in series design, J. Med. Chem., 22 (1979) 784–791.Google Scholar
- Austel, V., Experimental design in synthesisis planning and structure-property correlations, In van de Waterbeemd, H. (Ed.) Chemometric methods in molecular design, VCH, Weinheim, Germany, 1995, pp. 49–62.Google Scholar
- Downs, G.M. and Willett, P., Clustering in chemical-structure databases for compound selection. In van der Waterbeemd, H. (Ed.) Chemometric methods in molecular design, VCH, Weinheim, Germany, 1994, pp. 111–30.Google Scholar
- Martin, Y.C., Brown, R.D. and Bures, M.G., Quantifying diversity. In Kerwin, J.F. and Gordon, E.M. (Eds.) Combinatorial chemistry and molecular diversity, Wiley, New York, 1997 (in press).Google Scholar
- Turner, D.B., Tyrrell, S.M. and Willett, P., Rapid quantification of molecular diversity for selective database acquisition, J. Chem. Inf. Comput. Sci., 37 (1997) 18–22.CrossRefGoogle Scholar
- Simon, Z., Dragomir, N., Plauchitiu, M.G., Holban, S., Glatt, H. and Kerek, P., Receptor site mapping for cardiotoxic aglicones by the minimal steric difference method, Eur. J. Med. Chem., 15 (1980) 521–527.Google Scholar
- Hopfinger, A.J., A QSAR investigation of dihydrofolate reductase inhibition by Baker triazines based upon molecular shape analysis, J. Am. Chem. Soc., 102 (1980) 7196–7206.CrossRefGoogle Scholar
- Höltje, H.-D. and Kier, L.B., Sweet taste receptor studies using model interaction energy calculations, J. Pharm. Sci., 63 (1974) 1722–1725.Google Scholar
- Goodford, P.J., A computational procedure for determining energetically favorable binding sites on biologically important macromolecules, J. Med. Chem., 28 (1985) 849–857.CrossRefGoogle Scholar
- Kato, Y., Itai, A. and Iitaka, Y., A novel method for superimposing molecules and receptor mapping, Tetrahedron, 43 (1987) 5229–5234.Google Scholar
- Doweyko, A.M., The hypothetical active site lattice: An approach to modeling active sites from data on inhibitor molecules, J. Med. Chem., 31 (1988) 1396–1406.CrossRefGoogle Scholar
- Wold, S., Ruhe. A., Wold, H. and Dunn, W.J., The collinearity problem in linear regression: The partial least square (PLS) approach to generalized inverses, Siam J. Sci. Stat. Comput., 5 (1984) 735–743.CrossRefGoogle Scholar
- Cramer III, R.D., Patterson, D.E. and Buncc, J.D., Comparative molecular field analysis (CoMFA): 1. Effect of shape on binding of steroids to carrier proteins, J. Am. Chem. Soc., 110 (1988) 5959–5967.CrossRefGoogle Scholar
- Kim, K.H. and Martin, Y.C., Direct prediction of dissociation constants (pK a-s) of clonidine-like imidazolines, 2-.substituted imidazoles, and 1-melhyl-2-substituted-imidazoles from 3D structures using a comparative molecular field analysis (CoMFA) approach, J. Med. Chem., 34 (1991) 2056–2060.Google Scholar
- Kim, K.H., Comparison of classical and 3D QSAR, In Kubinyi, H. (Ed.) 3D QSAR in drug design: Theory methods and applications, ESCOM, Leiden, The Netherlands, 1993, pp. 619–642.Google Scholar
- Waller, C.L., Oprea, T.I., Giolitti, A. and Marshall, G.R., Three-dimensional QSAR of human immunodeficiency virus (I) protease inhibitors: 1. A CoMFA study employing experimentally-determined alignment rules, J. Med. Chem., 36 (1993) 4152–4160.Google Scholar
- Klebe, G. and Abraham, U., On the prediction of binding properties of drug molecules by comparative molecular field analysis, J. Med. Chem., 36 (1993) 70–80.CrossRefGoogle Scholar
- Watson, K.A., Mitchcll, E.P., Johnson, L.N., Cruciani, G., Son, J.C., Bichard, C.J.F., Fleet, G.W.J., Oikonomakos, N.G., Kontou, M. and Zographos, S.E., Glucose analog inhibitors of glycogenphosphorylase—from crystallographic analysis to drug prediction using grid force-field and GOLPE bariable selection, Acta Crystallogr., Sec. D, 51 (1995) 458–472.Google Scholar
- Jorgensen, W.L. and Tiradorives, J., Free-energies of hydration for organic-molecules from Monte Carlo Simulations, Persp. Drug Discov. Design, 3 (1995) 123–138.Google Scholar
- Marrone, T.J., Gilson, M.K. and McCammon, J.A., Comparison of continuum and explicit models of salvation—potentials of mean force for allanine dipeptide, J. Phys. Chem., 100 (1996) 1439–1441.CrossRefGoogle Scholar
- Madura, J.D., Nakajima, Y., Hamilton, R.M., Wierzbicki, A. and Warshel, A., Calculations of the electrostatic free-energy contributions to the binding free-energy of sulfonamides to carbonic-anhydrase. Struct. Chem, 7 (1996) 131–138.CrossRefGoogle Scholar
- Aqvist, J. and Mowbray, S.L., Sugar recognition by a gliico.se/galactose receptor: Evaluation of binding energetics from molecular dynamics simulations, J. Biol. Chem., 270 (1995) 9978–9981.Google Scholar
- Hansson, T. and Aqvist, J., Estimation of binding free-energies for HIV proteinase-inhibitors by molecular-dynamics simulations, Protein Eng., 8 (1995) 1137–1144.Google Scholar
- Paulsen, M.D. and Ornstein, R.L., Binding free-energy calculations for P450cam-subslrate complexes, Protein Eng., 9 (1996) 567–571.CrossRefGoogle Scholar
- Hulten, J., Bonham, N.M., Nillroth. U., Hansson, T., Zuccarello, G., Bouzide, A., Åqvist, J., Classon, B., Danielson, U.H., Karlen, A., Kvarnstrom, I., Samuelsson, B. and Hallberg, A., Cyclic HIV-1 protease inhibitors derived from mannitol: synthesis, inhibitory potencies, and computational predictions of binding affinities, J. Med. Chem., 40 (1997) 885–897.CrossRefGoogle Scholar
- Backbro, K., Lowgren, S., Osterlund, K., Atepo, J., Unge, T., Hulten, J., Bonham, N.M., Schaal, W., Karlen, A. and Hallberg, A., Unexpected binding mode of a cvelic sulfamide HIV-1 protease inhibitor, J. Med. Chem., 40 (1997) 898–902.CrossRefGoogle Scholar
- Blaney, J.M. and Dixon, J.S., A good ligand is hard to find: Automated docking methods, Persp. Drug Discovery Design, 1 (1993) 301–319.Google Scholar
- Böhm, H.-J., Ligand design, In H. Kubinyi (Ed.) 3D QSAR in drug design: theory, methods and applications, ESCOM, Leiden, The Netherlands, 1993, pp. 386–405.Google Scholar
- Böhm, H.-J., The development of a simple empirical scoring function to estimate the binding constant for a protein-ligand complex of known three-dimensional structure, J. Comput.-Aided Mol. Design, 8 (1994) 243–256.Google Scholar
- Head, R.D., Smythe, M.L., Oprea, T.I., Waller, C.L., Green, S.M. and Marshall, G.R., VALIDATE: A new method for the receptor-based prediction of binding affinities of novel ligands, J. Am. Chem. Soc., 118 (1996) 3959–3969.CrossRefGoogle Scholar
- Jain, A.N., Scoring noncovalent protein-ligand interactions: a continuous differentiable function tuned to compute binding affinities, J. Comput.-Aided Mol. Design, 10 (1996) 427–40.CrossRefGoogle Scholar
- Dixon, S. and Blaney, J., Docking, In Martin, Y.C. and Willett, P. (Eds.) Designing bioactivc molecules: Three-dimensional techniques and applications, American Chemical Society, Washington, DC, 1997 (in press).Google Scholar
- Holloway, M.K., Wai, J.M., Halgren, T.A., Fitzgerald, P.M.D., Vacua, J.P., Dorsey, B.D., Levin, R.B., Thompson, W.J., Chen, L.J., deSolms, S.J., Gaffin, N., Ghosh, A.K., Giuliani, E.A., Graham, S.L., Guare, J.P., Hungate, R.W., Lyle, T.A., Sanders, W.M., Tucker, T.J., Wiggins, M., Wiscount, C.M., Woltersdorf, O.W., Young, S.D., Darke, P.L. and Zugay, J.A., A priori predict/on of activity for HIV-1 protease inhibitors employing energy minimization in the active site, J. Med. Chem., 38 (1995) 305–317.CrossRefGoogle Scholar
- Ortiz, A.R., Pisaharro, M.T., Gago, F. and Wade, R.C., Prediction of drug binding affinities by comparative binding energy ana/ysis, 3. Med. Chem., 38 (1995) 2681–2691.CrossRefGoogle Scholar
- Reddy, B.V.B., Gopal, V. and Chatterji, D., Recognition of promoter DNA by subdomain-2 in-4.2 of Escherichia-Coli-sign(70): A knowledge-based model of-35-hexamer interaction with 4.2-helix-lurnhelix motif, J. Biomol. Struct. Dynamics, 14 (1997) 407–419.Google Scholar
- Weber, I.T. and Harrison, R.W., Molecular mechanics calculations on protein-ligand complexes, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 2, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 115–127.Google Scholar
- Wallqvist, A., Jeering, R.L. and Coeval, D.G., A preference-based free-energy parameterization of enzyme-inhibitor binding: Applications to HIV-1-protease inhibitor design, Protein Science, 4 (1995) 1881–1903.CrossRefGoogle Scholar
- Wallqvist, A. and Covell, D.G., Docking enzyme-inhibitor complexes using a preference-based free-energy surface, Proteins: Struct. Funk. Genet., 25 (1996) 403–411.Google Scholar
- Dewitt, R.S. and Shakhnovich, E.I., Smog — de novo design method based on simple, fast, and accurate free-energv estimates: 1. Methodology and supporting evidence, J. Am. Chem. Soc., 118 (1996) 11733–11744.Google Scholar
- Mattos, C., and Ringe, D., Multiple binding modes. In Kubinyi, H. (Ed.) 3D QSAR in drug design: Theory, methods and applications, ESCOM, Leiden, The Netherlands, 1993, pp. 226–254.Google Scholar
- Meyer, E.F., Boots, I., Scapozza, L. and Zhang, D., Backward binding and other structural surprises. Persp. Drug Discov. Design, 3 (1996) 168–195.Google Scholar
- Klebe, G., Mietzner, T., and Weber, P., Different approaches toward an automatic structural alignment of drug molecules: Applications to sterol mimics, thrombin and thermolysin inhibitors. J. Comput.-Aided Mol. Design, 8 (1994) 751–778.Google Scholar
- Oprea, T.I., Waller, C.L. and Marshall, G.R., Three dimensional quantitative structure-activity relation-ship of human immunodeficiency virus (I) protease Inhibitors: 2. Predictive power using limited exploration of alternate binding modes, J. Med. Chem., 37 (1994) 2206–2215.CrossRefGoogle Scholar
- DePriest, S.A., Mayer, D., Naylor, C.B. and Marshall, G.R., 3D-QSAR of angiotensin-converting enzyme and thermolysin inhibitors: A comparison of CoMFA models based on deduced and experimentally determined active-site geometries, J. Am. Chem. Soc., 115 (1993) 5372–5384.CrossRefGoogle Scholar
- Schoenleber, R., Martin, Y.C., Wilson, M., DiDomenico, S., Mackenzie, R.G., Artman, L.D., Ackerman, M.S., DeBernardis, J.K, Meyer, M.D., De, B., Hsiao, C.W. and Kebabian, J.W., American Chemical Society Meeting, August, New York, 1991.Google Scholar
- Martin, Y.C., Kebabian, J.W., MacKenzie, R. and Schoenleber, R., Molecular Modeling-based Design of Novel, Selective, Potent D1 Dopamine Agonists, In Silipo, C. and Vittoria, A. (Eds.) QSAR: Rational approaches on the design of bioactive compounds, Elsevier, Amsterdam, The Netherlands, 1991, pp. 469–482.Google Scholar
- Glen, R., Martin, G., Hill, A., Hyde, R., Woollard, P., Salmon, J., Buckingham, J. and Robertson, A., Computer-aided-design and synthesis of 5-substituted tryptamines and their pharmacology at the 5-HT1D receptor — discovery of compounds with potential antimigraine properties, J. Med. Chem., 38 (1995) 3566–3580.Google Scholar
- Waller, C.L. and Marshall, G.R., Three-dimensional quantitative structure-activity relationship of angiotensin-converting enzyme and thermolysin inhibitors: 2. A comparison of CoMFA models incorporating molecular-orbital fields and desolvation free-energies based on active-analog and complementary-receptor field alignment rules., J. Med. Chem., 36 (1993) 2390–2403.Google Scholar
- Klebe, G., Structural alignment of molecules. In Kubinyi, H. (Ed.) 3D QSAR in drug design: Theory, methods and applications, ESCOM, Leiden, The Netherlands, 1993, pp. 173–99.Google Scholar
- Kroemer, R.T., Hecht, P., Guessregen, S. and Liedl, K.R., Improving the predictive quality of CoMFA models, In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 41–56.Google Scholar
- Norinder, U., Recent progress in CoMFA methodology and related techniques. In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 25–39.Google Scholar
- Lin, C.T., Pavlik, P.A. and Martin, Y.C., Use of molecular fields to compare series of potentially bioactive molecules designed by scientists or by computer. Tetrahedron Comput. Method., 3 (1990) 723–738.CrossRefGoogle Scholar
- Norinder, U., Experimental design based 3-D QSAR analysis of steroid-protein interactions: Application to human CRG complexes, J. Comput.-Aided Mol. Design, 4 (1990) 381–389.CrossRefGoogle Scholar
- Caliendo, G., Greco, G., Novellino, E., Perissutti, E. and Santagada, V., Combined use of factorial design and comparative molecular field analysis (CoMFA): A case study, Quant. Struct.-Act. Relat., 13 (1994) 249–261.Google Scholar
- Mabilia, M., Belvisi, L., Bravi, G., Catalano, G. and Scolastico, C., A PCA/PLS analysis on nonpeptide angiotensin II receptor antagonists. In Sanz, F., Giraldo, J. and Manaut, F. (Eds.) QSAR and molecular modeling: Concepts, computational tools and biological applications. Proceedings of the l0th European Symposium on Structure-Activity Relationships: QSAR and Molecular Modeling, Barcelona, 4–9 September 1994, Prous, Barcelona, 1995, pp. 456–60.Google Scholar
- Cramer III, R.D., Clark, R.D., Patterson, D.E. and Ferguson, A.M., Bioisosterism as a molecular diversity descriptor — steric fields ofsingle topomeric conformers, J. Med. Chem., 39 (1996) 3060–3069.CrossRefGoogle Scholar
- Mager, P.P., A random number experiment to simulate resample model evaluations, J. Chemometrics, 10 (1996) 221–240.3.0.CO%3B2-2'>CrossRefGoogle Scholar
- Clark, M. and Cramer III, R.D., The probability of chance correlation using partial least squares (PLS), Quant. Struct.-Act. Relat., 12 (1993) 137–145.Google Scholar
- Doweyko, A.M., Three-dimensional pharmacophores from binding data, J. Med. Chem., 37 (1994) 1769–1778.CrossRefGoogle Scholar
- Dunn III, W.J. and Rogers, D., Genetic partial least squares in QSAR, In Devillers, J. (Ed.) Genetic al-gorithms in molecular modeling, Academic Press, London, 1996, pp. 109–130.Google Scholar
- Coats, E.A., The CoMFA steroids as a benchmark data set for development of 3D QSAR methods. In Kubinyi, H., Folkers, G. and Martin, Y.C. (Ed.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 199–214.Google Scholar
- Tropsha, A. and Cho, S.J., Cross-validated region selection for CoMFA studies. In Kubinyi, H., Folkers, G. and Martin, Y.C. (Eds.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, TheNetherlands, 1998, pp. 57–69.Google Scholar
- Cruciani, G., Clementi, S. and Pastor. M., GOLPE-Guided Region Selection, In Kubinyi, H., Folkers, G. and Martin, Y. (Ed.) 3D QSAR in drug design: Vol. 3, Kluwer Academic Publishers, Dordrecht, The Netherlands, 1998, pp. 71–86.Google Scholar
- Dunn III, W.J. and Rogers, D., Genetic partial least-squares in QSAR, In J. Devillers (Ed.) Genetic algorithms in molecular modeling, Academic Press, London, 1996, p. 109–30.Google Scholar
- Wikel, J.H. W.J. and Dow, E.R., The use of neural networks for variable selection in QSAR, Bioorg. Medic. Chem. Lett., 3 (1993) 645–651.Google Scholar
- Kubinyi, H., Variable selection in QSAR Studies: 1. An Evolutionary Algorithm, Quant. Struct.-Act. Relat., 13 (1994) 285–294.Google Scholar
- Kubinyi, H., Variable selection in QSAR studies: 2. A highly efficient combination of systematic search and evolution. Quant. Struct.-Act. Relat., 13 (1994) 393–401.Google Scholar
- Rogers, D. and Hopfinger, A.J., Application of genetic function approximation to quantitative struc-ture-activity relationships and quantitative structure-property relationships, J. Chem. Inf. Comput. Sci., 34 (1994) 854–866.Google Scholar
- Lingren, F., Geladi, P., Berglund, A., Sjostrum, M. and Wold, S., Interactive variable selection (IVS) for PLS: 2. Chemical applications, J. Chemometrics, 9 (1995) 331–342.Google Scholar
- Tetko, I.V., Villa, A. and Livingslonc, D.J., Neural-network studies: 2. Variable selection, J. Chem. Inf. Comput. Sci., 36 (1996) 794–803.CrossRefGoogle Scholar
- Baldovin, A., Wu, W., Centner, V., Jouanrimbaud, D., Massarl, D.L., Favretto, L. and Turello, A., Feature-selection for the discrimination between pollution types with partial least-squares modeling, Analyst, 121 (1996) 1603–1608.CrossRefGoogle Scholar
- Centner, V., Massart, D.L., Denoord, O.E., Dejong, S., Vandeginste, B.M. and Sterna, C., Elimination of uninformative variables for multivariate calibration, Anal. Chem., 68 (1996) 3851–3858.CrossRefGoogle Scholar
- Hasegawa, K., Miyashita, Y. and Funatsu, K., GA strategy for variable selection in QSAR studies: GA-basecl PLS analysis of calcium-channel antagonists, J. Chem. Inf. Comput. Sci., 37 (1997) 306–310.CrossRefGoogle Scholar
- Höltje, H.-D., Anzali, S., Dall, N. and Höltje, M., Binding Site Models, In Kubinyi, H. (Ed.) 3D QSAR in drug design: Theory, methods and applications, ESCOM, Leiden, The Netherlands, 1993, pp. 320–335.Google Scholar
- Vedani, A., Zhinden, P., Snyder, J.P. and Greenidge, P.A., Pseudoreceptor modeling: The construction of three-dimensional receptor surrogates, J. Am. Chem. Soc., 117 (1995) 4987–4994.CrossRefGoogle Scholar
- Topliss, J.G. and Edwards, R.P., Chance factors in studies of quantitative structure-activity relation-ships, J. Med. Chem., 22 (1979) 1238–1244.CrossRefGoogle Scholar
- Hoskuldsson, A., Quadratic PLS regression, J. Chemometrics, 6 (1992) 307–334.Google Scholar
- Benigni, R. and Guiliani, A., Analysis of distance matrices for studying data structures and separating classes. Quant. Struct.-Act. Relat., 12 (1993) 397–401.Google Scholar
- Kubinyi, H., QSAR: Hansch analysis and related approaches, VCH, Weinheim, Germany, 1993, Vol. 1, pp. 240.Google Scholar
- Martin, Y.C., Lin, C.T., Hetti, C. and DeLazzer, J., PLS analysis of distance matrices detects non-linear relationships between biological potency and molecular properties, J. Med. Chem., 38 (1995) 3009–3015.Google Scholar
- Livingstone, D. and Manallack, D.T., Statics using neural networks: Chance effects, J. Med. Chem., 36 (1993) 1295–1297.CrossRefGoogle Scholar
- Tetko, I.V., Livingstone, D.J. and Luik, A.I., Neural-network studies: 1. Comparison of overfitting and overtraining, J. Chem. Inf. Comput. Sci., 35 (1995) 826–833.CrossRefGoogle Scholar
- Devries, S. and Terbraak, C., Prediction error in partial least-squares regression: A critique on the deviation used in the unscramble, Chemometrics Intelligent Lab. systems, 30 (1995) 239–245.Google Scholar
- Jonathan, P., Mccarthy, W.V. and Roberts, A., Discriminant-analysis with singular covariance matrices: A method incorporating cross-validation and efficient randomized permutation tests, J. Chemometrics, 10 (1996) 189–213.3.0.CO%3B2-I'>CrossRefGoogle Scholar
- Kemsley, E.K., Discriminant-analysis of high-dimensional data: A comparison of principal com-ponents-analysis and partial least-squares data reduction methods, Chemometrics Intelligent Lab. Systems, 33 (1996) 47–61.Google Scholar
- Shuker, S., Hajduk, P., Meadows, R. and Fesik, S., Discovering high-affinity ligands for proteins: SAR by NMR, Science, 274 (1996) 1531–1534.CrossRefGoogle Scholar
- Sheridan, R.P. and Kearsley, S.K., Using a genetic algorithm to suggest combinatorial libraries, J. Chem. Inf. Comput. Sci., 35 (1995) 310–320.CrossRefGoogle Scholar