首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Non-specific chemical modification of protein thiol groups continues to be a significant source of false positive hits from high-throughput screening campaigns and can even plague certain protein targets and chemical series well into lead optimization. While experimental tools exist to assess the risk and promiscuity associated with the chemical reactivity of existing compounds, computational tools are desired that can reliably identify substructures that are associated with chemical reactivity to aid in triage of HTS hit lists, external compound purchases, and library design. Here we describe a Bayesian classification model derived from more than 8,800 compounds that have been experimentally assessed for their potential to covalently modify protein targets. The resulting model can be implemented in the large-scale assessment of compound libraries for purchase or design. In addition, the individual substructures identified as highly reactive in the model can be used as look-up tables to guide chemists during hit-to-lead and lead optimization campaigns.  相似文献   

2.
Advances in high throughput screening technologies have led to the identification of many small molecules, "hits", with activities toward the target of interest. And, as the screening technologies become faster and more robust, the rate at which the molecules are identified continues to increase. This evolution of high throughput screening technologies has generated a significant strain on the laboratories involved with the downstream profiling of these hits using cell-based assays. The CellCard System, by enabling multiple targets and/or cell lines to be assayed simultaneously within a single well, provides a platform on which selectivity screening can be quickly and robustly performed. Here we describe two case studies using the beta-lactamase and beta-galactosidase reporter gene systems to characterize G protein-coupled receptor agonist activity. Using these examples we demonstrate how the implementation of this technology enables assay miniaturization without micro-fluidic devices as well as how the inclusion of intra-well controls can provide a means of data quality assessment within each well.  相似文献   

3.
Drug discovery is a complicated process that involves multiple synthetic chemistry tasks. Among them, lead generation and optimization is the core business in the discovery research. During the stage of lead generation, a large library of many thousands individual compounds will be screened against a biological target to identify a set of hits that showed desirable activity. Once a hit has been identified, analog synthesis and development of SAR around this hit and establishment of relationsh…  相似文献   

4.
Small molecule aggregators non‐specifically inhibit multiple unrelated proteins, rendering them therapeutically useless. They frequently appear as false hits and thus need to be eliminated in high‐throughput screening campaigns. Computational methods have been explored for identifying aggregators, which have not been tested in screening large compound libraries. We used 1319 aggregators and 128,325 non‐aggregators to develop a support vector machines (SVM) aggregator identification model, which was tested by four methods. The first is five fold cross‐validation, which showed comparable aggregator and significantly improved non‐aggregator identification rates against earlier studies. The second is the independent test of 17 aggregators discovered independently from the training aggregators, 71% of which were correctly identified. The third is retrospective screening of 13M PUBCHEM and 168K MDDR compounds, which predicted 97.9% and 98.7% of the PUBCHEM and MDDR compounds as non‐aggregators. The fourth is retrospective screening of 5527 MDDR compounds similar to the known aggregators, 1.14% of which were predicted as aggregators. SVM showed slightly better overall performance against two other machine learning methods based on five fold cross‐validation studies of the same settings. Molecular features of aggregation, extracted by a feature selection method, are consistent with published profiles. SVM showed substantial capability in identifying aggregators from large libraries at low false‐hit rates. © 2009 Wiley Periodicals, Inc.J Comput Chem, 2010  相似文献   

5.
6.
Ideally, a team of biologists, medicinal chemists and information specialists will evaluate the hits from high throughput screening. In practice, it often falls to nonmedicinal chemists to make the initial evaluation of HTS hits. Chemical genetics and high content screening both rely on screening in cells or animals where the biological target may not be known. There is a need to place active compounds into a context to suggest potential biological mechanisms. Our idea is to build an operating environment to help the biologist make the initial evaluation of HTS data. To this end the operating environment provides viewing of compound structure files, computation of basic biologically relevant chemical properties and searching against biologically annotated chemical structure databases. The benefit is to help the nonmedicinal chemist, biologist and statistician put compounds into a potentially informative biological context. Although there are several similar public and private programs used in the pharmaceutical industry to help evaluate hits, these programs are often built for computational chemists. Our program is designed for use by biologists and statisticians.  相似文献   

7.
Integration of flexible data-analysis tools with cheminformatics methods is a prerequisite for successful identification and validation of “hits” in high-throughput screening (HTS) campaigns. We have designed, developed, and implemented a suite of robust yet flexible cheminformatics tools to support HTS activities at the Broad Institute, three of which are described herein. The “hit-calling” tool allows a researcher to set a hit threshold that can be varied during downstream analysis. The results from the hit-calling exercise are reported to a database for record keeping and further data analysis. The “cherry-picking” tool enables creation of an optimized list of hits for confirmatory and follow-up assays from an HTS hit list. This tool allows filtering by computed chemical property and by substructure. In addition, similarity searches can be performed on hits of interest and sets of related compounds can be selected. The third tool, an “S/SAR viewer,” has been designed specifically for the Broad Institute’s diversity-oriented synthesis (DOS) collection. The compounds in this collection are rich in chiral centers and the full complement of all possible stereoisomers of a given compound are present in the collection. The S/SAR viewer allows rapid identification of both structure/activity relationships and stereo-structure/activity relationships present in HTS data from the DOS collection. Together, these tools enable the prioritization and analysis of hits from diverse compound collections, and enable informed decisions for follow-up biology and chemistry efforts.  相似文献   

8.
9.
The “Cheminformatics aspects of high throughput screening (HTS): from robots to models” symposium was part of the computers in chemistry technical program at the American Chemical Society National Meeting in Denver, Colorado during the fall of 2011. This symposium brought together researchers from high throughput screening centers and molecular modelers from academia and industry to discuss the integration of currently available high throughput screening data and assays with computational analysis. The topics discussed at this symposium covered the data-infrastructure at various academic, hospital, and National Institutes of Health-funded high throughput screening centers, the cheminformatics and molecular modeling methods used in real world examples to guide screening and hit-finding, and how academic and non-profit organizations can benefit from current high throughput screening cheminformatics resources. Specifically, this article also covers the remarks and discussions in the open panel discussion of the symposium and summarizes the following talks on “Accurate Kinase virtual screening: biochemical, cellular and selectivity”, “Selective, privileged and promiscuous chemical patterns in high-throughput screening” and “Visualizing and exploring relationships among HTS hits using network graphs”.  相似文献   

10.
From a medicinal chemistry point of view, one of the primary goals of high throughput screening (HTS) hit list assessment is the identification of chemotypes with an informative structure-activity relationship (SAR). Such chemotypes may enable optimization of the primary potency, as well as selectivity and phamacokinetic properties. A common way to prioritize them is molecular clustering of the hits. Typical clustering techniques, however, rely on a general notion of chemical similarity or standard rules of scaffold decomposition and are thus insensitive to molecular features that are enriched in biologically active compounds. This hinders SAR analysis, because compounds sharing the same pharmacophore might not end up in the same cluster and thus are not directly compared to each other by the medicinal chemist. Similarly, common chemotypes that are not related to activity may contaminate clusters, distracting from important chemical motifs. We combined molecular similarity and Bayesian models and introduce (I) a robust, activity-aware clustering approach and (II) a feature mapping method for the elucidation of distinct SAR determinants in polypharmacologic compounds. We evaluated the method on 462 dose-response assays from the Pubchem Bioassay repository. Activity-aware clustering grouped compounds sharing molecular cores that were specific for the target or pathway at hand, rather than grouping inactive scaffolds commonly found in compound series. Many of these core structures we also found in literature that discussed SARs of the respective targets. A numerical comparison of cores allowed for identification of the structural prerequisites for polypharmacology, i.e., distinct bioactive regions within a single compound, and pointed toward selectivity-conferring medchem strategies. The method presented here is generally applicable to any type of activity data and may help bridge the gap between hit list assessment and designing a medchem strategy.  相似文献   

11.
12.
In recent few years combinatorial methodology has been extensively used in material science research. Based on the desired properties of materials, various high throughput synthesizing and screening technologies were developed. These high throughput technologies can increase our speed to more than hundred folds for finding and optimizing materials. One of the most active areas is catalysis. Scientists are developing novel high throughput technologies to screen catalyst libraries to find and optimize new catalysts for chemical industry. In this area die key is combinatorial catalytic reactor design, catalyst library synthesis, and product detection. Systematic technologies for catalyst library synthesis and characterization were developed in our laboratory. In this work, catalyst in situ synthesis, parallel reactor design, and detection methods will be introduced. Combining with the powerful combinatorial methodology, good chemistry design will make our work even more efficient. Hence, as an example of combining combinatorial technologies with chemistry design, a successful catalyst design is also introduced.  相似文献   

13.
Fragment-based screening is an emerging technology which is used as an alternative to high-throughput screening (HTS), and often in parallel. Fragment screening focuses on very small compounds. Because of their small size and simplicity, fragments exhibit a low to medium binding affinity (mM to μM) and must therefore be screened at high concentration in order to detect binding events. Since some issues are associated with high-concentration screening in biochemical assays, biophysical methods are generally employed in fragment screening campaigns. Moreover, these techniques are very sensitive and some of them can give precise information about the binding mode of fragments, which facilitates the mandatory hit-to-lead optimization. One of the main advantages of fragment-based screening is that fragment hits generally exhibit a strong binding with respect to their size, and their subsequent optimization should lead to compounds with better pharmacokinetic properties compared to molecules evolved from HTS hits. In other words, fragments are interesting starting points for drug discovery projects. Besides, the chemical space of low-complexity compounds is very limited in comparison to that of drug-like molecules, and thus easier to explore with a screening library of limited size. Furthermore, the "combinatorial explosion" effect ensures that the resulting combinations of interlinked binding fragments may cover a significant part of "drug-like" chemical space. In parallel to experimental screening, virtual screening techniques, dedicated to fragments or wider compounds, are gaining momentum in order to further reduce the number of compounds to test. This article is a review of the latest news in both experimental and in silico virtual screening in the fragment-based discovery field. Given the specificity of this journal, special attention will be given to fragment library design.  相似文献   

14.
15.
The identification of promising hits and the generation of high quality leads are crucial steps in the early stages of drug discovery projects. The definition and assessment of both chemical and biological space have revitalized the screening process model and emphasized the importance of exploring the intrinsic complementary nature of classical and modern methods in drug research. In this context, the widespread use of combinatorial chemistry and sophisticated screening methods for the discovery of lead compounds has created a large demand for small organic molecules that act on specific drug targets. Modern drug discovery involves the employment of a wide variety of technologies and expertise in multidisciplinary research teams. The synergistic effects between experimental and computational approaches on the selection and optimization of bioactive compounds emphasize the importance of the integration of advanced technologies in drug discovery programs. These technologies (VS, HTS, SBDD, LBDD, QSAR, and so on) are complementary in the sense that they have mutual goals, thereby the combination of both empirical and in silico efforts is feasible at many different levels of lead optimization and new chemical entity (NCE) discovery. This paper provides a brief perspective on the evolution and use of key drug design technologies, highlighting opportunities and challenges.  相似文献   

16.
The aim of this tutorial review is to introduce the reader to the concept, synthesis and application of natural product-inspired compound collections as an important field in chemical biology. This review will discuss how potentially interesting scaffolds can be identified (structural classification of natural products), synthesized in an appropriate manner (including stereoselective transformations for solid phase-bound compounds) and tested in biological assays (cell-based screening as well as biochemical in vitro assays). These approaches will provide the opportunity to identify new and interesting compounds as well as new targets for chemical biology and medicinal chemistry research.  相似文献   

17.
The process of Drug Discovery is a complex and high risk endeavor that requires focused attention on experimental hypotheses, the application of diverse sets of technologies and data to facilitate high quality decision-making. All is aimed at enhancing the quality of the chemical development candidate(s) through clinical evaluation and into the market. In support of the lead generation and optimization phases of this endeavor, high throughput technologies such as combinatorial/high throughput synthesis and high throughput and ultra-high throughput screening, have allowed the rapid analysis and generation of large number of compounds and data. Today, for every analog synthesized 100 or more data points can be collected and captured in various centralized databases. The analysis of thousands of compounds can very quickly become a daunting task. In this article we present the process we have developed for both analyzing and prioritizing large sets of data starting from diversity and focused uHTS in support of lead generation and secondary screens supporting lead optimization. We will describe how we use informatics and computational chemistry to focus our efforts on asking relevant questions about the desired attributes of a specific library, and subsequently in guiding the generation of more information-rich sets of analogs in support of both processes.  相似文献   

18.
Summary Structure-based screening using fully flexible docking is still too slow for large molecular libraries. High quality docking of a million molecule library can take days even on a cluster with hundreds of CPUs. This performance issue prohibits the use of fully flexible docking in the design of large combinatorial libraries. We have developed a fast structure-based screening method, which utilizes docking of a limited number of compounds to build a 2D QSAR model used to rapidly score the rest of the database. We compare here a model based on radial basis functions and a Bayesian categorization model. The number of compounds that need to be actually docked depends on the number of docking hits found. In our case studies reasonable quality models are built after docking of the number of molecules containing 50 docking hits. The rest of the library is screened by the QSAR model. Optionally a fraction of the QSAR-prioritized library can be docked in order to find the true docking hits. The quality of the model only depends on the training set size – not on the size of the library to be screened. Therefore, for larger libraries the method yields higher gain in speed no change in performance. Prioritizing a large library with these models provides a significant enrichment with docking hits: it attains the values of 13 and 35 at the beginning of the score-sorted libraries in our two case studies: screening of the NCI collection and a combinatorial libraries on CDK2 kinase structure. With such enrichments, only a fraction of the database must actually be docked to find many of the true hits. The throughput of the method allows its use in screening of large compound collections and in the design of large combinatorial libraries. The strategy proposed has an important effect on efficiency but does not affect retrieval of actives, the latter being determined by the quality of the docking method itself. Electronic supplementary material is available at http://dx.doi.org/10.1007/s10822-005-9002-6.  相似文献   

19.
Increasingly, chemical libraries are being produced which are focused on a biological target or group of related targets, rather than simply being constructed in a combinatorial fashion. A screening collection compiled from such libraries will contain multiple analogues of a number of discrete series of compounds. The question arises as to how many analogues are necessary to represent each series in order to ensure that an active series will be identified. Based on a simple probabilistic argument and supported by in-house screening data, guidelines are given for the number of compounds necessary to achieve a "hit", or series of hits, at various levels of certainty. Obtaining more than one hit from the same series is useful since this gives early acquisition of SAR (structure-activity relationship) and confirms a hit is not a singleton. We show that screening collections composed of only small numbers of analogues of each series are sub-optimal for SAR acquisition. Based on these studies, we recommend a minimum series size of about 200 compounds. This gives a high probability of confirmatory SAR (i.e. at least two hits from the same series). More substantial early SAR (at least 5 hits from the same series) can be gained by using series of about 650 compounds each. With this level of information being generated, more accurate assessment of the likely success of the series in hit-to-lead and later stage development becomes possible.  相似文献   

20.
Molecular similarity methods for ligand-based virtual screening (VS) generally do not take compound potency as a variable or search parameter into account. We have incorporated a logarithmic potency scaling function into two conceptually distinct VS algorithms to account for relative compound potency during search calculations. A high-throughput screening (HTS) data set containing cathepsin B inhibitors was analyzed to evaluate the effects of potency scaling. Sets of template compounds were randomly selected from the HTS data and used to search for hits having varying potency levels in the presence or absence of potency scaling. Enrichment of potent compounds in small subsets of the HTS data set was observed as a consequence of potency scaling. In part, observed enrichments could be rationalized as a result of recentering chemical reference space on a subspace populated by potent compounds. Our findings suggest that VS calculations using multiple reference compounds can be directed toward the preferential detection of potent database hits by scaling compound contributions according to potency differences.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号