Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Florian Nigsch is active.

Publication


Featured researches published by Florian Nigsch.


Journal of Computer-aided Molecular Design | 2011

Online chemical modeling environment (OCHEM): web platform for data storage, model development and publishing of chemical information.

Iurii Sushko; Sergii Novotarskyi; Robert Körner; Anil Kumar Pandey; Matthias Rupp; Wolfram Teetz; Stefan Brandmaier; Ahmed Abdelaziz; Volodymyr V. Prokopenko; Vsevolod Yu. Tanchuk; Roberto Todeschini; Alexandre Varnek; Gilles Marcou; Peter Ertl; Vladimir Potemkin; Maria A. Grishina; Johann Gasteiger; Christof H. Schwab; I. I. Baskin; V. A. Palyulin; E. V. Radchenko; William J. Welsh; Vladyslav Kholodovych; Dmitriy Chekmarev; Artem Cherkasov; João Aires-de-Sousa; Qingyou Zhang; Andreas Bender; Florian Nigsch; Luc Patiny

The Online Chemical Modeling Environment is a web-based platform that aims to automate and simplify the typical steps required for QSAR modeling. The platform consists of two major subsystems: the database of experimental measurements and the modeling framework. A user-contributed database contains a set of tools for easy input, search and modification of thousands of records. The OCHEM database is based on the wiki principle and focuses primarily on the quality and verifiability of the data. The database is tightly integrated with the modeling framework, which supports all the steps required to create a predictive model: data search, calculation and selection of a vast variety of molecular descriptors, application of machine learning methods, validation, analysis of the model and assessment of the applicability domain. As compared to other similar systems, OCHEM is not intended to re-implement the existing tools or models but rather to invite the original authors to contribute their results, make them publicly available, share them with other users and to become members of the growing research community. Our intention is to make OCHEM a widely used platform to perform the QSPR/QSAR studies online and share it with other users on the Web. The ultimate goal of OCHEM is collecting all possible chemoinformatics tools within one simple, reliable and user-friendly resource. The OCHEM is free for web users and it is available online at http://www.ochem.eu.


Journal of Chemical Information and Modeling | 2008

Why are some properties more difficult to predict than others? A study of QSPR models of solubility, melting point, and Log P

Laura D. Hughes; David S. Palmer; Florian Nigsch; John B. O. Mitchell

This paper attempts to elucidate differences in QSPR models of aqueous solubility (Log S), melting point (Tm), and octanol-water partition coefficient (Log P), three properties of pharmaceutical interest. For all three properties, Support Vector Machine models using 2D and 3D descriptors calculated in the Molecular Operating Environment were the best models. Octanol-water partition coefficient was the easiest property to predict, as indicated by the RMSE of the external test set and the coefficient of determination (RMSE = 0.73, r2 = 0.87). Melting point prediction, on the other hand, was the most difficult (RMSE = 52.8 degrees C, r2 = 0.46), and Log S statistics were intermediate between melting point and Log P prediction (RMSE = 0.900, r2 = 0.79). The data imply that for all three properties the lack of measured values at the extremes is a significant source of error. This source, however, does not entirely explain the poor melting point prediction, and we suggest that deficiencies in descriptors used in melting point prediction contribute significantly to the prediction errors.


Journal of Chemical Information and Modeling | 2006

Melting point prediction employing k-nearest neighbor algorithms and genetic parameter optimization.

Florian Nigsch; Andreas Bender; Bernd van Buuren; Jos Tissen; Eduard A. Nigsch; John B. O. Mitchell

We have applied the k-nearest neighbor (kNN) modeling technique to the prediction of melting points. A data set of 4119 diverse organic molecules (data set 1) and an additional set of 277 drugs (data set 2) were used to compare performance in different regions of chemical space, and we investigated the influence of the number of nearest neighbors using different types of molecular descriptors. To compute the prediction on the basis of the melting temperatures of the nearest neighbors, we used four different methods (arithmetic and geometric average, inverse distance weighting, and exponential weighting), of which the exponential weighting scheme yielded the best results. We assessed our model via a 25-fold Monte Carlo cross-validation (with approximately 30% of the total data as a test set) and optimized it using a genetic algorithm. Predictions for drugs based on drugs (separate training and test sets each taken from data set 2) were found to be considerably better [root-mean-squared error (RMSE)=46.3 degrees C, r2=0.30] than those based on nondrugs (prediction of data set 2 based on the training set from data set 1, RMSE=50.3 degrees C, r2=0.20). The optimized model yields an average RMSE as low as 46.2 degrees C (r2=0.49) for data set 1, and an average RMSE of 42.2 degrees C (r2=0.42) for data set 2. It is shown that the kNN method inherently introduces a systematic error in melting point prediction. Much of the remaining error can be attributed to the lack of information about interactions in the liquid state, which are not well-captured by molecular descriptors.


Journal of Chemical Information and Modeling | 2008

Ligand-target prediction using Winnow and naive Bayesian algorithms and the implications of overall performance statistics.

Florian Nigsch; Andreas Bender; Jeremy L. Jenkins; John B. O. Mitchell

We compared two algorithms for ligand-target prediction, namely, the Laplacian-modified Bayesian classifier and the Winnow algorithm. A dataset derived from the WOMBAT database, spanning 20 pharmaceutically relevant activity classes with 13 000 compounds, was used for performance assessment in 24 different experiments, each of which was assessed using a 15-fold Monte Carlo cross-validation. Compounds were described by different circular fingerprints, ECFP_4 and MOLPRINT 2D. A detailed analysis of the resulting approximately 2.4 million predictions led to very similar measures for overall accuracy for both classifiers, whereas we observed significant differences for individual activity classes. Moreover, we analyzed our data with respect to the numbers of compounds which are exclusively retrieved by either of the algorithmsbut never by the otheror by neither of them. This provided detailed information that can never be obtained by considering the overall performance statistics alone.


Microbiological Research | 2014

High-resolution chemical dissection of a model eukaryote reveals targets, pathways and gene functions

Dominic Hoepfner; Stephen B. Helliwell; Heather Sadlish; Sven Schuierer; Ireos Filipuzzi; Sophie Brachat; Bhupinder Bhullar; Uwe Plikat; Yann Abraham; Marc Altorfer; Thomas Aust; Lukas Baeriswyl; Raffaele Cerino; Lena Chang; David Estoppey; Juerg Eichenberger; Mathias Frederiksen; Nicole Hartmann; Annika Hohendahl; Britta Knapp; Philipp Krastel; Nicolas Melin; Florian Nigsch; Virginie Petitjean; Frank Petersen; Ralph Riedl; Esther K. Schmitt; Frank Staedtler; Christian Studer; John A. Tallarico

Due to evolutionary conservation of biology, experimental knowledge captured from genetic studies in eukaryotic model organisms provides insight into human cellular pathways and ultimately physiology. Yeast chemogenomic profiling is a powerful approach for annotating cellular responses to small molecules. Using an optimized platform, we provide the relative sensitivities of the heterozygous and homozygous deletion collections for nearly 1800 biologically active compounds. The data quality enables unique insights into pathways that are sensitive and resistant to a given perturbation, as demonstrated with both known and novel compounds. We present examples of novel compounds that inhibit the therapeutically relevant fatty acid synthase and desaturase (Fas1p and Ole1p), and demonstrate how the individual profiles facilitate hypothesis-driven experiments to delineate compound mechanism of action. Importantly, the scale and diversity of tested compounds yields a dataset where the number of modulated pathways approaches saturation. This resource can be used to map novel biological connections, and also identify functions for unannotated genes. We validated hypotheses generated by global two-way hierarchical clustering of profiles for (i) novel compounds with a similar mechanism of action acting upon microtubules or vacuolar ATPases, and (ii) an un-annotated ORF, YIL060w, that plays a role in respiration in the mitochondria. Finally, we identify and characterize background mutations in the widely used yeast deletion collection which should improve the interpretation of past and future screens throughout the community. This comprehensive resource of cellular responses enables the expansion of our understanding of eukaryotic pathway biology.


Expert Opinion on Drug Metabolism & Toxicology | 2009

Computational toxicology: an overview of the sources of data and of modelling methods

Florian Nigsch; Nj Maximilan Macaluso; John B. O. Mitchell; Donatas Zmuidinavicius

Background: Toxicology has the goal of ensuring the safety of humans, animals and the environment. Computational toxicology is an area of active development and great potential. There are tangible reasons for the emerging interest in this discipline from academia, industry, regulatory bodies and governments. Results: Pharmaceuticals, personal health care products, nutritional ingredients and products of the chemical industries are all potential hazards and need to be assessed. Toxicological tests for these products are costly, frequently use laboratory animals and are time-consuming. This delays end-user access to improved products or, conversely, the timely withdrawal of dangerous substances from the market. The aim of computational toxicology is to accelerate the assessment of potentially dangerous substances through in silico models. Conclusions: In this review, we provide an overview of the development of models for computational toxicology. Addressing the significant divide between the experimental and computational worlds–believed to be a prime hindrance to computational toxicology–we briefly consider the fundamental issue of toxicological data and the assays they stem from. Different kinds of models that can be built using such data are presented: computational filters, models for specific toxicological endpoints and tools for the generation of testable hypotheses.


Future Medicinal Chemistry | 2011

Recent trends and observations in the design of high-quality screening collections

Steffen Renner; Maxim Popov; Ansgar Schuffenhauer; Hans-Joerg Roth; Werner Breitenstein; Andreas Marzinzik; Ian Lewis; Philipp Krastel; Florian Nigsch; Jeremy L. Jenkins; Edgar Jacoby

The design of a high-quality screening collection is of utmost importance for the early drug-discovery process and provides, in combination with high-quality assay systems, the foundation of future discoveries. Herein, we review recent trends and observations to successfully expand the access to bioactive chemical space, including the feedback from hit assessment interviews of high-throughput screening campaigns; recent successes with chemogenomics target family approaches, the identification of new relevant target/domain families, diversity-oriented synthesis and new emerging compound classes, and non-classical approaches, such as fragment-based screening and DNA-encoded chemical libraries. The role of in silico library design approaches are emphasized.


Journal of Cheminformatics | 2012

Predicting the mechanism of phospholipidosis

Robert Lowe; Hamse Y. Mussa; Florian Nigsch; Robert C. Glen; John B. O. Mitchell

The mechanism of phospholipidosis is still not well understood. Numerous different mechanisms have been proposed, varying from direct inhibition of the breakdown of phospholipids to the binding of a drug compound to the phospholipid, preventing breakdown. We have used a probabilistic method, the Parzen-Rosenblatt Window approach, to build a model from the ChEMBL dataset which can predict from a compounds structure both its primary pharmaceutical target and other targets with which it forms off-target, usually weaker, interactions. Using a small dataset of 182 phospholipidosis-inducing and non-inducing compounds, we predict their off-target activity against targets which could relate to phospholipidosis as a side-effect of a drug. We link these targets to specific mechanisms of inducing this lysosomal build-up of phospholipids in cells. Thus, we show that the induction of phospholipidosis is likely to occur by separate mechanisms when triggered by different cationic amphiphilic drugs. We find that both inhibition of phospholipase activity and enhanced cholesterol biosynthesis are likely to be important mechanisms. Furthermore, we provide evidence suggesting four specific protein targets. Sphingomyelin phosphodiesterase, phospholipase A2 and lysosomal phospholipase A1 are shown to be likely targets for the induction of phospholipidosis by inhibition of phospholipase activity, while lanosterol synthase is predicted to be associated with phospholipidosis being induced by enhanced cholesterol biosynthesis. This analysis provides the impetus for further experimental tests of these hypotheses.


Expert Opinion on Drug Metabolism & Toxicology | 2011

Computational methods for early predictive safety assessment from biological and chemical data

Florian Nigsch; Eugen Lounkine; Patrick McCarren; Ben Cornett; Meir Glick; Kamal Azzaoui; Laszlo Urban; Arne Müller; Florian Hahne; David J. Heard; Jeremy L. Jenkins

Introduction: The goal of early predictive safety assessment (PSA) is to keep compounds with detectable liabilities from progressing further in the pipeline. Such compounds jeopardize the core of pharmaceutical research and development and limit the timely delivery of innovative therapeutics to the patient. Computational methods are increasingly used to help understand observed data, generate new testable hypotheses of relevance to safety pharmacology, and supplement and replace costly and time-consuming experimental procedures. Areas covered: The authors survey methods operating on different scales of both physical extension and complexity. After discussing methods used to predict liabilities associated with structures of individual compounds, the article reviews the use of adverse event data and safety profiling panels. Finally, the authors examine the complexities of toxicology data from animal experiments and how these data can be mined. Expert opinion: A significant obstacle for data-driven safety assessment is the absence of integrated data sets due to a lack of sharing of data and of using standard ontologies for data relevant to safety assessment. Informed decisions to derive focused sets of compounds can help to avoid compound liabilities in screening campaigns, and improved hit assessment of such campaigns can benefit the early termination of undesirable compounds.


Chemistry Central Journal | 2008

Simultaneous feature selection and parameter optimisation using an artificial ant colony: case study of melting point prediction

Noel M. O'Boyle; David S. Palmer; Florian Nigsch; John B. O. Mitchell

BackgroundWe present a novel feature selection algorithm, Winnowing Artificial Ant Colony (WAAC), that performs simultaneous feature selection and model parameter optimisation for the development of predictive quantitative structure-property relationship (QSPR) models. The WAAC algorithm is an extension of the modified ant colony algorithm of Shen et al. (J Chem Inf Model 2005, 45: 1024–1029). We test the ability of the algorithm to develop a predictive partial least squares model for the Karthikeyan dataset (J Chem Inf Model 2005, 45: 581–590) of melting point values. We also test its ability to perform feature selection on a support vector machine model for the same dataset.ResultsStarting from an initial set of 203 descriptors, the WAAC algorithm selected a PLS model with 68 descriptors which has an RMSE on an external test set of 46.6°C and R2 of 0.51. The number of components chosen for the model was 49, which was close to optimal for this feature selection. The selected SVM model has 28 descriptors (cost of 5, ε of 0.21) and an RMSE of 45.1°C and R2 of 0.54. This model outperforms a kNN model (RMSE of 48.3°C, R2 of 0.47) for the same data and has similar performance to a Random Forest model (RMSE of 44.5°C, R2 of 0.55). However it is much less prone to bias at the extremes of the range of melting points as shown by the slope of the line through the residuals: -0.43 for WAAC/SVM, -0.53 for Random Forest.ConclusionWith a careful choice of objective function, the WAAC algorithm can be used to optimise machine learning and regression models that suffer from overfitting. Where model parameters also need to be tuned, as is the case with support vector machine and partial least squares models, it can optimise these simultaneously. The moving probabilities used by the algorithm are easily interpreted in terms of the best and current models of the ants, and the winnowing procedure promotes the removal of irrelevant descriptors.

Collaboration


Dive into the Florian Nigsch's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge