Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Charles Siegel is active.

Publication


Featured researches published by Charles Siegel.


knowledge discovery and data mining | 2018

Using Rule-Based Labels for Weak Supervised Learning: A ChemNet for Transferable Chemical Property Prediction

Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas

With access to large datasets, deep neural networks (DNN) have achieved human-level accuracy in image and speech recognition tasks. However, in chemistry data is inherently small and fragmented. In this work, we develop an approach of using rule-based knowledge for training ChemNet, a transferable and generalizable deep neural network for chemical property prediction that learns in a weak-supervised manner from large unlabeled chemical databases. When coupled with transfer learning approaches to predict other smaller datasets for chemical properties that it was not originally trained on, we show that ChemNets accuracy outperforms contemporary DNN models that were trained using conventional supervised learning. Furthermore, we demonstrate that the ChemNet pre-training approach is equally effective on both CNN (Chemception) and RNN (SMILES2vec) models, indicating that this approach is network architecture agnostic and is effective across multiple data modalities. Our results indicate a pre-trained ChemNet that incorporates chemistry domain knowledge and enables the development of generalizable neural networks for more accurate prediction of novel chemical properties.


high performance distributed computing | 2018

Desh: deep learning for system health prediction of lead times to failure in HPC

Anwesha Das; Frank Mueller; Charles Siegel; Abhinav Vishnu

Todays large-scale supercomputers encounter faults on a daily basis. Exascale systems are likely to experience even higher fault rates due to increased component count and density. Triggering resilience-mitigating techniques remains a challenge due to the absence of well defined failure indicators. System logs consist of unstructured text that obscures essential system health information contained within. In this context, efficient failure prediction via log mining can enable proactive recovery mechanisms to increase reliability. This work aims to predict node failures that occur in supercomputing systems via long short-term memory (LSTM) networks that exploit recurrent neural networks (RNNs). Our framework, Desh1 (Deep Learning for System Health), diagnoses and predicts failures with short lead times. Desh identifies failure indicators with enhanced training and classification for generic applicability to logs from operating systems and software components without the need to modify any of them. Desh uses a novel three-phase deep learning approach to (1) train to recognize chains of log events leading to a failure, (2) re-train chain recognition of events augmented with expected lead times to failure, and (3) predict lead times during testing/inference deployment to predict which specific node fails in how many minutes. Desh obtains as high as 3 minutes average lead time with no less than 85% recall and 83% accuracy to take proactive actions on the failing nodes, which could be used to migrate computation to healthy nodes.


arXiv: Distributed, Parallel, and Cluster Computing | 2016

Distributed TensorFlow with MPI

Abhinav Vishnu; Charles Siegel; Jeffrey A. Daily


arXiv: Machine Learning | 2017

Chemception: A Deep Neural Network with Minimal Chemistry Knowledge Matches the Performance of Expert-developed QSAR/QSPR Models.

Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas; Nathan A. Baker


workshop on applications of computer vision | 2018

How Much Chemistry Does a Deep Neural Network Need to Know to Make Accurate Predictions

Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas; Nathan A. Baker


Archive | 2017

ChemNet: A Transferable and Generalizable Deep Neural Network for Small-Molecule Property Prediction.

Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas


arXiv: Machine Learning | 2017

SMILES2Vec: An Interpretable General-Purpose Deep Neural Network for Predicting Chemical Properties.

Garrett B. Goh; Nathan O. Hodas; Charles Siegel; Abhinav Vishnu


arXiv: Learning | 2018

Multimodal Deep Neural Networks using Both Engineered and Learned Representations for Biodegradability Prediction.

Garrett B. Goh; Khushmeen Sakloth; Charles Siegel; Abhinav Vishnu; Jim Pfaendtner


arXiv: Distributed, Parallel, and Cluster Computing | 2018

GossipGraD: Scalable Deep Learning using Gossip Communication based Asynchronous Gradient Descent.

Jeff Daily; Abhinav Vishnu; Charles Siegel; Thomas Warfel; Vinay C. Amatya


EasyChair Preprints | 2018

Effective Machine Learning Based Format Selection and Performance Modeling for SpMV on GPUs

Israt Nisa; Charles Siegel; Aravind Sukumaran Rajam; Abhinav Vishnu; P. Sadayappan

Collaboration


Dive into the Charles Siegel's collaboration.

Top Co-Authors

Avatar

Abhinav Vishnu

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Garrett B. Goh

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Nathan O. Hodas

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Jeff Daily

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Nathan A. Baker

Pacific Northwest National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Anwesha Das

North Carolina State University

View shared research outputs
Top Co-Authors

Avatar

Frank Mueller

North Carolina State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jeffrey A. Daily

Pacific Northwest National Laboratory

View shared research outputs
Researchain Logo
Decentralizing Knowledge