Charles Siegel
Cray
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Charles Siegel.
knowledge discovery and data mining | 2018
Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas
With access to large datasets, deep neural networks (DNN) have achieved human-level accuracy in image and speech recognition tasks. However, in chemistry data is inherently small and fragmented. In this work, we develop an approach of using rule-based knowledge for training ChemNet, a transferable and generalizable deep neural network for chemical property prediction that learns in a weak-supervised manner from large unlabeled chemical databases. When coupled with transfer learning approaches to predict other smaller datasets for chemical properties that it was not originally trained on, we show that ChemNets accuracy outperforms contemporary DNN models that were trained using conventional supervised learning. Furthermore, we demonstrate that the ChemNet pre-training approach is equally effective on both CNN (Chemception) and RNN (SMILES2vec) models, indicating that this approach is network architecture agnostic and is effective across multiple data modalities. Our results indicate a pre-trained ChemNet that incorporates chemistry domain knowledge and enables the development of generalizable neural networks for more accurate prediction of novel chemical properties.
high performance distributed computing | 2018
Anwesha Das; Frank Mueller; Charles Siegel; Abhinav Vishnu
Todays large-scale supercomputers encounter faults on a daily basis. Exascale systems are likely to experience even higher fault rates due to increased component count and density. Triggering resilience-mitigating techniques remains a challenge due to the absence of well defined failure indicators. System logs consist of unstructured text that obscures essential system health information contained within. In this context, efficient failure prediction via log mining can enable proactive recovery mechanisms to increase reliability. This work aims to predict node failures that occur in supercomputing systems via long short-term memory (LSTM) networks that exploit recurrent neural networks (RNNs). Our framework, Desh1 (Deep Learning for System Health), diagnoses and predicts failures with short lead times. Desh identifies failure indicators with enhanced training and classification for generic applicability to logs from operating systems and software components without the need to modify any of them. Desh uses a novel three-phase deep learning approach to (1) train to recognize chains of log events leading to a failure, (2) re-train chain recognition of events augmented with expected lead times to failure, and (3) predict lead times during testing/inference deployment to predict which specific node fails in how many minutes. Desh obtains as high as 3 minutes average lead time with no less than 85% recall and 83% accuracy to take proactive actions on the failing nodes, which could be used to migrate computation to healthy nodes.
arXiv: Distributed, Parallel, and Cluster Computing | 2016
Abhinav Vishnu; Charles Siegel; Jeffrey A. Daily
arXiv: Machine Learning | 2017
Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas; Nathan A. Baker
workshop on applications of computer vision | 2018
Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas; Nathan A. Baker
Archive | 2017
Garrett B. Goh; Charles Siegel; Abhinav Vishnu; Nathan O. Hodas
arXiv: Machine Learning | 2017
Garrett B. Goh; Nathan O. Hodas; Charles Siegel; Abhinav Vishnu
arXiv: Learning | 2018
Garrett B. Goh; Khushmeen Sakloth; Charles Siegel; Abhinav Vishnu; Jim Pfaendtner
arXiv: Distributed, Parallel, and Cluster Computing | 2018
Jeff Daily; Abhinav Vishnu; Charles Siegel; Thomas Warfel; Vinay C. Amatya
EasyChair Preprints | 2018
Israt Nisa; Charles Siegel; Aravind Sukumaran Rajam; Abhinav Vishnu; P. Sadayappan