Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Søren Kaae Sønderby is active.

Publication


Featured researches published by Søren Kaae Sønderby.


arXiv: Quantitative Methods | 2015

Convolutional LSTM Networks for Subcellular Localization of Proteins

Søren Kaae Sønderby; Casper Kaae Sønderby; Henrik Nielsen; Ole Winther

Machine learning is widely used to analyze biological sequence data. Non-sequential models such as SVMs or feed-forward neural networks are often used although they have no natural way of handling sequences of varying length. Recurrent neural networks such as the long short term memory (LSTM) model on the other hand are designed to handle sequences. In this study we demonstrate that LSTM networks predict the subcellular location of proteins given only the protein sequence with high accuracy (0.902) outperforming current state of the art algorithms. We further improve the performance by introducing convolutional filters and experiment with an attention mechanism which lets the LSTM focus on specific parts of the protein. Lastly we introduce new visualizations of both the convolutional filters and the attention mechanisms and show how they can be used to extract biological relevant knowledge from the LSTM networks.


Bioinformatics | 2017

DeepLoc: prediction of protein subcellular localization using deep learning

José Juan Almagro Armenteros; Casper Kaae Sønderby; Søren Kaae Sønderby; Henrik Nielsen; Ole Winther

Motivation The prediction of eukaryotic protein subcellular localization is a well‐studied topic in bioinformatics due to its relevance in proteomics research. Many machine learning methods have been successfully applied in this task, but in most of them, predictions rely on annotation of homologues from knowledge databases. For novel proteins where no annotated homologues exist, and for predicting the effects of sequence variants, it is desirable to have methods for predicting protein properties from sequence information only. Results Here, we present a prediction algorithm using deep neural networks to predict protein subcellular localization relying only on sequence information. At its core, the prediction model uses a recurrent neural network that processes the entire protein sequence and an attention mechanism identifying protein regions important for the subcellular localization. The model was trained and tested on a protein dataset extracted from one of the latest UniProt releases, in which experimentally annotated proteins follow more stringent criteria than previously. We demonstrate that our model achieves a good accuracy (78% for 10 categories; 92% for membrane‐bound or soluble), outperforming current state‐of‐the‐art algorithms, including those relying on homology information. Availability and implementation The method is available as a web server at http://www.cbs.dtu.dk/services/DeepLoc. Example code is available at https://github.com/JJAlmagro/subcellular_localization. The dataset is available at http://www.cbs.dtu.dk/services/DeepLoc/data.php. Contact [email protected]


Bioinformatics | 2017

An introduction to deep learning on biological sequence data: examples and solutions

Vanessa Isabell Jurtz; Alexander Rosenberg Johansen; Morten Nielsen; José Juan Almagro Armenteros; Henrik Nielsen; Casper Kaae Sønderby; Ole Winther; Søren Kaae Sønderby

Motivation Deep neural network architectures such as convolutional and long short‐term memory networks have become increasingly popular as machine learning tools during the recent years. The availability of greater computational resources, more data, new algorithms for training deep models and easy to use libraries for implementation and training of neural networks are the drivers of this development. The use of deep learning has been especially successful in image recognition; and the development of tools, applications and code examples are in most cases centered within this field rather than within biology. Results Here, we aim to further the development of deep learning methods within biology by providing application examples and ready to apply and adapt code templates. Given such examples, we illustrate how architectures consisting of convolutional and long short‐term memory neural networks can relatively easily be designed and trained to state‐of‐the‐art performance on three biological sequence problems: prediction of subcellular localization, protein secondary structure and the binding of peptides to MHC Class II molecules. Availability and implementation All implementations and datasets are available online to the scientific community at https://github.com/vanessajurtz/lasagne4bio. Contact [email protected] Supplementary information Supplementary data are available at Bioinformatics online.


international conference on bioinformatics | 2017

Deep Recurrent Conditional Random Field Network for Protein Secondary Prediction

Alexander Rosenberg Johansen; Casper Kaae Sønderby; Søren Kaae Sønderby; Ole Winther

Deep learning has become the state-of-the-art method for predicting protein secondary structure from only its amino acid residues and sequence profile. Building upon these results, we propose to combine a bi-directional recurrent neural network (biRNN) with a conditional random field (CRF), which we call the biRNN-CRF. The biRNN-CRF may be seen as an improved alternative to an auto-regressive uni-directional RNN where predictions are performed sequentially conditioning on the prediction in the previous time-step. The CRF is instead nearest neighbor-aware and models for the joint distribution of the labels for all time-steps. We condition the CRF on the output of biRNN, which learns a distributed representation based on the entire sequence. The biRNN-CRF is therefore close to ideally suited for the secondary structure task because a high degree of cross-talk between neighboring elements can be expected. We validate the model on several benchmark datasets. For example, on CB513, a model with 1.7 million parameters, achieves a Q8 accuracy of 69.4 for single model and 70.9 for ensemble, which to our knowledge is state-of-the-art.


international conference on machine learning | 2016

Autoencoding beyond pixels using a learned similarity metric

Anders Boesen Lindbo Larsen; Søren Kaae Sønderby; Hugo Larochelle; Ole Winther


international conference on machine learning | 2016

Auxiliary deep generative models

Lars Maaløe; Casper Kaae Sønderby; Søren Kaae Sønderby; Ole Winther


neural information processing systems | 2016

Ladder Variational Autoencoders

Casper Kaae Sønderby; Tapani Raiko; Lars Maaløe; Søren Kaae Sønderby; Ole Winther


neural information processing systems | 2016

Sequential Neural Models with Stochastic Layers

Marco Fraccaro; Søren Kaae Sønderby; Ulrich Paquet; Ole Winther


international conference on machine learning | 2016

How to Train Deep Variational Autoencoders and Probabilistic Ladder Networks

Casper Kaae Sønderby; Tapani Raiko; Lars Maaløe; Søren Kaae Sønderby; Ole Winther


arXiv: Computer Vision and Pattern Recognition | 2015

Recurrent Spatial Transformer Networks

Søren Kaae Sønderby; Casper Kaae Sønderby; Lars Maaløe; Ole Winther

Collaboration


Dive into the Søren Kaae Sønderby's collaboration.

Top Co-Authors

Avatar

Ole Winther

Technical University of Denmark

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Lars Maaløe

Technical University of Denmark

View shared research outputs
Top Co-Authors

Avatar

Henrik Nielsen

Technical University of Denmark

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Marco Fraccaro

Technical University of Denmark

View shared research outputs
Top Co-Authors

Avatar

Morten Nielsen

Technical University of Denmark

View shared research outputs
Researchain Logo
Decentralizing Knowledge