Sharath Adavanne
Tampere University of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Sharath Adavanne.
international conference on acoustics, speech, and signal processing | 2017
Sharath Adavanne; Pasi Pertilä; Tuomas Virtanen
This paper proposes to use low-level spatial features extracted from multichannel audio for sound event detection. We extend the convolutional recurrent neural network to handle more than one type of these multichannel features by learning from each of them separately in the initial stages. We show that instead of concatenating the features of each channel into a single feature vector the network learns sound events in multichannel audio better when they are presented as separate layers of a volume. Using the proposed spatial features over monaural features on the same network gives an absolute F-score improvement of 6.1% on the publicly available TUT-SED 2016 dataset and 2.7% on the TUT-SED 2009 dataset that is fifteen times larger.
european signal processing conference | 2017
Emre Cakir; Sharath Adavanne; Giambattista Parascandolo; Konstantinos Drossos; Tuomas Virtanen
Bird sounds possess distinctive spectral structure which may exhibit small shifts in spectrum depending on the bird species and environmental conditions. In this paper, we propose using convolutional recurrent neural networks on the task of automated bird audio detection in real-life environments. In the proposed method, convolutional layers extract high dimensional, local frequency shift invariant features, while recurrent layers capture longer term dependencies between the features extracted from short time frames. This method achieves 88.5% Area Under ROC Curve (AUC) score on the unseen evaluation data and obtains the second place in the Bird Audio Detection challenge.
european signal processing conference | 2017
Sharath Adavanne; Konstantinos Drossos; Emre Cakir; Tuomas Virtanen
This paper studies the detection of bird calls in audio segments using stacked convolutional and recurrent neural networks. Data augmentation by blocks mixing and domain adaptation using a novel method of test mixing are proposed and evaluated in regard to making the method robust to unseen data. The contributions of two kinds of acoustic features (dominant frequency and log mel-band energy) and their combinations are studied in the context of bird audio detection. Our best achieved AUC measure on five cross-validations of the development data is 95.5% and 88.1% on the unseen evaluation data.
arXiv: Sound | 2017
Sharath Adavanne; Giambattista Parascandolo; Pasi Pertilä; Toni Heittola; Tuomas Virtanen
arXiv: Sound | 2017
Miroslav Malik; Sharath Adavanne; Konstantinos Drossos; Tuomas Virtanen; Dasa Ticha; Roman Jarina
Archive | 2017
Sharath Adavanne; Tuomas Virtanen
arXiv: Sound | 2017
Sharath Adavanne; Tuomas Virtanen
arXiv: Sound | 2017
Sharath Adavanne; Archontis Politis; Tuomas Virtanen
international symposium on neural networks | 2018
Sharath Adavanne; Archontis Politis; Tuomas Virtanen
arXiv: Sound | 2018
Sharath Adavanne; Archontis Politis; Joonas Nikunen; Tuomas Virtanen