Milos Zelezný
University of West Bohemia
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Milos Zelezný.
SPECOM | 2018
Miroslav Hlaváč; Ivan Gruber; Milos Zelezný; Alexey Karpov
This paper presents a proposition for a method inspired by iVectors for improvement of visual speech recognition in the similar way iVectors are used to improve the recognition rate of audio speech recognition. A neural network for feature extraction is presented with training parameters and evaluation. The network is trained as a classifier for a closed set of 64 speakers from the UWB-HSCAVC dataset and then the last softmax fully connected layer is removed to gain a feature vector of size 256. The network is provided with sequences of 15 frames and outputs the softmax classification to 64 classes. The training data consists of approximately 20000 sequences of grayscale images from the first 50 sentences that are common to every speaker. The network is then evaluated on the 60000 sequences created from 150 sentences from each speaker. The testing sentences are different for each speaker.
SPECOM | 2018
Denis Ivanko; Dmitry Ryumin; Alexandr Axyonov; Milos Zelezný
The use of video information plays an increasingly important role for automatic speech recognition. Nowadays, audio-only based systems have reached a certain accuracy threshold and many researchers see a solution to the problem in the use of visual modality to obtain better results. Despite the fact that audio modality of speech is much more representative than video, their proper fusion can improve both quality and robustness of the entire recognition system that was proved in practice by many researches. However, no agreement between researchers on the optimal set of visual features was reached. In this paper, we investigate this issue in more detail and propose advanced geometry-based visual features for automatic Russian lip-reading system. The experiments were conducted using collected HAVRUS audio-visual speech database. The average viseme recognition accuracy of our system trained on the entire corpus is 40.62%. We also tested the main state-of-the-art methods for visual speech recognition, applying them to continuous Russian speech with high-speed recordings (200 frames per seconds).
Proceedings of the Auditory-Visual Speech Processing International Conference 2005 | 2005
Petr Císar; Milos Zelezný; Zdenek Krnoul; Jakub Kanis; Jan Zelinka; Ludek Müller
conference of the international speech communication association | 2014
Alexey Karpov; Lale Akarun; Hulya Yalcin; Alexander L. Ronzhin; Barış Evrim Demiröz; Aysun Çoban; Milos Zelezný
conference of the international speech communication association | 2009
Alexey Karpov; Liliya Tsirulnik; Zdenek Krnoul; Andrey Ronzhin; Boris Lobanov; Milos Zelezný
conference of the international speech communication association | 2006
Zdenek Krnoul; Milos Zelezný; Ludek Müller; Jakub Kanis
International Journal of Computer Science & Applications | 2013
Irina S. Kipyatkova; Alexey Karpov; Vasilisa Verkhodanova; Milos Zelezný
conference of the international speech communication association | 2010
Alexey Karpov; Andrey Ronzhin; Konstantin Markov; Milos Zelezný
Proceedings of AVSP 2003 | 2003
Milos Zelezný; Petr Císar
conference of the international speech communication association | 2008
Zdenek Krnoul; Milos Zelezný