Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Milos Zelezný is active.

Publication


Featured researches published by Milos Zelezný.


SPECOM | 2018

LipsID Using 3D Convolutional Neural Networks

Miroslav Hlaváč; Ivan Gruber; Milos Zelezný; Alexey Karpov

This paper presents a proposition for a method inspired by iVectors for improvement of visual speech recognition in the similar way iVectors are used to improve the recognition rate of audio speech recognition. A neural network for feature extraction is presented with training parameters and evaluation. The network is trained as a classifier for a closed set of 64 speakers from the UWB-HSCAVC dataset and then the last softmax fully connected layer is removed to gain a feature vector of size 256. The network is provided with sequences of 15 frames and outputs the softmax classification to 64 classes. The training data consists of approximately 20000 sequences of grayscale images from the first 50 sentences that are common to every speaker. The network is then evaluated on the 60000 sequences created from 150 sentences from each speaker. The testing sentences are different for each speaker.


SPECOM | 2018

Designing Advanced Geometric Features for Automatic Russian Visual Speech Recognition

Denis Ivanko; Dmitry Ryumin; Alexandr Axyonov; Milos Zelezný

The use of video information plays an increasingly important role for automatic speech recognition. Nowadays, audio-only based systems have reached a certain accuracy threshold and many researchers see a solution to the problem in the use of visual modality to obtain better results. Despite the fact that audio modality of speech is much more representative than video, their proper fusion can improve both quality and robustness of the entire recognition system that was proved in practice by many researches. However, no agreement between researchers on the optimal set of visual features was reached. In this paper, we investigate this issue in more detail and propose advanced geometry-based visual features for automatic Russian lip-reading system. The experiments were conducted using collected HAVRUS audio-visual speech database. The average viseme recognition accuracy of our system trained on the entire corpus is 40.62%. We also tested the main state-of-the-art methods for visual speech recognition, applying them to continuous Russian speech with high-speed recordings (200 frames per seconds).


Proceedings of the Auditory-Visual Speech Processing International Conference 2005 | 2005

Design and recording of Czech speech corpus for audio-visual continuous speech recognition

Petr Císar; Milos Zelezný; Zdenek Krnoul; Jakub Kanis; Jan Zelinka; Ludek Müller


conference of the international speech communication association | 2014

Audio-visual signal processing in a multimodal assisted living environment.

Alexey Karpov; Lale Akarun; Hulya Yalcin; Alexander L. Ronzhin; Barış Evrim Demiröz; Aysun Çoban; Milos Zelezný


conference of the international speech communication association | 2009

Audio-Visual Speech Asynchrony Modeling in a Talking Head

Alexey Karpov; Liliya Tsirulnik; Zdenek Krnoul; Andrey Ronzhin; Boris Lobanov; Milos Zelezný


conference of the international speech communication association | 2006

Training of Coarticulation Models using Dominance Functions and Visual Unit Selection Methods for Audio-Visual Speech Synthesis

Zdenek Krnoul; Milos Zelezný; Ludek Müller; Jakub Kanis


International Journal of Computer Science & Applications | 2013

Modeling of Pronunciation, Language and Nonverbal Units at Conversational Russian Speech Recognition.

Irina S. Kipyatkova; Alexey Karpov; Vasilisa Verkhodanova; Milos Zelezný


conference of the international speech communication association | 2010

Viseme-Dependent Weight Optimization for CHMM-Based Audio-Visual Speech Recognition

Alexey Karpov; Andrey Ronzhin; Konstantin Markov; Milos Zelezný


Proceedings of AVSP 2003 | 2003

Czech audio-visual speech corpus of a car driver for in-vehicle audio-visual speech recognition

Milos Zelezný; Petr Císar


conference of the international speech communication association | 2008

A development of Czech talking head.

Zdenek Krnoul; Milos Zelezný

Collaboration


Dive into the Milos Zelezný's collaboration.

Top Co-Authors

Avatar

Zdenek Krnoul

University of West Bohemia

View shared research outputs
Top Co-Authors

Avatar

Alexey Karpov

Russian Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Andrey Ronzhin

Russian Academy of Sciences

View shared research outputs
Top Co-Authors

Avatar

Jakub Kanis

University of West Bohemia

View shared research outputs
Top Co-Authors

Avatar

Jan Zelinka

University of West Bohemia

View shared research outputs
Top Co-Authors

Avatar

Ludek Müller

University of West Bohemia

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Jana Trojanová

University of West Bohemia

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge