Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Houman Ghaemmaghami is active.

Publication


Featured researches published by Houman Ghaemmaghami.


international conference of the ieee engineering in medicine and biology society | 2009

Normal probability testing of snore signals for diagnosis of obstructive sleep apnea

Houman Ghaemmaghami; Udantha R. Abeyratne; Craig Hukins

Obstructive Sleep Apnea (OSA) is a highly prevalent disease in which upper airways are collapsed during sleep, leading to serious consequences. The standard method of OSA diagnosis is known as Polysomnography (PSG), which requires an overnight stay in a specifically equipped facility, connected to over 15 channels of measurements. PSG requires (i) contact instrumentation and, (ii) the expert human scoring of a vast amount of data based on subjective criteria. PSG is expensive, time consuming and is difficult to use in community screening or pediatric assessment. Snoring is the most common symptom of OSA. Despite the vast potential, however, it is not currently used in the clinical diagnosis of OSA. In this paper, we propose a novel method of snore signal analysis for the diagnosis of OSA. The method is based on a novel feature that quantifies the non-Gaussianity of individual episodes of snoring. The proposed method was evaluated using overnight clinical snore sound recordings of 86 subjects. The recordings were made concurrently with routine PSG, which was used to establish the ground truth via standard clinical diagnostic procedures. The results indicated that the developed method has a detectability accuracy of 97.34%.


international conference on acoustics, speech, and signal processing | 2010

Noise robust voice activity detection using normal probability testing and time-domain histogram analysis

Houman Ghaemmaghami; David Dean; Sridha Sridharan; Iain A. McCowan

This paper presents a method of voice activity detection (VAD) suitable for high noise scenarios, based on the fusion of two complementary systems. The first system uses a proposed non-Gaussianity score (NGS) feature based on normal probability testing. The second system employs a histogram distance score (HDS) feature that detects changes in the signal through conducting a template-based similarity measure between adjacent frames. The decision outputs by the two systems are then merged using an open-by-reconstruction fusion stage. Accuracy of the proposed method was compared to several baseline VAD methods on a database created using real recordings of a variety of high-noise environments.


international conference on signal processing and communication systems | 2008

Speech Endpoint Detection Using Gradient Based Edge Detection Techniques

Houman Ghaemmaghami; Robert J. Vogt; Sridha Sridharan; Michael Mason

This paper proposes a novel method for speech endpoint detection. The developed method utilises gradient based edge detection algorithms, used in image processing, to detect boundaries of continuous speech in noisy conditions. It is simple and has low computational complexity. The accuracy of the proposed method was evaluated and compared to the ITU-T G.729 Annex-B voice activity detection (VAD) algorithm. To do this, the two algorithms were tested using a synthetically produced noisy-speech database, consisting of noisy-speech signals at various lengths and SNR. The results indicated that the developed method outperforms the G.729-B VAD algorithm at various signal-to-noise ratios.


conference of the international speech communication association | 2016

Speakers In The Wild (SITW): The QUT Speaker Recognition System.

Houman Ghaemmaghami; Md. Hafizur Rahman; Ivan Himawan; David Dean; Ahilan Kanagasundaram; Sridha Sridharan; Clinton Fookes

This paper presents the QUT speaker recognition system, as a competing system in the Speakers In The Wild (SITW) speaker recognition challenge. Our proposed system achieved an overall ranking of second place, in the main core-core condition evaluations of the SITW challenge. This system uses an ivector/ PLDA approach, with domain adaptation and a deep neural network (DNN) trained to provide feature statistics. The statistics are accumulated by using class posteriors from the DNN, in place of GMM component posteriors in a typical GMM UBM i-vector/PLDA system. Once the statistics have been collected, the i-vector computation is carried out as in a GMM-UBM based system. We apply domain adaptation to the extracted i-vectors to ensure robustness against dataset variability, PLDA modelling is used to capture speaker and session variability in the i-vector space, and the processed i-vectors are compared using the batch likelihood ratio. The final scores are calibrated to obtain the calibrated likelihood scores, which are then used to carry out speaker recognition and evaluate the performance of the system. Finally, we explore the practical application of our system to the core-multi condition recordings of the SITW data and propose a technique for speaker recognition in recordings with multiple speakers.


acm multimedia | 2015

Acoustic Adaptation in Cross Database Audio Visual SHMM Training for Phonetic Spoken Term Detection

Shahram Kalantari; David Dean; Sridha Sridharan; Houman Ghaemmaghami; Clinton Fookes

Visual information in the form of lip movements of the speaker has been shown to improve the performance of speech recognition and search applications. In our previous work, we proposed cross database training of synchronous hidden Markov models (SHMMs) to make use of external large and publicly available audio databases in addition to the relatively small given audio visual database. In this work, the cross database training approach is improved by performing an additional audio adaptation step, which enables audio visual SHMMs to benefit from audio observations of the external audio models before adding visual modality to them. The proposed approach outperforms the baseline cross database training approach in clean and noisy environments in terms of phone recognition accuracy as well as spoken term detection (STD) accuracy.


Faculty of Built Environment and Engineering; Information Security Institute | 2010

Noise robust voice activity detection using features extracted from the time-domain autocorrelation function

Houman Ghaemmaghami; Brendan Baker; Robert J. Vogt; Sridha Sridharan


Archive | 2009

MULTI-PARAMETRIC ANALYSIS OF SNORE SOUNDS FOR THE COMMUNITY SCREENING OF SLEEP APNEA WITH NON-GAUSSIANITY INDEX

Udantha R. Abeyratne; Asela Samantha Karunajeewa; Houman Ghaemmaghami


Faculty of Built Environment and Engineering; Information Security Institute | 2011

Extending the task of diarization to speaker attribution

Houman Ghaemmaghami; David Dean; Robbie Vogt; Sridha Sridharan


Faculty of Built Environment and Engineering; Information Security Institute | 2013

Speaker Attribution of Australian Broadcast News Data

Houman Ghaemmaghami; David Dean; Sridha Sridharan


conference of the international speech communication association | 2015

The QUT-NOISE-SRE protocol for the evaluation of noisy speaker recognition

David Dean; Ahilan Kanagasundaram; Houman Ghaemmaghami; Hafizur Rahman; Sridha Sridharan

Collaboration


Dive into the Houman Ghaemmaghami's collaboration.

Top Co-Authors

Avatar

Sridha Sridharan

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar

David Dean

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar

Clinton Fookes

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar

Shahram Kalantari

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar

Ahilan Kanagasundaram

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar

Robert J. Vogt

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Robbie Vogt

Queensland University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Brendan Baker

Queensland University of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge