Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Muhammad S. A. Zilany is active.

Publication


Featured researches published by Muhammad S. A. Zilany.


Journal of the Acoustical Society of America | 2006

Modeling auditory-nerve responses for high sound pressure levels in the normal and impaired auditory periphery

Muhammad S. A. Zilany; Ian C. Bruce

This paper presents a computational model to simulate normal and impaired auditory-nerve (AN) fiber responses in cats. The model responses match physiological data over a wider dynamic range than previous auditory models. This is achieved by providing two modes of basilar membrane excitation to the inner hair cell (IHC) rather than one. The two modes are generated by two parallel filters, component 1 (C1) and component 2 (C2), and the outputs are subsequently transduced by two separate functions. The responses are then added and passed through the IHC low-pass filter followed by the IHC-AN synapse model and discharge generator. The C1 filter is a narrow-band, chirp filter with the gain and bandwidth controlled by a nonlinear feed-forward control path. This filter is responsible for low and moderate level responses. A linear, static, and broadly tuned C2 filter followed by a nonlinear, inverted and nonrectifying C2 transduction function is critical for producing transition region and high-level effects. Consistent with Kiangs two-factor cancellation hypothesis, the interaction between the two paths produces effects such as the C1/C2 transition and peak splitting in the period histogram. The model responses are consistent with a wide range of physiological data from both normal and impaired ears for stimuli presented at levels spanning the dynamic range of hearing.


Journal of the Acoustical Society of America | 2007

Representation of the vowel /ε/ in normal and impaired auditory nerve fibers: Model predictions of responses in cats

Muhammad S. A. Zilany; Ian C. Bruce

The temporal response of auditory-nerve (AN) fibers to a steady-state vowel is investigated using a computational auditory-periphery model. The model predictions are validated against a wide range of physiological data for both normal and impaired fibers in cats. The model incorporates two parallel filter paths, component 1 (C1) and component 2 (C2), which correspond to the active and passive modes of basilar membrane vibration, respectively, in the cochlea. The outputs of the two filters are subsequently transduced by two separate functions, added together, and then low-pass filtered by the inner hair cell (IHC) membrane, which is followed by the IHC-AN synapse and discharge generator. The C1 response dominates at low and moderate levels and is responsible for synchrony capture and multiformant responses seen in the vowel responses. The C2 response dominates at high levels and contributes to the loss of synchrony capture observed in normal and impaired fibers. The interaction between C1 and C2 responses explains the behavior of AN fibers in the transition region, which is characterized by two important observations in the vowel responses: First, all components of the vowel undergo the C1/C2 transition simultaneously, and second, the responses to the nonformant components of the vowel become substantial.


Journal of the Acoustical Society of America | 2014

Updated parameters and expanded simulation options for a model of the auditory periphery

Muhammad S. A. Zilany; Ian C. Bruce; Laurel H. Carney

A phenomenological model of the auditory periphery in cats was previously developed by Zilany and colleagues [J. Acoust. Soc. Am. 126, 2390-2412 (2009)] to examine the detailed transformation of acoustic signals into the auditory-nerve representation. In this paper, a few issues arising from the responses of the previous version have been addressed. The parameters of the synapse model have been readjusted to better simulate reported physiological discharge rates at saturation for higher characteristic frequencies [Liberman, J. Acoust. Soc. Am. 63, 442-455 (1978)]. This modification also corrects the responses of higher-characteristic frequency (CF) model fibers to low-frequency tones that were erroneously much higher than the responses of low-CF model fibers in the previous version. In addition, an analytical method has been implemented to compute the mean discharge rate and variance from the models synapse output that takes into account the effects of absolute refractoriness.


The Journal of Neuroscience | 2010

Power-law dynamics in an auditory-nerve model can account for neural adaptation to sound-level statistics

Muhammad S. A. Zilany; Laurel H. Carney

Neurons in the auditory system respond to recent stimulus-level history by adapting their response functions according to the statistics of the stimulus, partially alleviating the so-called “dynamic-range problem.” However, the mechanism and source of this adaptation along the auditory pathway remain unknown. Inclusion of power-law dynamics in a phenomenological model of the inner hair cell (IHC)–auditory nerve (AN) synapse successfully explained neural adaptation to sound-level statistics, including the time course of adaptation of the mean firing rate and changes in the dynamic range observed in AN responses. A direct comparison between model responses to a dynamic stimulus and to an “inversely gated” static background suggested that AN dynamic-range adaptation largely results from the adaptation produced by the response history. These results support the hypothesis that the potential mechanism underlying the dynamic-range adaptation observed at the level of the auditory nerve is located peripheral to the spike generation mechanism and central to the IHC receptor potential.


Biomedical Signal Processing and Control | 2015

Artifacts-matched blind source separation and wavelet transform for multichannel EEG denoising

Rakibul Mowla; Siew-Cheok Ng; Muhammad S. A. Zilany; Raveendran Paramesran

Abstract The physiological artifacts such as electromyogram (EMG) and electrooculogram (EOG) remain a major problem in electroencephalogram (EEG) research. A number of techniques are currently in use to remove these artifacts with the hope that the process does not unduly degrade the quality of the obscured EEG. In this paper, a new method has been proposed by combining two techniques: a canonical correlation analysis (CCA) followed by a stationary wavelet transform (SWT) to remove EMG artifacts and a second-order blind identification (SOBI) technique followed by SWT to remove EOG artifacts. The simulation results show that these combinations are more effective than either using the individual techniques alone or using other combinations of techniques. The quality of the artifact removal is evaluated by calculating correlations between processed and unprocessed data, and the practicability of the technique is judged by comparing execution times of the algorithms.


Journal of Neuroscience Methods | 2012

Semi-supervised spike sorting using pattern matching and a scaled Mahalanobis distance metric.

Douglas M. Schwarz; Muhammad S. A. Zilany; Melissa Skevington; Nicholas J. Huang; Brian C. Flynn; Laurel H. Carney

Sorting action potentials (spikes) from tetrode recordings can be time consuming, labor intensive, and inconsistent, depending on the methods used and the experience of the operator. The techniques presented here were designed to address these issues. A feature related to the slope of the spike during repolarization is computed. A small subsample of the features obtained from the tetrode (ca. 10,000-20,000 events) is clustered using a modified version of k-means that uses Mahalanobis distance and a scaling factor related to the cluster size. The cluster-size-based scaling improves the clustering by increasing the separability of close clusters, especially when they are of disparate size. The full data set is then classified from the statistics of the clusters. The technique yields consistent results for a chosen number of clusters. A MATLAB implementation is able to classify more than 5000 spikes per second on a modern workstation.


IEEE Transactions on Audio, Speech, and Language Processing | 2015

Prediction of speech intelligibility using a neurogram orthogonal polynomial measure (NOPM)

Nursadul Mamun; Wissam A. Jassim; Muhammad S. A. Zilany

Sensorineural hearing loss (SNHL) is an increasingly prevalent condition, resulting from damage to the inner ear and causing a reduction in speech intelligibility. This paper proposes a new speech intelligibility prediction metric, the neurogram orthogonal polynomial measure (NOPM). This metric applies orthogonal moments to the auditory neurogram to predict speech intelligibility for listeners with and without hearing loss. The model simulates the responses of auditory-nerve fibers to speech signals under quiet and noisy conditions. Neurograms were created using a physiologically based computational model of the auditory periphery. A well-known orthogonal polynomial measure, Krawtchouk moments, was applied to extract features from the auditory neurogram. The predicted intelligibility scores were compared to subjective results, and NOPM showed a good fit with the subjective scores for normal listeners and also for listeners with hearing loss. The proposed metric has a realistic and wider dynamic range than corresponding existing metrics, such as mean structural similarity index measure and neurogram similarity index measure, and the predicted scores are also well-separated as a function of hearing loss. The application of this metric could be extended for assessing hearing-aid and speech-enhancement algorithms.


Iet Signal Processing | 2014

Enhancing noisy speech signals using orthogonal moments

Wissam A. Jassim; Raveendran Paramesran; Muhammad S. A. Zilany

This study describes a new approach to enhance noisy speech signals using the discrete Tchebichef transform (DTT) and the discrete Krawtchouk transform (DKT). The DTT and DKT are based on well-known orthogonal moments: the Tchebichef and Krawtchouk moments, respectively. The representations of speech signals using a limited number of moment coefficients and their behaviour in the domain of orthogonal moments are shown. The method involves removing noise from the signal using a minimum-mean-square error in the domain of the DTT or DKT. According to comparisons with traditional methods, the initial experiments yield promising results and show that orthogonal moments are applicable in the field of speech signal enhancement. The application of orthogonal moments could be extended to speech analysis, compression and recognition.


PLOS ONE | 2016

A Robust Speaker Identification System Using the Responses from a Model of the Auditory Periphery

Md. Atiqul Islam; Wissam A. Jassim; Ng Siew Cheok; Muhammad S. A. Zilany

Speaker identification under noisy conditions is one of the challenging topics in the field of speech processing applications. Motivated by the fact that the neural responses are robust against noise, this paper proposes a new speaker identification system using 2-D neurograms constructed from the responses of a physiologically-based computational model of the auditory periphery. The responses of auditory-nerve fibers for a wide range of characteristic frequency were simulated to speech signals to construct neurograms. The neurogram coefficients were trained using the well-known Gaussian mixture model-universal background model classification technique to generate an identity model for each speaker. In this study, three text-independent and one text-dependent speaker databases were employed to test the identification performance of the proposed method. Also, the robustness of the proposed method was investigated using speech signals distorted by three types of noise such as the white Gaussian, pink, and street noises with different signal-to-noise ratios. The identification results of the proposed neural-response-based method were compared to the performances of the traditional speaker identification methods using features such as the Mel-frequency cepstral coefficients, Gamma-tone frequency cepstral coefficients and frequency domain linear prediction. Although the classification accuracy achieved by the proposed method was comparable to the performance of those traditional techniques in quiet, the new feature was found to provide lower error rates of classification under noisy environments.


PLOS ONE | 2016

Reference-Free Assessment of Speech Intelligibility Using Bispectrum of an Auditory Neurogram

Mohammad E. Hossain; Wissam A. Jassim; Muhammad S. A. Zilany

Sensorineural hearing loss occurs due to damage to the inner and outer hair cells of the peripheral auditory system. Hearing loss can cause decreases in audibility, dynamic range, frequency and temporal resolution of the auditory system, and all of these effects are known to affect speech intelligibility. In this study, a new reference-free speech intelligibility metric is proposed using 2-D neurograms constructed from the output of a computational model of the auditory periphery. The responses of the auditory-nerve fibers with a wide range of characteristic frequencies were simulated to construct neurograms. The features of the neurograms were extracted using third-order statistics referred to as bispectrum. The phase coupling of neurogram bispectrum provides a unique insight for the presence (or deficit) of supra-threshold nonlinearities beyond audibility for listeners with normal hearing (or hearing loss). The speech intelligibility scores predicted by the proposed method were compared to the behavioral scores for listeners with normal hearing and hearing loss both in quiet and under noisy background conditions. The results were also compared to the performance of some existing methods. The predicted results showed a good fit with a small error suggesting that the subjective scores can be estimated reliably using the proposed neural-response-based metric. The proposed metric also had a wide dynamic range, and the predicted scores were well-separated as a function of hearing loss. The proposed metric successfully captures the effects of hearing loss and supra-threshold nonlinearities on speech intelligibility. This metric could be applied to evaluate the performance of various speech-processing algorithms designed for hearing aids and cochlear implants.

Collaboration


Dive into the Muhammad S. A. Zilany's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge