Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Moncef Gabbouj is active.

Publication


Featured researches published by Moncef Gabbouj.


international conference on image processing | 2002

The error concealment feature in the H.26L test model

Ye-Kui Wang; Miska M. Hannuksela; Viktor Varsa; Ari Hourunranta; Moncef Gabbouj

This paper presents the error concealment (EC) feature implemented by the authors in the test model of the draft ITU-T video coding standard H.26L. The selected EC algorithms are based on weighted pixel value averaging for INTRA. pictures and boundary-matching-based motion vector recovery for INTER pictures. The specific concealment strategy and some special methods, including handling of B-pictures, multiple reference frames and entire frame losses, are described. Both subjective and objective results are given based on simulations under Internet conditions. The feature was adopted and is now included in the latest H.26L reference software TML-9.0.


acm sigmm conference on multimedia systems | 2011

Rate adaptation for adaptive HTTP streaming

Chenghao Liu; Imed Bouazizi; Moncef Gabbouj

Recently, HTTP has been widely used for the delivery of real-time multimedia content over the Internet, such as in video streaming applications. To combat the varying network resources of the Internet, rate adaptation is used to adapt the transmission rate to the varying network capacity. A key research problem of rate adaptation is to identify network congestion early enough and to probe the spare network capacity. In adaptive HTTP streaming, this problem becomes challenging because of the difficulties in differentiating between the short-term throughput variations, incurred by the TCP congestion control, and the throughput changes due to more persistent bandwidth changes. In this paper, we propose a novel rate adaptation algorithm for adaptive HTTP streaming that detects bandwidth changes using a smoothed HTTP throughput measured based on the segment fetch time (SFT). The smoothed HTTP throughput instead of the instantaneous TCP transmission rate is used to determine if the bitrate of the current media matches the end-to-end network bandwidth capacity. Based on the smoothed throughput measurement, this paper presents a receiver-driven rate adaptation method for HTTP/TCP streaming that deploys a step-wise increase/ aggressive decrease method to switch up/down between the different representations of the content that are encoded at different bitrates. Our rate adaptation method does not require any transport layer information such as round trip time (RTT) and packet loss rates which are available at the TCP layer. Simulation results show that the proposed rate adaptation algorithm quickly adapts to match the end-to-end network capacity and also effectively controls buffer underflow and overflow.


EURASIP Journal on Advances in Signal Processing | 2009

The emerging MVC standard for 3D video services

Ying Chen; Ye-Kui Wang; Kemal Ugur; Miska Hannuksela; Jani Lainema; Moncef Gabbouj

Multiview video has gained a wide interest recently. The huge amount of data needed to be processed by multiview applications is a heavy burden for both transmission and decoding. The joint video team has recently devoted part of its effort to extend the widely deployed H.264/AVC standard to handle multiview video coding (MVC). The MVC extension of H.264/AVC includes a number of new techniques for improved coding efficiency, reduced decoding complexity, and new functionalities for multiview operations. MVC takes advantage of some of the interfaces and transport mechanisms introduced for the scalable video coding (SVC) extension of H.264/AVC, but the system level integration of MVC is conceptually more challenging as the decoder output may contain more than one view and can consist of any combination of the views with any temporal level. The generation of all the output views also requires careful consideration and control of the available decoder resources. In this paper, multiview applications and solutions to support generic multiview as well as 3D services are introduced. The proposed solutions, which have been adopted to the draft MVC specification, cover a wide range of requirements for 3D video related to interface, transport of the MVC bitstreams, and MVC decoder resource management. The features that have been introduced in MVC to support these solutions include marking of reference pictures, supporting for efficient view switching, structuring of the bitstream, signalling of view scalability supplemental enhancement information (SEI) and parallel decoding SEI.


multidimensional signal processing workshop | 1989

Optimal stack filtering and the estimation and structural approaches to image processing

Edward J. Coyle; Jeanhsang Lin; Moncef Gabbouj

Summary form only given. Two approaches have been used in the past to design rank-order-based nonlinear filters to enhance or restore images: the structural approach and the estimation approach. The first approach requires structural descriptions of the image and the process which has altered it, whereas the second required statistical descriptions. The many different classes of rank-order-based filters that have been developed over the last few decades have been reviewed in the context of these two approaches. One of these filter classes, stack filters, has been investigated. These filters, which are defined by a weak superposition property and an ordering property, contain all compositions of 2D rank-order operations. The recently developed theory of minimum-mean-absolute-error (MMAE) stack filtering has been extended to two dimensions. A theory of optimal stack filtering under structural constraints and goals has been developed for the structural approach to image processing. These two optimal stack filtering theories have been combined into a single design theory for rank-order-based filters. >


IEEE Transactions on Signal Processing | 1995

Optimal weighted median filtering under structural constraints

Ruikang Yang; Lin Yin; Moncef Gabbouj; Jaakko Astola

A new expression for the output moments of weighted median filtered data is derived. The noise attenuation capability of a weighted median filter can now be assessed using the L-vector and M-vector parameters in the new expression. The second major contribution of the paper is the development of a new optimality theory for weighted median filters. This theory is based on the new expression for the output moments, and combines the noise attenuation and some structural constraints on the filters behavior. In certain special cases, the optimal weighted median filter can be obtained by merely solving a set of linear inequalities. This leads in some cases to closed form solutions for optimal weighted median filters. Some applications of the theory developed in this paper, in 1-D signal processing and image processing are discussed. Throughout the analysis, some striking similarities are pointed out between linear FIR filters and weighted median filters. >


Neural Networks | 2009

Evolutionary artificial neural networks by multi-dimensional particle swarm optimization

Serkan Kiranyaz; Turker Ince; E. Alper Yildirim; Moncef Gabbouj

In this paper, we propose a novel technique for the automatic design of Artificial Neural Networks (ANNs) by evolving to the optimal network configuration(s) within an architecture space. It is entirely based on a multi-dimensional Particle Swarm Optimization (MD PSO) technique, which re-forms the native structure of swarm particles in such a way that they can make inter-dimensional passes with a dedicated dimensional PSO process. Therefore, in a multidimensional search space where the optimum dimension is unknown, swarm particles can seek both positional and dimensional optima. This eventually removes the necessity of setting a fixed dimension a priori, which is a common drawback for the family of swarm optimizers. With the proper encoding of the network configurations and parameters into particles, MD PSO can then seek the positional optimum in the error space and the dimensional optimum in the architecture space. The optimum dimension converged at the end of a MD PSO process corresponds to a unique ANN configuration where the network parameters (connections, weights and biases) can then be resolved from the positional optimum reached on that dimension. In addition to this, the proposed technique generates a ranked list of network configurations, from the best to the worst. This is indeed a crucial piece of information, indicating what potential configurations can be alternatives to the best one, and which configurations should not be used at all for a particular problem. In this study, the architecture space is defined over feed-forward, fully-connected ANNs so as to use the conventional techniques such as back-propagation and some other evolutionary methods in this field. The proposed technique is applied over the most challenging synthetic problems to test its optimality on evolving networks and over the benchmark problems to test its generalization capability as well as to make comparative evaluations with the several competing techniques. The experimental results show that the MD PSO evolves to optimum or near-optimum networks in general and has a superior generalization capability. Furthermore, the MD PSO naturally favors a low-dimension solution when it exhibits a competitive performance with a high dimension counterpart and such a native tendency eventually yields the evolution process to the compact network configurations in the architecture space rather than the complex ones, as long as the optimality prevails.


IEEE Transactions on Biomedical Engineering | 2009

A Generic and Robust System for Automated Patient-Specific Classification of ECG Signals

Turker Ince; Serkan Kiranyaz; Moncef Gabbouj

This paper presents a generic and patient-specific classification system designed for robust and accurate detection of ECG heartbeat patterns. The proposed feature extraction process utilizes morphological wavelet transform features, which are projected onto a lower dimensional feature space using principal component analysis, and temporal features from the ECG data. For the pattern recognition unit, feedforward and fully connected artificial neural networks, which are optimally designed for each patient by the proposed multidimensional particle swarm optimization technique, are employed. By using relatively small common and patient-specific training data, the proposed classification system can adapt to significant interpatient variations in ECG patterns by training the optimal network structure, and thus, achieves higher accuracy over larger datasets. The classification experiments over a benchmark database demonstrate that the proposed system achieves such average accuracies and sensitivities better than most of the current state-of-the-art algorithms for detection of ventricular ectopic beats (VEBs) and supra-VEBs (SVEBs). Over the entire database, the average accuracy-sensitivity performances of the proposed system for VEB and SVEB detections are 98.3%-84.6% and 97.4%-63.5%, respectively. Finally, due to its parameter-invariant nature, the proposed system is highly generic, and thus, applicable to any ECG dataset.


IEEE Transactions on Audio, Speech, and Language Processing | 2010

Voice Conversion Using Partial Least Squares Regression

Elina Helander; Tuomas Virtanen; Jani Nurminen; Moncef Gabbouj

Voice conversion can be formulated as finding a mapping function which transforms the features of the source speaker to those of the target speaker. Gaussian mixture model (GMM)-based conversion is commonly used, but it is subject to overfitting. In this paper, we propose to use partial least squares (PLS)-based transforms in voice conversion. To prevent overfitting, the degrees of freedom in the mapping can be controlled by choosing a suitable number of components. We propose a technique to combine PLS with GMMs, enabling the use of multiple local linear mappings. To further improve the perceptual quality of the mapping where rapid transitions between GMM components produce audible artefacts, we propose to low-pass filter the component posterior probabilities. The conducted experiments show that the proposed technique results in better subjective and objective quality than the baseline joint density GMM approach. In speech quality conversion preference tests, the proposed method achieved 67% preference score against the smoothed joint density GMM method and 84% preference score against the unsmoothed joint density GMM method. In objective tests the proposed method produced a lower Mel-cepstral distortion than the reference methods.


Circuits Systems and Signal Processing | 1992

An overview of median and stack filtering

Edward J. Coyle; Neal C. Gallagher; Moncef Gabbouj

Within the last two decades a small group of researchers has built a useful, nontrivial theory of nonlinear signal processing around the median-related filters known as rank-order filters, order-statistic filters, weighted median filters, and stack filters. This required significant effort to overcome the bias, both in education and research, toward linear theory, which has been dominant since the days of Fourier, Laplace, and “Convolute.”We trace the development of this theory of nonlinear filtering from its beginnings in the study of noise-removal properties and structural behavior of the median filter to the recently developed theory of optimal stack filtering.The theory of stack filtering provides a point of view which unifies many different filter classes, including morphological filters, so it is discussed in detail. Of particular importance is the way this theory has brought together, in a single analytical framework, both the estimation-based and the structural-based approaches to the design of these filters.Some recent applications of median and stack filters are provided to demonstrate the effectiveness of this approach to nonlinear filtering. They include: the design of an optimal stack filter for image restoration; the use of vector median filters to attenuate impulsive noise in color images and to eliminate cross luminance and cross color in TV images; and the use of median-based filters for image sequence coding, reconstruction, and scan rate conversion in normal TV and HDTV systems.


IEEE Transactions on Acoustics, Speech, and Signal Processing | 1990

Minimum mean absolute error stack filtering with structural constraint and goals

Moncef Gabbouj; Edward J. Coyle

A theory for the structural behavior of stack filters is developed. This theory provides a test which can determine if a given stack filter has any root signals; a method for classifying the root signal behavior of any stack filter found to have roots; and, perhaps most important, a method for designing stack filters with specific root signals or other structural behavior. This theory of root signals for stack filters is then combined with the theory of minimum mean absolute error stack filtering. This unified theory allows the designer to pick a filter which minimizes noise subject to constraints on its structural behavior. >

Collaboration


Dive into the Moncef Gabbouj's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar

Alexandros Iosifidis

Tampere University of Technology

View shared research outputs
Top Co-Authors

Avatar

Turker Ince

İzmir University of Economics

View shared research outputs
Top Co-Authors

Avatar

Lazhar Khriji

Sultan Qaboos University

View shared research outputs
Top Co-Authors

Avatar

Bogdan Cramariuc

Tampere University of Technology

View shared research outputs
Top Co-Authors

Avatar

Ioan Tabus

Tampere University of Technology

View shared research outputs
Top Co-Authors

Avatar

Payman Aflaki

Tampere University of Technology

View shared research outputs
Top Co-Authors

Avatar

Imed Bouazizi

Tampere University of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge