Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Muhammad Sarim is active.

Publication


Featured researches published by Muhammad Sarim.


IEEE Transactions on Circuits and Systems for Video Technology | 2012

Outdoor Dynamic 3-D Scene Reconstruction

Hansung Kim; Jean-Yves Guillemaut; Takeshi Takai; Muhammad Sarim; Adrian Hilton

Existing systems for 3-D reconstruction from multiple view video use controlled indoor environments with uniform illumination and backgrounds to allow accurate segmentation of dynamic foreground objects. In this paper, we present a portable system for 3-D reconstruction of dynamic outdoor scenes that require relatively large capture volumes with complex backgrounds and nonuniform illumination. This is motivated by the demand for 3-D reconstruction of natural outdoor scenes to support film and broadcast production. Limitations of existing multiple view 3-D reconstruction techniques for use in outdoor scenes are identified. Outdoor 3-D scene reconstruction is performed in three stages: 1) 3-D background scene modeling using spherical stereo image capture; 2) multiple view segmentation of dynamic foreground objects by simultaneous video matting across multiple views; and 3) robust 3-D foreground reconstruction and multiple view segmentation refinement in the presence of segmentation and calibration errors. Evaluation is performed on several outdoor productions with complex dynamic scenes including people and animals. Results demonstrate that the proposed approach overcomes limitations of previous indoor multiple view reconstruction approaches enabling high-quality free-viewpoint rendering and 3-D reference models for production.


international conference on image processing | 2010

Natural image matting for multiple wide-baseline views

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut; Takeshi Takai; Hansung Kim

In this paper we present a novel approach to estimate the alpha mattes of a foreground object captured by a wide-baseline circular camera rig provided a single key frame trimap. Bayesian inference coupled with camera calibration information are used to propagate high confidence trimaps labels across the views. Recent techniques have been developed to estimate an alpha matte of an image using multiple views but they are limited to narrow baseline views with low foreground variation. The proposed wide-baseline trimap propagation is robust to inter-view foreground appearance changes, shadows and similarity in foreground/background appearance for cameras with opposing views enabling high quality alpha matte extraction using any state-of-the-art image matting algorithm.


Proceedings of the 1st international workshop on 3D video processing | 2010

Wide-baseline multi-view video segmentation for 3D reconstruction

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut; Hansung Kim; Takeshi Takai

Obtaining a foreground silhouette across multiple views is one of the fundamental steps in 3D reconstruction. In this paper we present a novel video segmentation approach, to obtain a foreground silhouette, for scenes captured by a wide-baseline camera rig given a sparse manual interaction in a single view. The algorithm is based on trimap propagation, a framework used in video matting. Bayesian inference coupled with camera calibration information are used to spatio-temporally propagate high confidence trimap labels across the multi-view video to obtain coarse silhouettes which are later refined using a matting algorithm. Recent techniques have been developed for foreground segmentation, based on image matting, in multiple views but they are limited to narrow baseline with low foreground variation. The proposed wide-baseline silhouette propagation is robust to inter-view foreground appearance changes, shadows and similarity in foreground/background appearance. The approach has demonstrated good performance in silhouette estimation for views up to 180 degree baseline (opposing views). The segmentation technique has been fully integrated in a multi-view reconstruction pipeline. The results obtained demonstrate the suitability of the technique for multi-view reconstruction with wide-baseline camera set-ups and natural background


international conference on image processing | 2009

Non-parametric natural image matting

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut; Hansung Kim

Natural image matting is an extremely challenging image processing problem due to its ill-posed nature. It often requires skilled user interaction to aid definition of foreground and background regions. Current algorithms use these predefined regions to build local foreground and background colour models. In this paper we propose a novel approach which uses non-parametric statistics to model image appearance variations. This technique overcomes the limitations of previous parametric approaches which are purely colour-based and thereby unable to model natural image structure. The proposed technique consists of three successive stages: (i) background colour estimation, (ii) foreground colour estimation, (iii) alpha estimation. Colour estimation uses patch-based matching techniques to efficiently recover the optimum colour by comparison against patches from the known regions. Quantitative evaluation against ground truth demonstrates that the technique produces better results and successfully recovers fine details such as hair where many other algorithms fail.


international conference on image processing | 2011

Temporal trimap propagation for video matting using inferential statistics

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut

This paper introduces a statistical inference framework to temporally propagate trimap labels from sparsely defined key frames to estimate trimaps for the entire video sequence. Trimap is a fundamental requirement for digital image and video matting approaches. Statistical inference is coupled with Bayesian statistics to allow robust trimap labelling in the presence of shadows, illumination variation and overlap between the foreground and background appearance. Results demonstrate that trimaps are sufficiently accurate to allow high quality video matting using existing natural image matting algorithms. Quantitative evaluation against ground-truth demonstrates that the approach achieves accurate matte estimation with less amount of user interaction compared to the state-of-the-art techniques.


international conference on image processing | 2010

Stereoscopic content production of complex dynamic scenes using a wide-baseline monoscopic camera set-up

Jean-Yves Guillemaut; Muhammad Sarim; Adrian Hilton

Conventional stereoscopic video content production requires use of dedicated stereo camera rigs which is both costly and lacking video editing flexibility. In this paper, we propose a novel approach which only requires a small number of standard cameras sparsely located around a scene to automatically convert the monocular inputs into stereoscopic streams. The approach combines a probabilistic spatio-temporal segmentation framework with a state-of-the-art multi-view graph-cut reconstruction algorithm, thus providing full control of the stereoscopic settings at render time. Results with studio sequences of complex human motion demonstrate the suitability of the method for high quality stereoscopic content generation with minimum user interaction.


conference on visual media production | 2009

Wide-Baseline Matte Propagation for Indoor Scenes

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut

Digital image matting is the process of extracting foreground objects from an image. This is extremely challenging for natural images and videos because of its ill posed nature. Initial user interaction is required to aid the algorithms in identifying the definite foreground and background regions. Recently techniques have been developed to estimate the alpha matte of an image using multi-view images of a foreground object. However these algorithms are only capable of handling narrow baseline views having small intensity and structural variations in the foreground. In this paper, we propose a novel non-parametric approach to generate alpha matte for wide-baseline multi-view images having different inter-view foreground appearance.


british machine vision conference | 2009

Non-parametric patch based video matting

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut

In computer vision, matting is the process of accurate foreground estimation in images and videos. In this paper we presents a novel patch based approach to video matting relying on non-parametric statistics to represent image variations in appearance. This overcomes the limitation of parametric algorithms which only rely on strong colour correlation between the nearby pixels. Initially we construct a clean background by utilising the foreground object’s movement across the background. For a given frame, a trimap is constructed using the background and the last frame’s trimap. A patch-based approach is used to estimate the foreground colour for every unknown pixel and finally the alpha matte is extracted. Quantitative evaluation shows that the technique performs better, in terms of the accuracy and the required user interaction, than the current state-of-the-art parametric approaches.


International Journal of Pattern Recognition and Artificial Intelligence | 2017

“Cellular-Cut”-Interactive n-Dimensional Image Segmentation Using Cellular Automata

Muhammad Ashraf; Muhammad Sarim; Abdul Basit Shaikh

Interactive segmentation of images has become an integral part of image processing applications. Several graph based segmentation techniques have been developed, which depend upon global minimization of the energy cost function. An adequate scheme of interactive segmentation still needs a skilled initialization of regions with user-defined seeds pixels distributed over the entire image. We propose an iterative segmentation technique based on Cellular Automaton which focuses to reduce the user efforts required to provide initialization. The existing algorithms based on Cellular Automaton only use local smoothness term in label propagation making them highly sensitive to user-defined seeds pixels. To reduce the sensitivity towards initial user definition of regions, global constraints are introduced along with local information to propagate labels. The results obtained are comparable to the state-of-the-art interactive segmentation techniques on a standard dataset.


conference on visual media production | 2010

Multiple View Wide-Baseline Trimap Propagation for Natural Video Matting

Muhammad Sarim; Adrian Hilton; Jean-Yves Guillemaut; Hansung Kim; Takeshi Takai

This paper presents a method to estimate alpha mattes for video sequences of the same foreground scene from wide-baseline views given sparse key-frame trimaps in a single view. A statistical inference framework is introduced for spatio-temporal propagation of high-confidence trimap labels between video sequences without a requirement for correspondence or camera calibration and motion estimation. Multiple view trimap propagation integrates appearance information between views and over time to achieve robust labelling in the presence of shadows, changes in appearance with view point and overlap between foreground and background appearance. Results demonstrate that trimaps are sufficiently accurate to allow high-quality video matting using existing single view natural image matting algorithms. Quantitative evaluation against ground-truth demonstrates that the approach achieves accurate matte estimation for camera views separated by up to 180°, with the same amount of manual interaction required for conventional single view video matting

Collaboration


Dive into the Muhammad Sarim's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Adnan Nadeem

Federal Urdu University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Kamran Ahsan

Staffordshire University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Abdul Salam

Federal Urdu University

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge