Stephan Würmlin
ETH Zurich
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Stephan Würmlin.
international conference on computer graphics and interactive techniques | 2003
Markus H. Gross; Stephan Würmlin; Martin Naef; Edouard Lamboray; Christian P. Spagno; Andreas Kunz; Esther Koller-Meier; Tomáš Svoboda; Luc Van Gool; Silke Lang; Kai Strehlke; Andrew Vande Moere; Oliver G. Staadt
In this paper, we report ongoing work in a new project, The Blue-C. The goal of this project is to build a collaborative, immersive virtual environment which will eventually integrate real humans, captured by a set of video cameras. Two Blue-C.s will be interconnected via a high-speed network. This will allow for bi-directional collaboration and interaction between two persons sharing virtual spaces. The video streams are used for both texture and geometry extraction. We will generate a 3-D light field inlay enriched with the reconstructed geometry, which will be integrated into the virtual environment. The design and construction of the Blue-C. environment, including both hardware and software, is an interdisciplinary effort with participants from the departments of computer science, architecture, product development, and electrical engineering. Parallel to the development of the core system, we are designing new applications in the areas of computer aided architectural design, product reviewing, and medicine, which will highlight the versatility of the Blue-C.
Computers & Graphics | 2004
Stephan Würmlin; Edouard Lamboray; Markus H. Gross
Abstract We present 3D video fragments, a dynamic point sample framework for real-time free-viewpoint video. By generalizing 2D video pixels towards 3D irregular point samples we combine the simplicity of conventional 2D video processing with the power of more complex polygonal representations for free-viewpoint video. We propose a differential update scheme exploiting the spatio-temporal coherence of the video streams of multiple cameras. Updates are issued by operators such as inserts and deletes accounting for changes in the input video images. The operators from multiple cameras are processed, merged into a 3D video stream and transmitted to a remote site. We also introduce a novel concept for camera control which dynamically selects the set of relevant cameras for reconstruction. Moreover, it adapts to the processing load and rendering platform. Our framework is generic in the sense that it works with any real-time 3D reconstruction method which extracts depth from images. The video renderer displays free-viewpoint videos using an efficient point-based splatting scheme and makes use of state-of-the-art vertex and pixel processing hardware for real-time visual processing.
pacific conference on computer graphics and applications | 2002
Stephan Würmlin; Edouard Lamboray; Oliver G. Staadt; Markus H. Gross
We present the 3D video recorder, a system capable of recording, processing, and playing three-dimensional video from multiple points of view. We first record 2D video streams from several synchronized digital video cameras and store pre-processed images to disk. An off-line processing stage converts these images into a time-varying three-dimensional hierarchical point-based data structure and stores this 3D video to disk. We show how we can trade-off 3D video quality with processing performance and devise efficient compression and coding schemes for our novel 3D video representation. A typical sequence is encoded at less than 7 megabits per second at a frame rate of 8.5 frames per second. The 3D video player decodes and renders 3D videos from hard-disk in real-time, providing interaction features known from common video cassette recorders, like variable-speed forward and reverse, and slow motion. 3D video playback can be enhanced with novel 3D video effects such as freeze-and-rotate and arbitrary scaling. The player builds upon point-based rendering techniques and is thus capable of rendering high-quality images in real-time. Finally, we demonstrate the 3D video recorder on multiple real-life video sequences.
The Visual Computer | 2005
Michael Waschbüsch; Stephan Würmlin; Daniel Cotting; Filip Sadlo; Markus H. Gross
In this paper we present a scalable 3D video framework for capturing and rendering dynamic scenes. The acquisition system is based on multiple sparsely placed 3D video bricks, each comprising a projector, two grayscale cameras, and a color camera. Relying on structured light with complementary patterns, texture images and pattern-augmented views of the scene are acquired simultaneously by time-multiplexed projections and synchronized camera exposures. Using space–time stereo on the acquired pattern images, high-quality depth maps are extracted, whose corresponding surface samples are merged into a view-independent, point-based 3D data structure. This representation allows for effective photo-consistency enforcement and outlier removal, leading to a significant decrease of visual artifacts and a high resulting rendering quality using EWA volume splatting. Our framework and its view-independent representation allow for simple and straightforward editing of 3D video. In order to demonstrate its flexibility, we show compositing techniques and spatiotemporal effects.
eurographics | 2004
Michael Waschbüsch; Markus H. Gross; Felix Eberhard; Edouard Lamboray; Stephan Würmlin
decomposition of the point set and thus easily allows for progressive decoding. Our method is generic in the sense that it can handle arbitrary point attributes using attribute-specific coding operations. Furthermore, no resampling of the model is needed and thus we do not introduce additional smoothing artifacts. We provide coding operators for the point position, normal and color. Particularly, by transforming the point positions into a local reference frame, we exploit the fact that all point samples are living on a surface. Our framework enables for compressing both geometry and appearance of the model in a unified manner. We show the performance of our framework on a diversity of point-based models.
Computer Graphics Forum | 2010
Marcel Germann; Alexander Hornung; Richard Keiser; Remo Ziegler; Stephan Würmlin; Markus H. Gross
We present a novel representation and rendering method for free‐viewpoint video of human characters based on multiple input video streams. The basic idea is to approximate the articulated 3D shape of the human body using a subdivision into textured billboards along the skeleton structure. Billboards are clustered to fans such that each skeleton bone contains one billboard per source camera. We call this representation articulated billboards.
Computer Graphics Forum | 2003
Stephan Würmlin; Edouard Lamboray; Oliver G. Staadt; Markus H. Gross
We present the 3D Video Recorder, a system capable of recording, processing, and playing three‐dimensional video from multiple points of view. We first record 2D video streams from several synchronized digital video cameras and store pre‐processed images to disk. An off‐line processing stage converts these images into a time‐varying 3D hierarchical point‐based data structure and stores this 3D video to disk. We show how we can trade‐off 3D video quality with processing performance and devise efficient compression and coding schemes for our novel 3D video representation. A typical sequence is encoded at less than 7 Mbps at a frame rate of 8.5 frames per second. The 3D video player decodes and renders 3D videos from hard‐disk in real‐time, providing interaction features known from common video cassette recorders, like variable‐speed forward and reverse, and slow motion. 3D video playback can be enhanced with novel 3D video effects such as freeze‐and‐rotate and arbitrary scaling. The player builds upon point‐based rendering techniques and is thus capable of rendering high‐quality images in real‐time. Finally, we demonstrate the 3D Video Recorder on multiple real‐life video sequences.
Signal Processing-image Communication | 2007
Michael Waschbüsch; Stephan Würmlin; Daniel Cotting; Markus H. Gross
This paper presents a point-sampled approach for capturing 3D video footage and subsequent re-rendering of real-world scenes. The acquisition system is composed of multiple sparsely placed 3D video bricks. The bricks contain a low-cost projector, two grayscale cameras and a high-resolution color camera. To improve on depth calculation we rely on structured light patterns. Texture images and pattern-augmented views of the scene are acquired simultaneously by time multiplexed projections of complementary patterns and synchronized camera exposures. High-resolution depth maps are extracted using depth-from-stereo algorithms performed on the acquired pattern images. The surface samples corresponding to the depth values are merged into a view-independent, point-based 3D data structure. This representation allows for efficient post-processing algorithms and leads to a high resulting rendering quality using enhanced probabilistic EWA volume splatting. In this paper, we focus on the 3D video acquisition system and necessary image and video processing techniques.
IEEE Transactions on Visualization and Computer Graphics | 2005
Edouard Lamboray; Stephan Würmlin; Markus H. Gross
In this paper, we discuss data transmission in telepresence environments for collaborative virtual reality applications. We analyze data streams in the context of networked virtual environments and classify them according to their traffic characteristics. Special emphasis is put on geometry-enhanced (3D) video. We review architectures for real-time 3D video pipelines and derive theoretical bounds on the minimal system latency as a function of the transmission and processing delays. Furthermore, we discuss bandwidth issues of differential update coding for 3D video. In our telepresence system - the blue-c - we use a point-based 3D video technology which allows for differentially encoded 3D representations of human users. While we discuss the considerations which lead to the design of our three-stage 3D video pipeline, we also elucidate some critical implementation details regarding decoupling of acquisition, processing and rendering frame rates, and audio/video synchronization. Finally, we demonstrate the communication and networking features of the blue-c system in its full deployment. We show how the system can possibly be controlled to face processing or networking bottlenecks by adapting the multiple system components like audio, application data, and 3D video.
Computer Graphics Forum | 2007
Michael Waschbüsch; Stephan Würmlin; Markus H. Gross
3D video billboard clouds reconstruct and represent a dynamic three‐dimensional scene using displacement‐mapped billboards. They consist of geometric proxy planes augmented with detailed displacement maps and combine the generality of geometry‐based 3D video with the regularization properties of image‐based 3D video. 3D video billboards are an image‐based representation placed in the disparity space of the acquisition cameras and thus provide a regular sampling of the scene with a uniform error model. We propose a general geometry filtering framework which generates time‐coherent models and removes reconstruction and quantization noise as well as calibration errors. This replaces the complex and time‐consuming sub‐pixel matching process in stereo reconstruction with a bilateral filter. Rendering is performed using a GPU‐accelerated algorithm which generates consistent view‐dependent geometry and textures for each individual frame. In addition, we present a semi‐automatic approach for modeling dynamic three‐dimensional scenes with a set of multiple 3D video billboards clouds.