Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Dushyant Mehta is active.

Publication


Featured researches published by Dushyant Mehta.


international conference on computer graphics and interactive techniques | 2017

VNect: real-time 3D human pose estimation with a single RGB camera

Dushyant Mehta; Srinath Sridhar; Oleksandr Sotnychenko; Helge Rhodin; Mohammad Shafiei; Hans-Peter Seidel; Weipeng Xu; Dan Casas; Christian Theobalt

We present the first real-time method to capture the full global 3D skeletal pose of a human in a stable, temporally consistent manner using a single RGB camera. Our method combines a new convolutional neural network (CNN) based pose regressor with kinematic skeleton fitting. Our novel fully-convolutional pose formulation regresses 2D and 3D joint positions jointly in real time and does not require tightly cropped input frames. A real-time kinematic skeleton fitting method uses the CNN output to yield temporally stable 3D global pose reconstructions on the basis of a coherent kinematic skeleton. This makes our approach the first monocular RGB method usable in real-time applications such as 3D character control---thus far, the only monocular methods for such applications employed specialized RGB-D cameras. Our methods accuracy is quantitatively on par with the best offline 3D monocular RGB pose estimation methods. Our results are qualitatively comparable to, and sometimes better than, results from monocular RGB-D approaches, such as the Kinect. However, we show that our approach is more broadly applicable than RGB-D solutions, i.e., it works for outdoor scenes, community videos, and low quality commodity RGB cameras.


international conference on computer vision | 2017

Real-Time Hand Tracking under Occlusion from an Egocentric RGB-D Sensor

Franziska Mueller; Dushyant Mehta; Oleksandr Sotnychenko; Srinath Sridhar; Dan Casas; Christian Theobalt

We present an approach for real-time, robust and accurate hand pose estimation from moving egocentric RGB-D cameras in cluttered real environments. Existing methods typically fail for hand-object interactions in cluttered scenes imaged from egocentric viewpoints—common for virtual or augmented reality applications. Our approach uses two subsequently applied Convolutional Neural Networks (CNNs) to localize the hand and regress 3D joint locations. Hand localization is achieved by using a CNN to estimate the 2D position of the hand center in the input, even in the presence of clutter and occlusions. The localized hand position, together with the corresponding input depth value, is used to generate a normalized cropped image that is fed into a second CNN to regress relative 3D hand joint locations in real time. For added accuracy, robustness and temporal stability, we refine the pose estimates using a kinematic pose tracking energy. To train the CNNs, we introduce a new photorealistic dataset that uses a merged reality approach to capture and synthesize large amounts of annotated data of natural hand interaction in cluttered scenes. Through quantitative and qualitative evaluation, we show that our method is robust to self-occlusion and occlusions by objects, particularly in moving egocentric perspectives.


Computer Graphics Forum | 2017

Deep Shading: Convolutional Neural Networks for Screen Space Shading

Oliver Nalbach; Elena Arabadzhiyska; Dushyant Mehta; Hans-Peter Seidel; Tobias Ritschel

In computer vision, convolutional neural networks (CNNs) achieve unprecedented performance for inverse problems where RGB pixel appearance is mapped to attributes such as positions, normals or reflectance. In computer graphics, screen space shading has boosted the quality of real‐time rendering, converting the same kind of attributes of a virtual scene back to appearance, enabling effects like ambient occlusion, indirect light, scattering and many more. In this paper we consider the diagonal problem: synthesizing appearance from given per‐pixel attributes using a CNN. The resulting Deep Shading renders screen space effects at competitive quality and speed while not being programmed by human experts but learned from example images.


international conference on computer graphics and interactive techniques | 2018

MonoPerfCap: Human Performance Capture from Monocular Video

Weipeng Xu; Avishek Chatterjee; Michael Zollhoefer; Helge Rhodin; Dushyant Mehta; Hans-Peter Seidel; Christian Theobalt

We present the first marker-less approach for temporally coherent 3D performance capture of a human with general clothing from monocular video. Our approach reconstructs articulated human skeleton motion as well as medium-scale non-rigid surface deformations in general scenes. Human performance capture is a challenging problem due to the large range of articulation, potentially fast motion, and considerable non-rigid deformations, even from multi-view data. Reconstruction from monocular video alone is drastically more challenging, since strong occlusions and the inherent depth ambiguity lead to a highly ill-posed reconstruction problem. We tackle these challenges by a novel approach that employs sparse 2D and 3D human pose detections from a convolutional neural network using a batch-based pose estimation strategy. Joint recovery of per-batch motion allows to resolve the ambiguities of the monocular reconstruction problem based on a low dimensional trajectory subspace. In addition, we propose refinement of the surface geometry based on fully automatically extracted silhouettes to enable medium-scale non-rigid alignment. We demonstrate state-of-the-art performance capture results that enable exciting applications such as video editing and free viewpoint video, previously infeasible from monocular video. Our qualitative and quantitative evaluation demonstrates that our approach significantly outperforms previous monocular methods in terms of accuracy, robustness and scene complexity that can be handled.


international conference on 3d vision | 2017

Monocular 3D Human Pose Estimation in the Wild Using Improved CNN Supervision

Dushyant Mehta; Helge Rhodin; Dan Casas; Pascal Fua; Oleksandr Sotnychenko; Weipeng Xu; Christian Theobalt


arXiv: Computer Vision and Pattern Recognition | 2016

Monocular 3D Human Pose Estimation Using Transfer Learning and Improved CNN Supervision.

Dushyant Mehta; Helge Rhodin; Dan Casas; Oleksandr Sotnychenko; Weipeng Xu; Christian Theobalt


computer vision and pattern recognition | 2018

GANerated Hands for Real-Time 3D Hand Tracking From Monocular RGB

Franziska Mueller; Florian Bernard; Oleksandr Sotnychenko; Dushyant Mehta; Srinath Sridhar; Dan Casas; Christian Theobalt


Archive | 2017

Single-Shot Multi-Person 3D Body Pose Estimation From Monocular RGB Input

Dushyant Mehta; Oleksandr Sotnychenko; Franziska Mueller; Weipeng Xu; Srinath Sridhar; Gerard Pons-Moll; Christian Theobalt


international conference on 3d vision | 2018

Single-Shot Multi-Person 3D Pose Estimation From Monocular RGB

Dushyant Mehta; Oleksandr Sotnychenko; Franziska Mueller; Weipeng Xu; Srinath Sridhar; Gerard Pons-Moll; Christian Theobalt


international conference on computer graphics and interactive techniques | 2017

Monocular 3D Human Pose Estimation In The Wild Using Improved CNN Supervision

Dushyant Mehta; Helge Rhodin; Dan Casas; Pascal Fua; Oleksandr Sotnychenko; Weipeng Xu; Christian Theobalt

Collaboration


Dive into the Dushyant Mehta's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge