Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Ozan Sener is active.

Publication


Featured researches published by Ozan Sener.


computer vision and pattern recognition | 2016

3D Semantic Parsing of Large-Scale Indoor Spaces

Iro Armeni; Ozan Sener; Amir Roshan Zamir; Helen Jiang; Ioannis Brilakis; Silvio Savarese

In this paper, we propose a method for semantic parsing the 3D point cloud of an entire building using a hierarchical approach: first, the raw data is parsed into semantically meaningful spaces (e.g. rooms, etc) that are aligned into a canonical reference coordinate system. Second, the spaces are parsed into their structural and building elements (e.g. walls, columns, etc). Performing these with a strong notation of global 3D space is the backbone of our method. The alignment in the first step injects strong 3D priors from the canonical coordinate system into the second step for discovering elements. This allows diverse challenging scenarios as man-made indoor spaces often show recurrent geometric patterns while the appearance features can change drastically. We also argue that identification of structural elements in indoor spaces is essentially a detection problem, rather than segmentation which is commonly used. We evaluated our method on a new dataset of several buildings with a covered area of over 6, 000m2 and over 215 million points, demonstrating robust results readily useful for practical applications.


acm multimedia | 2012

Error-tolerant interactive image segmentation using dynamic and iterated graph-cuts

Ozan Sener; Kemal Ugur; A. Aydin Alatan

Efficient and accurate interactive image segmentation have significant importance in many multimedia applications. For mobile touchscreen-based applications, efficiency is more crucial. Moreover, due to small screens of the mobile devices, error tolerance is also a crucial factor. In this paper, a method for interactive image segmentation, tailored for mobile touch screen devices, is proposed. As an interaction methodology, coloring is presented. An automatic stroke-error correction methodology to correct the inaccurate user interaction is also proposed. For the efficient computation of the solution, a novel dynamic and iterative graph-cut solution is formulated. Efficiency and error tolerance of the proposed method are tested by using various sample images. Subjective evaluation of the interactive segmentation algorithms for mobile-touch screen is also performed. Indeed, for the challenging examples, the superior performance of the proposed method is obtained by the experiments.


robotics science and systems | 2015

rCRF: Recursive Belief Estimation over CRFs in RGB-D Activity Videos

Ozan Sener; Ashutosh Saxena

For assistive robots, anticipating the future actions of humans is an essential task. This requires modelling both the evolution of the activities over time and the rich relationships between humans and the objects. Since the future activities of humans are quite ambiguous, robots need to assess all the future possibilities in order to choose an appropriate action. Therefore, a successful anticipation algorithm needs to compute all plausible future activities and their corresponding probabilities. In this paper, we address the problem of efficiently computing beliefs over future human activities from RGB-D videos. We present a new recursive algorithm that we call Recursive Conditional Random Field (rCRF) which can compute an accurate belief over a temporal CRF model. We use the rich modelling power of CRFs and describe a computationally tractable inference algorithm based on Bayesian filtering and structured diversity. In our experiments, we show that incorporating belief, computed via our approach, significantly outperforms the stateof-the-art methods, in terms of accuracy and computation time.


international conference on image processing | 2012

Interactive 2D-3D image conversion for mobile devices

Yagiz Aksoy; Ozan Sener; A. Aydin Alatan; Kemal Ugur

We propose a complete still image based 2D-3D mobile conversion system for touch screen use. The system consists of interactive segmentation followed by 3D rendering. The interactive segmentation is conducted dynamically by color Gaussian mixture model updates and dynamic-iterative graph-cut. A coloring gesture is used to guide the way and entertain the user during the process. Output of the image segmentation is then fed to the 3D rendering stage of the system. For rendering stage, two novel improvements are proposed to handle holes resulting from depth image based rendering process. These improvements are also expected to enhance the 3D perception. These two methods are subjectively tested and their results are presented.


IEEE Transactions on Multimedia | 2014

Efficient MRF Energy Propagation for Video Segmentation via Bilateral Filters

Ozan Sener; Kemal Ugur; A. Aydin Alatan

Segmentation of an object from a video is a challenging task in multimedia applications. Depending on the application, automatic or interactive methods are desired; however, regardless of the application type, efficient computation of video object segmentation is crucial for time-critical applications; specifically, mobile and interactive applications require near real-time efficiencies. In this paper, we address the problem of video segmentation from the perspective of efficiency. We initially redefine the problem of video object segmentation as the propagation of MRF energies along the temporal domain. For this purpose, a novel and efficient method is proposed to propagate MRF energies throughout the frames via bilateral filters without using any global texture, color or shape model. Recently presented bi-exponential filter is utilized for efficiency, whereas a novel technique is also developed to dynamically solve graph-cuts for varying, non-lattice graphs in general linear filtering scenario. These improvements are experimented for both automatic and interactive video segmentation scenarios. Moreover, in addition to the efficiency, segmentation quality is also tested both quantitatively and qualitatively. Indeed, for some challenging examples, significant time efficiency is observed without loss of segmentation quality.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2018

Watch-n-Patch: Unsupervised Learning of Actions and Relations

Chenxia Wu; Jiemi Zhang; Ozan Sener; Bart Selman; Silvio Savarese; Ashutosh Saxena

There is a large variation in the activities that humans perform in their everyday lives. We consider modeling these composite human activities which comprises multiple basic level actions in a completely unsupervised setting. Our model learns high-level co-occurrence and temporal relations between the actions. We consider the video as a sequence of short-term action clips, which contains human-words and object-words. An activity is about a set of action-topics and object-topics indicating which actions are present and which objects are interacting with. We then propose a new probabilistic model relating the words and the topics. It allows us to model long-range action relations that commonly exist in the composite activities, which is challenging in previous works. We apply our model to the unsupervised action segmentation and clustering, and to a novel application that detects forgotten actions, which we call action patching. For evaluation, we contribute a new challenging RGB-D activity video dataset recorded by the new Kinect v2, which contains several human daily activities as compositions of multiple actions interacting with different objects. Moreover, we develop a robotic system that watches and reminds people using our action patching algorithm. Our robotic setup can be easily deployed on any assistive robots.


neural information processing systems | 2016

Learning Transferrable Representations for Unsupervised Domain Adaptation

Ozan Sener; Hyun Oh Song; Ashutosh Saxena; Silvio Savarese


arXiv: Artificial Intelligence | 2014

RoboBrain: Large-Scale Knowledge Engine for Robots.

Ashutosh Saxena; Ashesh Jain; Ozan Sener; Aditya Jami; Dipendra Kumar Misra; Hema Swetha Koppula


international conference on computer vision | 2015

Unsupervised Semantic Parsing of Video Collections

Ozan Sener; Amir Roshan Zamir; Silvio Savarese; Ashutosh Saxena


international conference on learning representations | 2018

Active Learning for Convolutional Neural Networks: A Core-Set Approach

Ozan Sener; Silvio Savarese

Collaboration


Dive into the Ozan Sener's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

A. Aydin Alatan

Middle East Technical University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Hyun Oh Song

University of California

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge