Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Kihong Park is active.

Publication


Featured researches published by Kihong Park.


european conference on computer vision | 2016

Unified Depth Prediction and Intrinsic Image Decomposition from a Single Image via Joint Convolutional Neural Fields

Seungryong Kim; Kihong Park; Kwanghoon Sohn; Stephen Lin

We present a method for jointly predicting a depth map and intrinsic images from single-image input. The two tasks are formulated in a synergistic manner through a joint conditional random field (CRF) that is solved using a novel convolutional neural network (CNN) architecture, called the joint convolutional neural field (JCNF) model. Tailored to our joint estimation problem, JCNF differs from previous CNNs in its sharing of convolutional activations and layers between networks for each task, its inference in the gradient domain where there exists greater correlation between depth and intrinsic images, and the incorporation of a gradient scale network that learns the confidence of estimated gradients in order to effectively balance them in the solution. This approach is shown to surpass state-of-the-art methods both on single-image depth estimation and on intrinsic image decomposition.


international conference on pattern recognition | 2016

Multi-spectral pedestrian detection based on accumulated object proposal with fully convolutional networks

Hangil Choi; Seungryong Kim; Kihong Park; Kwanghoon Sohn

This paper presents a method for detecting a pedestrian by leveraging multi-spectral image pairs. Our approach is based on the observation that a multi-spectral image, especially far-infrared (FIR) image, enables us to overcome inherent limitations for pedestrian detection under challenging circumstances, such as even dark environments. For that task, multi-spectral color-FIR image pairs are used in a synergistic manner for pedestrian detection through deep convolutional neural networks (CNNs) learning and support vector regression (SVR). For inferring the confidence of a pedestrian, we first learn CNNs between color images (or FIR images) and bounding box annotations of pedestrians, respectively. Furthermore, for each object proposal, we extract intermediate activation features from network, and learn the probability of pedestrian using SVR. To improve the detection performance, the learned probability of pedestrian for each proposal is accumulated on the image domain. Based on the pedestrian confidence estimated from each network and accumulated pedestrian probabilities, the most probable pedestrian is finally localized among object proposal candidates. Thanks to its high robustness of multi-spectral imaging in dark environments and its high discriminative power of deep CNNs, our framework is shown to surpass state-of-the-art pedestrian detection methods on multi-spectral pedestrian benchmark.


Pattern Recognition | 2018

Unified multi-spectral pedestrian detection based on probabilistic fusion networks

Kihong Park; Seungryong Kim; Kwanghoon Sohn

Abstract Despite significant progress in machine learning, pedestrian detection in the real-world is still regarded as one of the challenging problems, limited by occluded appearances, cluttered backgrounds, and bad visibility at night. This has caused detection approaches using multi-spectral sensors such as color and thermal which could be complementary to each other. In this paper, we propose a novel sensor fusion framework for detecting pedestrians even in challenging real-world environments. We design a convolutional neural network (CNN) architecture that consists of three-branch detection models taking different modalities as inputs. Unlike existing methods, we consider all detection probabilities from each modality in a unified CNN framework and selectively use them through a channel weighting fusion (CWF) layer to maximize the detection performance. An accumulated probability fusion (APF) layer is also introduced to combine probabilities from different modalities at the proposal-level. We formulate these sub-networks into a unified network, so that it is possible to train the whole network in an end-to-end manner. Our extensive evaluation demonstrates that the proposed method outperforms the state-of-the-art methods on the challenging KAIST, CVC-14, and DIML multi-spectral pedestrian datasets.


british machine vision conference | 2015

Randomized Global Transformation Approach for Dense Correspondence.

Kihong Park; Seungryong Kim; Seungchul Ryu; Kwanghoon Sohn

This paper describes a randomized global transformation approach to estimate dense correspondence for image pairs taken under challengingly different photometric and geometric conditions. Our approach assumes that a correspondence field consists of piecewise parametric transformation model. While conventional approaches consider large search space including flow and geometric fields exhaustively, our approach is based on an inference of optimal global transformation model from transformation candidates. To build a reliable global transformation hypothesis, we build optimal global transformation candidates with a randomized manner from an initial sparse feature correspondence, followed by a transformation clustering. Furthermore, the optimal global transformation is estimated as a cost filtering scheme with fast edge-aware filtering to provide a geometrical smoothness. Experiments demonstrate outstanding performance of our approach in terms of correspondence accuracy and computational complexity.


IEEE Transactions on Image Processing | 2017

Modality-Invariant Image Classification Based on Modality Uniqueness and Dictionary Learning

Seungryong Kim; Rui Cai; Kihong Park; Sunok Kim; Kwanghoon Sohn

We present a unified framework for the image classification of image sets taken under varying modality conditions. Our method is motivated by a key observation that the image feature distribution is simultaneously influenced by the semantic-class and the modality category label, which limits the performance of conventional methods for that task. With this insight, we introduce modality uniqueness as a discriminative weight that divides each modality cluster from all other clusters. By leveraging the modality uniqueness, our framework is formulated as unsupervised modality clustering and classifier learning based on modality-invariant similarity kernel. Specifically, in the assignment step, each training image is first assigned to the most similar cluster according to its modality. In the update step, based on the current cluster hypothesis, the modality uniqueness and the sparse dictionary are updated. These two steps are formulated in an iterative manner. Based on the final clusters, a modality-invariant marginalized kernel is then computed, where the similarities between the reconstructed features of each modality are aggregated across all clusters. Our framework enables the reliable inference of semantic-class category for an image, even across large photometric variations. Experimental results show that our method outperforms conventional methods on various benchmarks, such as landmark identification under severely varying weather conditions, domain-adapting image classification, and RGB and near-infrared image classification.We present a unified framework for the image classification of image sets taken under varying modality conditions. Our method is motivated by a key observation that the image feature distribution is simultaneously influenced by the semantic-class and the modality category label, which limits the performance of conventional methods for that task. With this insight, we introduce modality uniqueness as a discriminative weight that divides each modality cluster from all other clusters. By leveraging the modality uniqueness, our framework is formulated as unsupervised modality clustering and classifier learning based on modality-invariant similarity kernel. Specifically, in the assignment step, each training image is first assigned to the most similar cluster according to its modality. In the update step, based on the current cluster hypothesis, the modality uniqueness and the sparse dictionary are updated. These two steps are formulated in an iterative manner. Based on the final clusters, a modality-invariant marginalized kernel is then computed, where the similarities between the reconstructed features of each modality are aggregated across all clusters. Our framework enables the reliable inference of semantic-class category for an image, even across large photometric variations. Experimental results show that our method outperforms conventional methods on various benchmarks, such as landmark identification under severely varying weather conditions, domain-adapting image classification, and RGB and near-infrared image classification.


asia pacific signal and information processing association annual summit and conference | 2016

Homography flow for dense correspondences

Kihong Park; Seungryoung Kim; Kwanghoon Sohn

We present a unified framework for dense correspondence estimation, called Homography flow, to handle large photometric and geometric deformations in an efficient manner. Our algorithm is inspired by recent successes of the sparse to dense framework. The main intuition is that dense flows located in same plane can be represented as a single geometric transform. Tailored to dense correspondence task, the Homography flow differs from previous methods in the flow domain clustering and the trilateral interpolation. By estimating and propagating sparsely estimated transforms, dense flow field is estimated with very low computation time. The Homography flow highly improves the performance of dense correspondences, especially in flow discontinuous area. Experimental results on challenging image pairs show that our approach suppresses the state-of-the-art algorithms in both accuracy and computation time.


electronic imaging | 2015

Statistical approach for supervised codeword selection

Kihong Park; Seungchul Ryu; Seungryong Kim; Kwanghoon Sohn

Bag-of-words (BoW) is one of the most successful methods for object categorization. This paper proposes a statistical codeword selection algorithm where the best subset is selected from the initial codewords based on the statistical characteristics of codewords. For this purpose, we defined two types of codeword-confidences: cross- and within-category confidences. The cross- and within-category confidences eliminate indistinctive codewords across categories and inconsistent codewords within each category, respectively. An informative subset of codewords is then selected based on these two codeword-confidences. The experimental evaluation for a scene categorization dataset and a Caltech-101 dataset shows that the proposed method improves the categorization performance up to 10% in terms of error rate reduction when cooperated with BoW, sparse coding (SC), and locality-constrained liner coding (LLC). Furthermore, the codeword size is reduced by 50% leading a low computational complexity.


international conference on robotics and automation | 2018

High-Precision Depth Estimation with the 3D LiDAR and Stereo Fusion

Kihong Park; Seungryong Kim; Kwanghoon Sohn


computer vision and pattern recognition | 2018

Learning Descriptor, Confidence, and Depth Estimation in Multi-View Stereo

Sungil Choi; Seungryong Kim; Kihong Park; Kwanghoon Sohn


international conference on image processing | 2017

Pedestrian proposal generation using depth-aware scale estimation

Kihong Park; Seungryong Kim; Kwanghoon Sohn

Collaboration


Dive into the Kihong Park's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge