Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Sungmin Eum is active.

Publication


Featured researches published by Sungmin Eum.


Archive | 2017

Page-to-Word Extraction from Unconstrained Handwritten Document Images

Pawan Kumar Singh; Sagnik Pal Chowdhury; Shubham Sinha; Sungmin Eum; Ram Sarkar

Extraction of words directly from handwritten document images is still a challenging problem in the development of a complete Optical Character Recognition (OCR) system. In this paper, a robust word extraction scheme is reported. Firstly, applying Harris corner point detection algorithm, key points are generated from the document images which are then clustered using well-known DBSCAN technique. Finally, the boundary of the text words present in the document images are estimated based on the convex hull drawn for each of the clustered key points. The proposed technique is tested on randomly selected 50 images from CMATERdb1database and the success rate is found to be 90.48 % which is equivalent to the state-of-the-art.


british machine vision conference | 2015

JH2R: Joint Homography Estimation for Highlight Removal.

Sungmin Eum; Hyungtae Lee; David S. Doermann

Imagine being in an art museum where there are paintings or pictures held inside glass-frames for protection. There are pieces which you wish to capture using a camera, but you experience difficulties avoiding highlights which are generated by indoor lighting reflected off the glossy surfaces. Similar problems occur when capturing contents off of whiteboards, documents printed on glossy surfaces, objects such as books or CDs with plastic covers. In this work, we address the problem of removing unwanted highlight regions in images generated by reflections of light sources on glossy surfaces. Although there have been efforts made to synthetically fill in the missing regions using the neighboring patterns by applying methods like inpainting [3, 4], it is impossible to recover the missing information in completely saturated regions. Therefore, we need to use multiple images where corresponding regions are not covered by the saturated highlights. Unlike other methods, our method uses the relationship between the highlight regions resulting in more robust removal of saturated highlights. Our method Overview Our method was motivated by a widely acknowledged physical phenomenon referred to as the ‘motion parallax’. Without loss of generality, we can similarly view the relationship between the desired content (e.g., a painting) and the highlights. Since the highlights caused by the light source are the result of the reflection on the glossy surface before they reach the camera, the light source can be modeled to virtually exist on the other side of the content. Note that, the distance from the light source is always larger than the distance from the content (D > d, in Figure 1).


international conference on image processing | 2014

Sharpness-aware document image mosaicing using graphcuts

Sungmin Eum; David S. Doermann

There are numerous types of documents which are difficult to scan or capture in a single pass due to their physical size or the size of their content. One possible solution that has been proposed is mosaicing multiple overlapping images to capture the complete document. In this paper, we present a novel Graphcut-based document image mosaicing method which seeks to overcome the known limitations of the previous approaches. First, our method does not require any prior knowledge of the content of the given document images, making it more widely applicable and robust. Second, information regarding the geometrical disposition between the overlapping images is exploited to minimize the errors at the boundary regions. Third, our method incorporates a sharpness measure which induces cut generation in a way that results in the mosaic including the sharpest pixels. Our method is shown to outperform previous methods, both quantitatively and qualitatively.


Signal Processing, Sensor/Information Fusion, and Target Recognition XXVII | 2018

Going deeper with CNN in malicious crowd event classification

Hyungtae Lee; Sungmin Eum; Heesung Kwon

Terror attacks are often targeted towards the civilians gathered in one location (e.g., Boston Marathon bombing). Distinguishing such ’malicious’ scenes from the ’normal’ ones, which are semantically different, is a difficult task as both scenes contain large groups of people with high visual similarity. To overcome the difficulty, previous methods exploited various contextual information, such as language-driven keywords or relevant objects. Although useful, they require additional human effort or dataset. In this paper, we show that using more sophisticated and deeper Convolutional Neural Networks (CNNs) can achieve better classification accuracy even without using any additional information outside the image domain. We have conducted a comparative study where we train and compare seven different CNN architectures (AlexNet, VGG-M, VGG16, GoogLeNet, ResNet- 50, ResNet-101, and ResNet-152). Based on the experimental analyses, we found out that deeper networks typically show better accuracy, and that GoogLeNet is the most favorable among the seven architectures for the task of malicious event classification.


international conference on pattern recognition | 2016

Content selection using frontalness evaluation of multiple frames

Sungmin Eum; David S. Doermann

This paper addresses the problem of selecting instances of a planar object in a video or from a set of images based on an evaluation of its “frontalness”. We introduce the idea of “evaluating the frontalness” by computing how close the objects surface normal aligns with the optical axis of a camera. The unique and novel aspect of our method is that unlike previous planar object pose estimation methods, our method does not require the true frontal image as a reference. The intuition is that a true frontal image can be used to produce other non-frontal images by perspective projection, while the non-frontal images have limited ability to produce other non-frontal images. We show that this intuition of comparing ‘frontal’ and ‘non-frontal’ can be extended to comparing ‘more frontal’ and ‘less frontal’ images. Based on this observation, our method estimates the relative frontalness of an image by exploiting the objective space error. We also propose the usage of K-invariant space to evaluate the frontalness even when the camera intrinsic parameters are unknown (e.g., images/videos from the web). We show that our method outperforms the homography decomposition-based method which also does not require reference images. In addition, a qualitative evaluation is carried out to show that our method can be applied in selecting the most frontal characters from a set of images captured in various viewpoints.


arXiv: Computer Vision and Pattern Recognition | 2016

Joint Deep Exploitation of Semantic Keywords and Visual Features for Malicious Crowd Image Classification.

Joel Levis; Hyungtae Lee; Heesung Kwon; James Michaelis; Michael Kolodny; Sungmin Eum


international conference on acoustics, speech, and signal processing | 2018

Exploitation of Semantic Keywords for Malicious Event Classification.

Hyungtae Lee; Sungmin Eum; Joel Levis; Heesung Kwon; James Michaelis; Michael Kolodny


arXiv: Computer Vision and Pattern Recognition | 2018

Cross-domain CNN for Hyperspectral Image Classification.

Hyungtae Lee; Sungmin Eum; Heesung Kwon


arXiv: Computer Vision and Pattern Recognition | 2018

Object and Text-guided Semantics for CNN-based Activity Recognition.

Sungmin Eum; Christopher Reale; Heesung Kwon; Claire Bonial; Clare R. Voss


arXiv: Computer Vision and Pattern Recognition | 2017

ME R-CNN: Multi-Expert R-CNN for Object Detection

Hyungtae Lee; Sungmin Eum; Heesung Kwon

Collaboration


Dive into the Sungmin Eum's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sagnik Pal Chowdhury

Indian Institute of Engineering Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Shubham Sinha

Indian Institute of Engineering Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Claire Bonial

University of Colorado Boulder

View shared research outputs
Researchain Logo
Decentralizing Knowledge