Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Hwa-Mook Yoon is active.

Publication


Featured researches published by Hwa-Mook Yoon.


The Journal of the Korea Contents Association | 2011

Development of Web Crawler for Archiving Web Resources

Kwang-Young Kim; Wongoo Lee; Min-Ho Lee; Hwa-Mook Yoon; Sung-Ho Shin

There are no way of collection, preservation and utilization for web resources after the service is terminated and is gone. However, these Web resources, regardless of the importance of periodically or aperiodically updated or have been destroyed. Therefore, to collect and preserve Web resources Web archive is being emphasized. Web resources collected periodically in order to develop Web archiving crawlers only was required. In this study, from the collection of Web resources to be used for archiving existing web crawlers to analyze the strengths and weaknesses. We have developed web archiving systems for the best collection of web resources.


The Journal of the Korea Contents Association | 2011

Comparison and Analysis of Science and Technology Journal Metadata

Min-Ho Lee; Wongoo Lee; Hwa-Mook Yoon; Sung-Ho Shin; Jae-Cheol Ryou

It is important to manage large amount of information from various information providers for supporting recent information services such as providing global research trends, detecting emerging signal and listing leading researchers. For integrated management, definition of integrated metadata schema, data transformation and schema matching are needed. It is first necessary to analyze existing various metadata for defining integrated metadata schema. In this paper, we have analyzed several metadata of scientific journal papers by classifying semantics, content rules and syntax, and looked around considerations to make integrated schema or transform metadata. We have known that XML is used as a syntax for supporting convenience and various usage condition, and hierarchy element names and common elements in semantics are needed. We also have looked at elements having various content rules and related standards. We hope that this study will be used as basic research material of metadata integrated management, data transform and schema matching for interoperability.


text speech and dialogue | 2007

On the evaluation of Korean wordnet

Altangerel Chagnaa; Ho-Seop Choe; Cheol-Young Ock; Hwa-Mook Yoon

WordNet has become an important and useful resource for the natural language processing field. Recently, many countries have been developing their own WordNet. In this paper we show an evaluation of the Korean WordNet (U-WIN). The purpose of the work is to study how well the manually created lexical taxonomy U-WIN is built. Evaluation is done level by level, and the reason for selecting words for each level is that we want to compare each level and to find relations between them. As a result the words at a certain level (level 6) give the best score, for which we can make a conclusion that the words at this level are better organized than those at other levels. The score decreases as the level goes up or down from this particular level.


international conference on advanced language processing and web information technology | 2007

Toward DB-IR Integration: Per-Document Basis Transactional Index Maintenance

Jinsuk Kim; Du-Seok Jin; Yun-Soo Choi; Chang-Hoo Jeong; Kwang-Young Kim; Sung-Pil Choi; Min-Ho Lee; Min-Hee Cho; Ho-Seop Choe; Hwa-Mook Yoon; Jeong-Hyun Seo

While information retrieval(IR) and databases(DB) have been developed independently, there have been emerging requirements that both data management and efficient text retrieval should be supported simultaneously in an information system such as health care systems, bulletin boards, XML data management, and digital libraries. Recently DB-IR integration issue has been budded in the research field. The great divide between DB and IR has caused different manners in index maintenance for newly arriving documents. While DB has extended its SQL layer to cope with text fields due to lack of intact mechanism to build IR-like index, IR usually treats a block of new documents as a logical unit of index maintenance since it has no concept of integrity constraint. However, towards DB-IR integration, a transaction on adding or updating a document should include maintenance of the postings lists accompanied by the document - hence per-document basis transactional index maintenance. In this paper, performance of a few strategies for per-document basis transaction for inserting documents -- direct index update, stand-alone auxiliary index and pulsing auxiliary index - will be evaluated. The result tested on the KRISTAL-IRMS shows that the pulsing auxiliary strategy, where long postings lists in the auxiliary index are in-place updated to the main index whereas short lists are directly updated in the auxiliary index, can be a challenging candidate for text field indexing in DB-IR integration.


Archive | 2011

Automatic Metadata Conversion System with Ontology

Min-Ho Lee; Hwa-Mook Yoon; Won-Kyung Sung; Jae-Cheol Ryou

Some information providers or information circulation centers are collecting various metadata to serve users. The collected information has various structures and expression types of contents as well as file types. For information service, these various types of metadata should be processed and converted into one unified form. Converting metadata has lots of difficulties and mapping between elements of metadata schema is most important and difficult. We are making an automatic metadata conversion system with ontology. This paper describes the conversion method with ontology and schema mapping tool in the system.


parallel and distributed computing: applications and technologies | 2007

Service-Centric Object Fragmentation for Efficient Retrieval and Management of Huge XML Documents

Chang-Hoo Jeong; Yun-Soo Choi; Du-Seok Jin; Min-Ho Lee; Sung-Pil Choi; Kwang-Young Kim; Min-Hee Cho; Won-Kyun Joo; Hwa-Mook Yoon; Jeong-Hyun Seo; Jinsuk Kim

Vast amount of XML documents raise interests in how they will be used and how far their usage can be expanded, This paper has two central goals: 1) easy and fast retrieval of XML documents or relevant elements; and 2) efficient and stable management of large-size XML documents, The keys to develop such a practical system are how to segment a large XML document to smaller fragments and how to store them. In order to achieve these goals, we designed SCOF(Service-centric Object Fragmentation) model, which is a semi-decomposition method based on conversion rules provided by XML database managers. Keyword-based search using SCOF model then retrieves the specific elements or attributes of XML documents, just as typical XML query language does. Even though this approach needs the wisdom of managers in XML document collection, SCOF model makes it efficient both retrieval and management of massive XML documents.


Multimedia Tools and Applications | 2014

ETL-based interoperable data management system

Wongoo Lee; Min-Ho Lee; Yun-Soo Choi; Donghoon Choi; Min-Hee Cho; Sa-Kwang Song; Hanmin Jung; DongHwi Lee; Hwa-Mook Yoon

An explosive increase of information and demand for inter-contents association knowledge has made clear the need for integrated contents management. With contents being managed in various ways, the issue of maintaining interoperability among data also has been a consistent concern. This paper proposes possible solutions for the integration of contents management and the unification of the management processes. It has been proven that the e-Tracer system is easy to use and appropriate to the integration of metadata through user’s usage and it is able to maintain interoperability for different types of contents.


The Journal of the Korea Contents Association | 2012

Optimization and Performance Analysis of Distributed Parallel Processing Platform for Terminology Recognition System

Yun-Soo Choi; Wongoo Lee; Min-Ho Lee; Donghoon Choi; Hwa-Mook Yoon; Sa-Kwang Song; Hanmin Jung

Many statistical methods have been adapted for terminology recognition to improve its accuracy. However, since previous studies have been carried out in a single core or a single machine, they have difficulties in real-time analysing explosively increasing documents. In this study, the task where bottlenecks occur in the process of terminology recognition is classified into linguistic processing in the process of `candidate terminology extraction` and collection of statistical information in the process of `terminology weight assignment`. A terminology recognition system is implemented and experimented to address each task by means of the distributed parallel processing-based MapReduce. The experiments were performed in two ways; the first experiment result revealed that distributed parallel processing by means of 12 nodes improves processing speed by 11.27 times as compared to the case of using a single machine and the second experiment was carried out on 1) default environment, 2) multiple reducers, 3) combiner, and 4) the combination of 2)and 3), and the use of 3) showed the best performance. Our terminology recognition system contributes to speed up knowledge extraction of large scale science and technology documents.


The Journal of the Korea Contents Association | 2011

Semi-Automatic Management of Classification Scheme with Interoperability

Wongoo Lee; Sung-Ho Shin; Kwang-Young Kim; Do-Heon Jeon; Hwa-Mook Yoon; Won-Kyung Sung; Min-Ho Lee

Under the knowledge-based economy in 21C, the convergence and complexity in science and technology are being more active. Therefore, we have science and technology are classified properly, make not easy to construct the system to new next generation area. Thus we suggest the systematic solution method to flexibly extend classification scheme in order for content management and service organizations. In this way, we expect that the difficult of classification scheme management is minimized and the expense of it is spared.


International Conference on U- and E-Service, Science and Technology | 2011

An Management of Digital Contents on Science and Technology

Wongoo Lee; Min-Ho Lee; Kwang-Young Kim; Sung-Ho Shin; Hwa-Mook Yoon; Won-Kyung Sung

It needs a management system that managers manage more conveniently, faster, more systematically and Users use faster, more various, more reliable, more discriminatory, higher value-added services on science & technology informations. Thereupon, We design and implement DiCMS(Digital Contents Management System) standardized and unified in a variety of literature data.

Collaboration


Dive into the Hwa-Mook Yoon's collaboration.

Top Co-Authors

Avatar

Min-Ho Lee

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Wongoo Lee

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Kwang-Young Kim

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Won-Kyung Sung

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Yun-Soo Choi

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Chang-Hoo Jeong

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Jinsuk Kim

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Du-Seok Jin

Korea Institute of Science and Technology Information

View shared research outputs
Top Co-Authors

Avatar

Jae-Cheol Ryou

Chungnam National University

View shared research outputs
Top Co-Authors

Avatar

Min-Hee Cho

Korea Institute of Science and Technology Information

View shared research outputs
Researchain Logo
Decentralizing Knowledge