Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Batselem Jagvaral is active.

Publication


Featured researches published by Batselem Jagvaral.


international conference on big data and smart computing | 2017

Large-scale incremental OWL/RDFS reasoning over fuzzy RDF data

Batselem Jagvaral; Lee Wangon; Hyun-Kyu Park; Myung-Joong Jeon; Nam-Gee Lee; Young-Tack Park

Ontological RDF data are extracted from multiple sources on the web through mapping and alignment for various purposes, but extracting and reasoning about ontologies from different sources causes information ambiguity and uncertainty. A reasonable solution to this problem is to annotate extracted ontology data with truth values to determine the reliability of information. However, the recent growth in data has brought forth difficulties in ascertaining the credibility of numerous ontologies during OWL/RDFS reasoning. In this paper, we present a distributed and incremental reasoning approach for RDF data with uncertainty. We focused on RDFS and OWL pD* semantics and developed methods for incremental OWL reasoning with uncertainty. We also introduced parallel algorithms that resolve the scalable reasoning problem. To evaluate the efficiency of the proposed system, we conducted OWL/RDFS reasoning over fuzzy LUBM3000 and achieved a performance three times higher than that achieved with the fastest reasoning system.


Journal of KIISE | 2015

SPARQL Query Processing in Distributed In-Memory System

Batselem Jagvaral; Wan-Gon Lee; KangPil Kim; Young-Tack Park

In this paper, we propose a query processing approach that uses the Spark functional programming and distributed memory system to solve the computational overhead of SPARQL. In the semantic web, RDF ontology data is produced at large scale, and the main challenge for the semantic web is to query and manipulate such a large ontology with a high throughput. The most existing studies on SPARQL have focused on deploying the Hadoop MapReduce framework, and although approaches based on Hadoop MapReduce have shown promising results, they achieve a low level of throughput due to the underlying distributed file processes. Therefore, in order to speed up the query processes, we suggest query- processing methods that are based on memory caching in distributed memory system. Our approach is also integrated with a clause unification method for propagating between the clauses that exploits Spark join, map and filter methods along with caching. In our experiments, we have achieved a high level of performance relative to other approaches. In particular, our performance was nearly similar to that of Sempala, which has been considered to be the fastest query processing system.


Journal of KIISE | 2014

Scalable RDFS Reasoning using Logic Programming Approach in a Single Machine

Batselem Jagvaral; Je-Min Kim; Wan Gon Lee; Young Tack Park

시맨틱 웹상에서 RDFS로 표현된 데이터의 사용 증가로 인하여, 대용량 데이터의 추론에 대한 많은 요구가 생겨나고 있다. 많은 연구자들은 대용량 온톨로지 추론을 수행하기 위해서 하둡과 같은 고가의 분산 프레임워크를 활용한다. 그러나, 적절한 사이즈의 RDFS 트리플 추론을 위해서는 굳이 고가의 분산 환경 시스템을 사용하지 않고 단일 머신에서도 논리적 프로그래밍을 이용하면 분산 환경과 유사한 추론 성능을 얻을 수 있다. 본 논문에서는 단일 머신에 논리적 프로그래밍 방식을 적용한 대용량 RDFS 추론 기법을 제안하였고 다중 머신을 기반으로 한 분산 환경 시스템과 비교하여 2억개 정도의 트리플에 대한 RDFS 추론 시스템을 적용한 경우 분산환경과 비슷한 성능을 보이는 것을 실험적으로 증명하였다. 효율적인 추론을 위해 온톨로지 모델을 세부적으로 분리한 메타데이터 구조와 대용량 트리플의 색인 방안을 제안하고 이를 위해서 전체 트리플을 하나의 모델로 로딩하는 것이 아니라 각각 온톨로지 추론 규칙에 따라 적절한 트리플 집합을 선택하였다. 또한 논리 프로그래밍이 제공하는 Unification 알고리즘 기반의 트리플 매칭, 검색, Conjunctive 질의어 처리 기반을 활용하는 온톨로지 추론 방식을 제안한다. 제안된 기법이 적용된 추론 엔진을 LUBM1500(트리플 수 2억개) 에 대해서 실험한 결과 166K/sec의 추론 성능을 얻었는데 이는 8개의 노드(8 코아/노드)환경에서 맵 리듀스로 수행한 WebPIE의 185K/sec의 추론 속도와 유사함을 실험적으로 증명하였다. 따라서 단일 머신에서 수행되는 본 연구 결과는 트리플의 수가 2억개 정도까지는 분산환경시스템을 활용하지 않고도 분산환경 시스템과 비교해서 비슷한 성능을 보이는 것을 확인할 수 있었다.


Journal of KIISE | 2016

Distributed Assumption-Based Truth Maintenance System for Scalable Reasoning

Batselem Jagvaral; Young-Tack Park

가정기반진리관리 시스템(ATMS)은 추론 시스템의 추론 과정을 저장하고 비단조추론을 지원할 수 있는 도구이다 또한 의존기반 backtracking을 지원하므로 매우 넓은 공간 탐색 문제를 해결 할 수 있는 강력한 도구이다. 모든 추론 과정을 기록하고, 특정한 컨텍스트에서 지능형시스템의 Belief를 매우 빠르게 확인하고 비단조 추론 문제에 대한 해결책을 효율적으로 제공할 수 있게 한다. 그러나 최근 데이터의 양이 방대해지면서 기존의 단일 머신을 사용하는 경우 문제 해결 프로그램의 대용량의 추론과정을 저장하는 것이 불가능하게 되었다. 대용량 데이터에 대한 문제 해결 과정을 기록하는 것은 많은 연산과 메모리 오버헤드를 야기한다. 이러한 단점을 극복하기 위해 본 논문에서는 Apache Spark 환경에서 functional 및 객체지향 방식 기반의 점진적 컨텍스트 추론을 유지할 수 있는 방법을 제안한다.. 이는 가정(Assumption)과 유도과정을 분산 환경에 저장하며, 실체화된 대용량 데이터셋의 변화를 효율적으로 수정가능하게 한다. 또한 ATMS의 Label, Environment를 분산 처리하여 대규모의 추론 과정을 효과적으로 관리할 수 있는 방안을 제시하고 있다. 제안하는 시스템의 성능을 측정하기 위해 5개의 노드로 구성된 클러스터에서 LUBM 데이터셋에 대한 OWL/RDFS 추론을 수행하고, 데이터의 추가, 설명, 제거에 대한 실험을 수행하였다. LUBM2000에 대하여 추론을 수행한 결과 80GB데이터가 추론되었고, ATMS에 적용하여 추가, 설명, 제거에 대하여 수초 내에 처리하는 성능을 보였다.


Journal of KIISE | 2016

Spark based Scalable RDFS Ontology Reasoning over Big Triples with Confidence Values

Hyun-Kyu Park; Wan-Gon Lee; Batselem Jagvaral; Young-Tack Park

Recently, due to the development of the Internet and electronic devices, there has been an enormous increase in the amount of available knowledge and information. As this growth has proceeded, studies on large-scale ontological reasoning have been actively carried out. In general, a machine learning program or knowledge engineer measures and provides a degree of confidence for each triple in a large ontology. Yet, the collected ontology data contains specific uncertainty and reasoning such data can cause vagueness in reasoning results. In order to solve the uncertainty issue, we propose an RDFS reasoning approach that utilizes confidence values indicating degrees of uncertainty in the collected data. Unlike conventional reasoning approaches that have not taken into account data uncertainty, by using the in-memory based cluster computing framework Spark, our approach computes confidence values in the data inferred through RDFS-based reasoning by applying methods for uncertainty estimating. As a result, the computed confidence values represent the uncertainty in the inferred data. To evaluate our approach, ontology reasoning was carried out over the LUBM standard benchmark data set with addition arbitrary confidence values to ontology triples. Experimental results indicated that the proposed system is capable of running over the largest data set LUBM3000 in 1179 seconds inferring 350K triples.


international conference on big data and smart computing | 2015

Distributed scalable RDFS reasoning

Batselem Jagvaral; Young-Tack Park

A number of reasoning studies on big ontology have been carried out in the recent years. However, most of the existing studies have focused heavily on Hadoop MapReduce. In this paper, we propose a reasoning approach for Resource Description Framework Schema (RDFS) that employs optimized methods based on Spark. Spark is a general distributed inmemory framework for large-scale data processing that is not tied to the two-stage MapReduce paradigm. In our work, we devised an extensive optimization method to cope with the communication bottleneck of data shuffling between machine nodes in a distributed system. From empirical evaluations, the proposed reasoning system produces at most the throughput of 4166KT/sec which is almost 80% faster than the MapReduce based reasoner WebPIE.


Journal of KIISE | 2015

Scalable RDFS Reasoning Using the Graph Structure of In-Memory based Parallel Computing

Myung-Joong Jeon; ChiSeoung So; Batselem Jagvaral; KangPil Kim; Jin Kim; JinYoung Hong; Young-Tack Park

In recent years, there has been a growing interest in RDFS Inference to build a rich knowledge base. However, it is difficult to improve the inference performance with large data by using a single machine. Therefore, researchers are investigating the development of a RDFS inference engine for a distributed computing environment. However, the existing inference engines cannot process data in real-time, are difficult to implement, and are vulnerable to repetitive tasks. In order to overcome these problems, we propose a method to construct an in-memory distributed inference engine that uses a parallel graph structure. In general, the ontology based on a triple structure possesses a graph structure. Thus, it is intuitive to design a graph structure-based inference engine. Moreover, the RDFS inference rule can be implemented by utilizing the operator of the graph structure, and we can thus design the inference engine according to the graph structure, and not the structure of the data table. In this study, we evaluate the proposed inference engine by using the LUBM1000 and LUBM3000 data to test the speed of the inference. The results of our experiment indicate that the proposed in-memory distributed inference engine achieved a performance of about 10 times faster than an in-storage inference engine.


Journal of KIISE | 2018

Knowledge Completion Modeling using Knowledge Base Embedding

Hyun-Young Choi; Ji-Hun Hong; Wan-Gon Lee; Batselem Jagvaral; Myung-Joong Jeon; Hyun-Kyu Park; Young-Tack Park


Journal of KIISE | 2018

SWAT: A Study on the Efficient Integration of SWRL and ATMS based on a Distributed In-Memory System

Myung-Joong Jeon; Wan-Gon Lee; Batselem Jagvaral; Hyun-Kyu Park; Young-Tack Park


Journal of KIISE | 2018

Extracting Rules from Neural Networks with Continuous Attributes

Batselem Jagvaral; Wan-Gon Lee; Myung-Joong Jeon; Hyun-Kyu Park; Young-Tack Park

Collaboration


Dive into the Batselem Jagvaral's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge