Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Keishi Tajima is active.

Publication


Featured researches published by Keishi Tajima.


ACM Transactions on Database Systems | 2004

Archiving scientific data

Peter Buneman; Sanjeev Khanna; Keishi Tajima; Wang Chiew Tan

Archiving is important for scientific data, where it is necessary to record all past versions of a database in order to verify findings based upon a specific version. Much scientific data is held in a hierachical format and has a key structure that provides a canonical identification for each element of the hierarchy. In this article, we exploit these properties to develop an archiving technique that is both efficient in its use of space and preserves the continuity of elements through versions of the database, something that is not provided by traditional minimum-edit-distance diff approaches. The approach also uses timestamps. All versions of the data are merged into one hierarchy where an element appearing in multiple versions is stored only once along with a timestamp. By identifying the semantic continuity of elements and merging them into one data structure, our technique is capable of providing meaningful change descriptions, the archive allows us to easily answer certain temporal queries such as retrieval of any specific version from the archive and finding the history of an element. This is in contrast with approaches that store a sequence of deltas where such operations may require undoing a large number of changes or significant reasoning with the deltas. A suite of experiments also demonstrates that our archive does not incur any significant space overhead when contrasted with diff approaches. Another useful property of our approach is that we use XML format to represent hierarchical data and the resulting archive is also in XML. Hence, XML tools can be directly applied on our archive. In particular, we apply an XML compressor on our archive, and our experiments show that our compressed archive outperforms compressed diff-based repositories in space efficiency. We also show how we can extend our archiving tool to an external memory archiver for higher scalability and describe various index structures that can further improve the efficiency of some temporal queries on our archive.


acm conference on hypertext | 1998

Cut as a querying unit for WWW, Netnews, and E-mail

Keishi Tajima; Yoshiaki Mizuuchi; Masatsugu Kitagawa; Katsumi Tanaka

In this paper, we propose a query framework for hypertext data in general, and for WWW pages, Netnews articles, and e-mails in particular. In existing query tools for hypertext data, such as search engines for WWW or intelligent news/mail readers, data units in query are typically individual nodes. In actual hypertext data, however, one topic is often described over a series of connected nodes, and therefore, the logical data unit should be such a series of nodes corresponding to one topic. This discrepancy between the data unit in query and the logical data unit hinders the e cient information discovery from hypertext data. To solve this problem, in our framework, we divide hypertexts into connected subgraphs corresponding to individual topics, and we use those subgraphs as the data units in queries.


acm conference on hypertext | 1999

Finding context paths for Web pages

Yoshiaki Mizuuchi; Keishi Tajima

The contents of Web pages are often not self-contained. A page author often assumes all the readers of the page come through the same path, and he sometimes omit the information described in the pages on that path because the readers must already know it. Therefore, indexes used by search engines based on the contents of each page are also incomplete. In this paper, we propose a method of discovering those paths assumed by page authors, and of complementing the incomplete indexes with keywords extracted from the pages on those paths.


very large data bases | 2004

Answering xpath queries over networks by sending minimal views

Keishi Tajima; Yoshiki Fukui

When a client submits a set of XPath queries to a XML database on a network, the set of answer sets sent back by the database may include redundancy in two ways: some elements may appear in more than one answer set, and some elements in some answer sets may be subelements of other elements in other (or the same) answer sets. Even when a client submits a single query, the answer can be self-redundant because some elements may be subelements of other elements in that answer. Therefore, sending those answers as they are is not optimal with respect to communication costs. In this paper, we propose a method of minimizing communication costs in XPath processing over networks. Given a single or a set of queries, we compute a minimal-size view set that can answer all the original queries. The database sends this view set to the client, and the client produces answers from it. We show algorithms for computing such a minimal view set for given queries. This view set is optimal; it only includes elements that appear in some of the final answers, and each element appears only once.


IEEE Transactions on Knowledge and Data Engineering | 2001

A query model to synthesize answer intervals from indexed video units

Sujeet Pradhan; Keishi Tajima; Katsumi Tanaka

While a query result in a traditional database is a subset of the database, in a video database, it is a set of subintervals extracted from the raw video sequence. It is very hard, if not impossible, to predetermine all the queries that will be issued in the future, and all the subintervals that will become necessary to answer them. As a result, conventional query frameworks are not applicable to video databases. We propose a new video query model that computes query results by dynamically synthesizing needed subintervals from fragmentary indexed intervals in the database. We introduce new interval operations required for that computation. We also propose methods to compute relative relevance of synthesized intervals to a given query. A query result is a list of synthesized intervals sorted in the order of their degree of relevance.


user interface software and technology | 2008

Browsing large HTML tables on small screens

Keishi Tajima; Kaori Ohnishi

We propose new interaction techniques that support better browsing of large HTML tables on small screen devices, such as mobile phones. We propose three modes for browsing tables: normal mode, record mode, and cell mode. Normal mode renders tables in the ordinary way, but provides various useful functions for browsing large tables, such as hiding unnecessary rows and columns. Record mode regards each row (or column) as the basic information unit and displays it in a record-like format with column (or row) headers, while cell mode regards each cell as the basic unit and displays each cell together with its corresponding row and column headers. For these table presentations, we need to identify row and column headers that explain the meaning of rows and columns. To provide users with both row and column headers even when the tables have attributes for only one of them, we introduce the concept of keys and develop a method of automatically discovering attributes and keys in tables. Another issue in these presentations is how to handle composite cells spanning multiple rows or columns. We determine the semantics of such composite cells and render them in appropriate ways in accordance with their semantics.


symposium on principles of database systems | 1994

A polymorphic calculus for views and object sharing (extended abstract)

Atsushi Ohori; Keishi Tajima

We present a typed polymorphic calculus that supports a general mechanism for view definition and object sharing among classes. In this calculus, a class can contain inclusion specifications of objects from other classes. Each such specification consists of a predicate determining the subset of objects to be included and a viewing function under which those included objects are manipulated. Both predicates and viewing functions can be any type consistent programs definable in the polymorphic calculus. Inclusion specifications among classes can be cyclic, allowing mutually recursive class definitions. These features achieve flexible view definitions and wide range of class organizations in a compact and elegant way. Moreover, the calculus provides a suitable set of operations for views and classes so that the programmer can manipulate views and classes just the same way as one deals with ordinary records and sets. The proposed calculus uniformly integrates views and classes in a polymorphic type system of a database programming language similar to Machiavelli. The calculus has a type inference algorithm that relieves the programmer from complicated type declarations of views and classes. The polymorphic type system of the calculus is also shown to be sound, which guarantees complete static check of type consistency of programs involving classes and views. Through these properties, the programmer can enjoy full advantages of polymorphism and type inference when writing object-oriented database programs.


web intelligence | 2005

Cache Replacement for Transcoding Proxy Caching

Keqiu Li; Keishi Tajima; Hong Shen

In this paper, we address the problem of cache replacement for transcoding proxy caching. First, an efficient cache replacement algorithm is proposed. Our algorithm considers both the aggregate effect of caching multiple versions of the same multimedia object and cache consistency. Second, a complexity analysis is presented to show the efficiency of our algorithm. Finally, some preliminary simulation experiments are conducted to compare the performance of our algorithm with some existing algorithms. The results show that our algorithm outperforms others in terms of the various performance metrics.


conference on information and knowledge management | 2012

Tweet classification based on their lifetime duration

Hikaru Takemura; Keishi Tajima

Many microblog messages remain useful only within a short time, and users often find such a message after its informational value has vanished. Users also sometimes miss old but still useful messages buried among outdated ones. To solve these problems, we develop a method of classifying messages into the following three categories: (1) messages that users should read now because their value will diminish soon, (2) messages that users may read later because their value will not largely change soon, and (3) messages that are not useful anymore because their value has vanished. Our method uses an error correcting output code consisting of binary classifiers each of which determines whether a given message has value at specific time point. Our experiments on Twitter data confirmed that it outperforms naive methods.


The Journal of Supercomputing | 2006

An effective cache replacement algorithm in transcoding-enabled proxies

Keqiu Li; Hong Shen; Keishi Tajima; Liusheng Huang

In this paper, we address the problem of cache replacement for transcoding proxy caching. Transcoding proxy is a proxy that has the functionality of transcoding a multimedia object into an appropriate format or resolution for each client. We first propose an effective cache replacement algorithm for transcoding proxy. In general, when a new object is to be cached, cache replacement algorithms evict some of the cached objects with the least profit to accommodate the new object. Our algorithm takes into account of the inter-relationships among different versions of the same multimedia object, and selects the versions to replace according to their aggregate profit which usually differs from simple summation of their individual profits as assumed in the existing algorithms. It also considers cache consistency, which is not considered in the existing algorithms. We then present a complexity analysis to show the efficiency of our algorithm. Finally, we give extensive simulation results to compare the performance of our algorithm with some existing algorithms. The results show that our algorithm outperforms others in terms of various performance metrics.

Collaboration


Dive into the Keishi Tajima's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Keqiu Li

Japan Advanced Institute of Science and Technology

View shared research outputs
Top Co-Authors

Avatar

Sujeet Pradhan

Kurashiki University of Science and the Arts

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Hong Shen

University of Adelaide

View shared research outputs
Researchain Logo
Decentralizing Knowledge