Tohgoroh Matsui
Nagoya Institute of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Tohgoroh Matsui.
knowledge discovery and data mining | 1999
Daisuke Shibata; Nobuhiro Inuzuka; Shohei Kato; Tohgoroh Matsui; Hidenori Itoh
This paper gives a formulation of inductive learning based on fuzzy logic programming (FLP) and a top-down algorithm for it by extending an inductive logic programming (ILP) algorithm FOIL. The algorithm was implemented and evaluated by experiments. Linguistic hedges, which modifies truth, are shown to have effect to adjust classification properties. The algorithm deals with structural domain as other ILP algorithms do and also works well with numeric attributes.
international conference on knowledge-based and intelligent information and engineering systems | 2003
Tohgoroh Matsui; Nobuhiro Inuzuka; Hirohisa Seki
Reinforcement learning constructs knowledge containing state-to-action decision rules from agent’s experiences. Most of reinforcement learning methods are action-value estimation methods which estimate the true values of state-action pairs and derive the optimal policy from the value estimates. However, these methods have a serious drawback that they stray when the values for the “opposite” actions, such as moving left and moving right, are equal. This paper describes the basic mechanism of on-line profit-sharing (OnPS) which is an action-preference learning method. The main contribution of this paper is to show the equivalence of off-line and on-line in profit sharing. We also show a typical benchmark example for comparison between OnPS and Q-learning.
pacific rim international conference on artificial intelligence | 2000
Tohgoroh Matsui; Nobuhiro Inuzuka; Hirohisa Seki
This paper proposes new architecture of agent which adapts its behavior by predicting the results of actions and avoiding taking ones predicted to be failures shown in Fig. 1. Our agent consists of five parts: Observer, Planner, Actor which are same as those of basic agent and write perceiving information in the first-order formalism, Learner which acquires prediction rules using inductive logic programming (ILP), and Checker which predicts the result of an action selected by Planner and changes action if it seems to be failure. The agent classifies past actions into successes and failures, then it learns rules from them to assort the current action without information after taking it.
discovery science | 2005
Masaki Yokoyama; Tohgoroh Matsui; Hayato Ohwada
This paper proposes a method for detecting misclassifications of a classification rule and then revising them. Given a rule and a set of examples, the method divides misclassifications by the rule into miscovered examples and uncovered examples, and then, separately, learns to detect them using Inductive Logic Programming (ILP). The method then combines the acquired rules with the initial rule and revises the labels of misclassified examples. The paper shows the effectiveness of the proposed method by theoretical analysis. In addition, it presents experimental results, using the Brill tagger for Part-Of-Speech (POS) tagging.
international conference on knowledge-based and intelligent information and engineering systems | 2003
Takahisa Ishiguro; Tohgoroh Matsui; Nobuhiro Inuzuka; Koichi Wada
Reinforcement learning methods for environment including actions with differing costs are investigated. Through experiments we examined treatment of this problem with Q-learning, R-learning and Profit sharing. Profit sharing with a credit assignment functions considering costs is shown to have good performance in a practical sense.
Archive | 1998
Tohgoroh Matsui; Nobuhiro Inuzuka; Hirohisa Seki; Hidenori Itoh
inductive logic programming | 2000
Tohgoroh Matsui; Nobuhiro Inuzuka; Hirohisa Seki
Artificial Intelligence and Applications | 2005
Yoko Ino; Tohgoroh Matsui; Hayato Ohwada
Archive | 1999
Tohgoroh Matsui; Kazuo Kashiwabara; Nobuhiro Inuzuka; Hirohisa Seki; Hidenori Itoh
全国大会講演論文集 | 2008
Shimpei Kawamura; Tohgoroh Matsui; Hidetaka Kaya; Hayato Ohwada; Kazuyuki Kuchitsu