Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where David M. J. Tax is active.

Publication


Featured researches published by David M. J. Tax.


Machine Learning | 2004

Support Vector Data Description

David M. J. Tax; Robert P. W. Duin

Data domain description concerns the characterization of a data set. A good description covers all target data but includes no superfluous space. The boundary of a dataset can be used to detect novel data or outliers. We will present the Support Vector Data Description (SVDD) which is inspired by the Support Vector Classifier. It obtains a spherically shaped boundary around a dataset and analogous to the Support Vector Classifier it can be made flexible by using other kernel functions. The method is made robust against outliers in the training set and is capable of tightening the description by using negative examples. We show characteristics of the Support Vector Data Descriptions using artificial and real data.


Pattern Recognition Letters | 1999

Support vector domain description

David M. J. Tax; Robert P. W. Duin

This paper shows the use of a data domain description method, inspired by the support vector machine by Vapnik, called the support vector domain description (SVDD). This data description can be used for novelty or outlier detection. A spherically shaped decision boundary around a set of objects is constructed by a set of support vectors describing the sphere boundary. It has the possibility of transforming the data to new feature spaces without much extra computational cost. By using the transformed data, this SVDD can obtain more flexible and more accurate data descriptions. The error of the first kind, the fraction of the training objects which will be rejected, can be estimated immediately from the description without the use of an independent test set, which makes this method data eAcient. The support vector domain description is compared with other outlier detection methods on real data. ” 1999 Elsevier Science B.V. All rights reserved.


Pattern Recognition | 2000

Combining multiple classifiers by averaging or by multiplying

David M. J. Tax; Martijn van Breukelen; Robert P. W. Duin; Josef Kittler

Abstract In classification tasks it may be wise to combine observations from different sources. Not only it decreases the training time but it can also increase the robustness and the performance of the classification. Combining is often done by just (weighted) averaging of the outputs of the different classifiers. Using equal weights for all classifiers then results in the mean combination rule. This works very well in practice, but the combination strategy lacks a fundamental basis as it cannot readily be derived from the joint probabilities. This contrasts with the product combination rule which can be obtained from the joint probability under the assumption of independency. In this paper we will show differences and similarities between this mean combination rule and the product combination rule in theory and in practice.


multiple classifier systems | 2000

Experiments with Classifier Combining Rules

Robert P. W. Duin; David M. J. Tax

A large experiment on combining classifiers is reported and discussed. It includes, both, the combination of different classifiers on the same feature set and the combination of classifiers on different feature sets. Various fixed and trained combining rules are used. It is shown that there is no overall winning combining rule and that bad classifiers as well as bad feature sets may contain valuable information for performance improvement by combining rules. Best performance is achieved by combining both, different feature sets and different classifiers.


international conference on pattern recognition | 2002

Using two-class classifiers for multiclass classification

David M. J. Tax; Robert P. W. Duin

The generalization from two-class classification to multiclass classification is not straightforward for discriminants which are not based on density estimation. Simple combining methods use voting, but this has the drawback of inconsequent labelings and ties. More advanced methods map the discriminant outputs to approximate posterior probability estimates and combine these, while other methods use error-correcting output codes. In this paper we want to show the possibilities of simple generalizations of the two-class classification, using voting and combinations of approximate posterior probabilities.


multiple classifier systems | 2001

Combining One-Class Classifiers

David M. J. Tax; Robert P. W. Duin

In the problem of one-class classification target objects should be distinguished from outlier objects. In this problem it is assumed that only information of the target class is available while nothing is known about the outlier class. Like standard two-class classifiers, one-class classifiers hardly ever fit the data distribution perfectly. Using only the best classifier and discarding the classifiers with poorer performance might waste valuable information. To improve performance the results of different classifiers (which may differ in complexity or training algorithm) can be combined. This can not only increase the performance but it can also increase the robustness of the classification. Because for one-class classifiers only information of one of the classes is present, combining one-class classifiers is more difficult. In this paper we investigate if and how one-class classifiers can be combined best in a handwritten digit recognition problem.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2008

Subclass Problem-Dependent Design for Error-Correcting Output Codes

Sergio Escalera; David M. J. Tax; Oriol Pujol; Petia Radeva; Robert P. W. Duin

A common way to model multiclass classification problems is by means of Error-Correcting Output Codes (ECOCs). Given a multiclass problem, the ECOC technique designs a code word for each class, where each position of the code identifies the membership of the class for a given binary problem. A classification decision is obtained by assigning the label of the class with the closest code. One of the main requirements of the ECOC design is that the base classifier is capable of splitting each subgroup of classes from each binary problem. However, we cannot guarantee that a linear classifier model convex regions. Furthermore, nonlinear classifiers also fail to manage some type of surfaces. In this paper, we present a novel strategy to model multiclass classification problems using subclass information in the ECOC framework. Complex problems are solved by splitting the original set of classes into subclasses and embedding the binary problems in a problem-dependent ECOC design. Experimental results show that the proposed splitting procedure yields a better performance when the class overlap or the distribution of the training objects conceal the decision boundaries for the base classifier. The results are even more significant when one has a sufficiently large training size.


Neurocomputing | 2009

Minimum spanning tree based one-class classifier

Piotr Juszczak; David M. J. Tax; Elzbieta Pekalska; Robert P. W. Duin

In the problem of one-class classification one of the classes, called the target class, has to be distinguished from all other possible objects. These are considered as non-targets. The need for solving such a task arises in many practical applications, e.g. in machine fault detection, face recognition, authorship verification, fraud recognition or person identification based on biometric data. This paper proposes a new one-class classifier, the minimum spanning tree class descriptor (MST_CD). This classifier builds on the structure of the minimum spanning tree constructed on the target training set only. The classification of test objects relies on their distances to the closest edge of that tree, hence the proposed method is an example of a distance-based one-class classifier. Our experiments show that the MST_CD performs especially well in case of small sample size problems and in high-dimensional spaces.


Lecture Notes in Computer Science | 1998

Classifier Conditional Posterior Probabilities

Robert P. W. Duin; David M. J. Tax

Classifiers based on probability density estimates can be used to find posterior probabilities for the objects to be classified. These probabilities can be used for rejection or for combining classifiers. Posterior probabilities for other classifiers, however, have to be conditional for the classifier., i.e. they yield class probabilities for a given value of the classifier outcome instead for a given input feature vector. In this paper they are studied for a set of individual classifiers as well as for combination rules.


Lecture Notes in Computer Science | 1998

Outlier Detection Using Classifier Instability

David M. J. Tax; Robert P. W. Duin

When a classifier is used to classify objects, it is important to know if these objects resemble the training objects the classifier is trained with. Several methods to detect novel objects exist. In this paper a new method is presented which is based on the instability of the output of simple classifiers on new objects. The performances of the outlier detection methods is shown in a handwritten digit recognition problem.

Collaboration


Dive into the David M. J. Tax's collaboration.

Top Co-Authors

Avatar

Robert P. W. Duin

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Marco Loog

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Veronika Cheplygina

Erasmus University Rotterdam

View shared research outputs
Top Co-Authors

Avatar

Dick de Ridder

Wageningen University and Research Centre

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Pavel Paclík

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Wenjie Pei

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Piotr Juszczak

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Yan Li

Delft University of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge