Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Elzbieta Pekalska is active.

Publication


Featured researches published by Elzbieta Pekalska.


Pattern Recognition Letters | 2002

Dissimilarity representations allow for building good classifiers

Elzbieta Pekalska; Robert P. W. Duin

In this paper, a classification task on dissimilarity representations is considered. A traditional way to discriminate between objects represented by dissimilarities is the nearest neighbor method. It suffers, however, from a number of limitations, i.e., high computational complexity, a potential loss of accuracy when a small set of prototypes is used and sensitivity to noise. To overcome these shortcomings, we propose to use a normal density-based classifier constructed on the same representation. We show that such a classifier, based on a weighted combination of dissimilarities, can significantly improve the nearest neighbor rule with respect to the recognition accuracy and computational effort.


IEEE Transactions on Pattern Analysis and Machine Intelligence | 2009

Kernel Discriminant Analysis for Positive Definite and Indefinite Kernels

Elzbieta Pekalska; Bernard Haasdonk

Kernel methods are a class of well established and successful algorithms for pattern analysis thanks to their mathematical elegance and good performance. Numerous nonlinear extensions of pattern recognition techniques have been proposed so far based on the so-called kernel trick. The objective of this paper is twofold. First, we derive an additional kernel tool that is still missing, namely kernel quadratic discriminant (KQD). We discuss different formulations of KQD based on the regularized kernel Mahalanobis distance in both complete and class-related subspaces. Secondly, we propose suitable extensions of kernel linear and quadratic discriminants to indefinite kernels. We provide classifiers that are applicable to kernels defined by any symmetric similarity measure. This is important in practice because problem-suited proximity measures often violate the requirement of positive definiteness. As in the traditional case, KQD can be advantageous for data with unequal class spreads in the kernel-induced spaces, which cannot be well separated by a linear discriminant. We illustrate this on artificial and real data for both positive definite and indefinite kernels.


Neurocomputing | 2009

Minimum spanning tree based one-class classifier

Piotr Juszczak; David M. J. Tax; Elzbieta Pekalska; Robert P. W. Duin

In the problem of one-class classification one of the classes, called the target class, has to be distinguished from all other possible objects. These are considered as non-targets. The need for solving such a task arises in many practical applications, e.g. in machine fault detection, face recognition, authorship verification, fraud recognition or person identification based on biometric data. This paper proposes a new one-class classifier, the minimum spanning tree class descriptor (MST_CD). This classifier builds on the structure of the minimum spanning tree constructed on the target training set only. The classification of test objects relies on their distances to the closest edge of that tree, hence the proposed method is an example of a distance-based one-class classifier. Our experiments show that the MST_CD performs especially well in case of small sample size problems and in high-dimensional spaces.


Lecture Notes in Computer Science | 2006

Non-Euclidean or non-metric measures can be informative

Elzbieta Pekalska; Artsiom Harol; Robert P. W. Duin; Barbara Spillmann; Horst Bunke

Statistical learning algorithms often rely on the Euclidean distance. In practice, non-Euclidean or non-metric dissimilarity measures may arise when contours, spectra or shapes are compared by edit distances or as a consequence of robust object matching [1,2]. It is an open issue whether such measures are advantageous for statistical learning or whether they should be constrained to obey the metric axioms. The k-nearest neighbor (NN) rule is widely applied to general dissimilarity data as the most natural approach. Alternative methods exist that embed such data into suitable representation spaces in which statistical classifiers are constructed [3]. In this paper, we investigate the relation between non-Euclidean aspects of dissimilarity data and the classification performance of the direct NN rule and some classifiers trained in representation spaces. This is evaluated on a parameterized family of edit distances, in which parameter values control the strength of non-Euclidean behavior. Our finding is that the discriminative power of this measure increases with increasing non-Euclidean and non-metric aspects until a certain optimum is reached. The conclusion is that statistical classifiers perform well and the optimal values of the parameters characterize a non-Euclidean and somewhat non-metric measure.


Lecture Notes in Computer Science | 2006

Transforming strings to vector spaces using prototype selection

Barbara Spillmann; Michel Neuhaus; Horst Bunke; Elzbieta Pekalska; Robert P. W. Duin

A common way of expressing string similarity in structural pattern recognition is the edit distance. It allows one to apply the kNN rule in order to classify a set of strings. However, compared to the wide range of elaborated classifiers known from statistical pattern recognition, this is only a very basic method. In the present paper we propose a method for transforming strings into n-dimensional real vector spaces based on prototype selection. This allows us to subsequently classify the transformed strings with more sophisticated classifiers, such as support vector machine and other kernel based methods. In a number of experiments, we show that the recognition rate can be significantly improved by means of this procedure.


international conference on pattern recognition | 2000

Classifiers for dissimilarity-based pattern recognition

Elzbieta Pekalska; Robert P. W. Duin

In the traditional way of learning from examples of objects the classifiers are built in a feature space. However, alternative ways can be found by constructing decision rules on dissimilarity (distance) representations, instead. In such a recognition process a new object is described by its distances to (a subset of) the training samples. In this paper, a number of methods to tackle this type of classification problem are investigated: the feature-based (interpreting the distance representation as a feature space) and rank-based (considering the given relations) decision rules. The experiments demonstrate that the feature-based (especially normal-based) classifiers often outperform the rank-based ones. This is to be expected, since summation-based distances are, under general conditions, approximately normally distributed. In addition, the support vector classifier also achieves a high accuracy.


Pattern Recognition Letters | 2012

The dissimilarity space

Robert P. W. Duin; Elzbieta Pekalska

Highlights? Consciousness divides human recognition in structural and statistical approaches. ? Dissimilarities, fundamental in human recognition are suited to integrate the two. ? The dissimilarity space is a good vector space for the dissimilarity representation. ? Classifiers in dissimilarity space are accurate and/or complexity. ? Combining dissimilarities by averaging may improve results further. Human experts constitute pattern classes of natural objects based on their observed appearance. Automatic systems for pattern recognition may be designed on a structural description derived from sensor observations. Alternatively, training sets of examples can be used in statistical learning procedures. They are most powerful for vectorial object representations. Unfortunately, structural descriptions do not match well with vectorial representations. Consequently it is difficult to combine the structural and statistical approaches to pattern recognition.Structural descriptions may be used to compare objects. This leads to a set of pairwise dissimilarities from which vectors can be derived for the purpose of statistical learning. The resulting dissimilarity representation bridges thereby the structural and statistical approaches.The dissimilarity space is one of the possible spaces resulting from this representation. It is very general and easy to implement. This paper gives a historical review and discusses the properties of the dissimilarity space approaches illustrated by a set of examples on real world datasets.


Lecture Notes in Computer Science | 2004

On Not Making Dissimilarities Euclidean

Elzbieta Pekalska; Robert P. W. Duin; Simon Günter; Horst Bunke

Non-metric dissimilarity measures may arise in practice e.g. when objects represented by sensory measurements or by structural de- scriptions are compared. It is an open issue whether such non-metric measures should be corrected in some way to be metric or even Eu- clidean. The reason for such corrections is the fact that pairwise metric distances are interpreted in metric spaces, while Euclidean distances can be embedded into Euclidean spaces. Hence, traditional learning methods can be used. The k-nearest neighbor rule is usually applied to dissimilarities. In our earlier study (12, 13), we proposed some alternative approaches to general dissimilarity representations (DRs). They rely either on an embedding to a pseudo-Euclidean space and building classifiers there or on construct- ing classifiers on the representation directly. In this paper, we investigate ways of correcting DRs to make them more Euclidean (metric) either by adding a proper constant or by some concave transformations. Classi- fication experiments conducted on five dissimilarity data sets indicate that non-metric dissimilarity measures can be more beneficial than their corrected Euclidean or metric counterparts. The discriminating power of the measure itself is more important than its Euclidean (or metric) properties.


international conference on pattern recognition | 2010

Feature-based dissimilarity space classification

Robert P. W. Duin; Marco Loog; Elzbieta Pekalska; David M. J. Tax

General dissimilarity-based learning approaches have been proposed for dissimilarity data sets [1,2]. They often arise in problems in which direct comparisons of objects are made by computing pairwise distances between images, spectra, graphs or strings. Dissimilarity-based classifiers can also be defined in vector spaces [3]. A large comparative study has not been undertaken so far. This paper compares dissimilarity-based classifiers with traditional feature-based classifiers, including linear and nonlinear SVMs, in the context of the ICPR 2010 Classifier Domains of Competence contest. It is concluded that the feature-based dissimilarity space classification performs similar or better than the linear and nonlinear SVMs, as averaged over all 301 datasets of the contest and in a large subset of its datasets. This indicates that these classifiers have their own domain of competence.


SSPR&SPR'10 Proceedings of the 2010 joint IAPR international conference on Structural, syntactic, and statistical pattern recognition | 2010

Non-Euclidean dissimilarities: causes and informativeness

Robert P. W. Duin; Elzbieta Pekalska

In the process of designing pattern recognition systems one may choose a representation based on pairwise dissimilarities between objects. This is especially appealing when a set of discriminative features is difficult to find. Various classification systems have been studied for such a dissimilarity representation: the direct use of the nearest neighbor rule, the postulation of a dissimilarity space and an embedding to a virtual, underlying feature vector space. It appears in several applications that the dissimilarity measures constructed by experts tend to have a non-Euclidean behavior. In this paper we first analyze the causes of such choices and then experimentally verify that the non-Euclidean property of the measure can be informative.

Collaboration


Dive into the Elzbieta Pekalska's collaboration.

Top Co-Authors

Avatar

Robert P. W. Duin

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

David M. J. Tax

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Artsiom Harol

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Carmen Lai

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Dick de Ridder

Wageningen University and Research Centre

View shared research outputs
Top Co-Authors

Avatar

Pavel Paclík

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Marco Loog

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Marina Skurichina

Delft University of Technology

View shared research outputs
Top Co-Authors

Avatar

Piotr Juszczak

Delft University of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge