Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yannis Kalantidis is active.

Publication


Featured researches published by Yannis Kalantidis.


computer vision and pattern recognition | 2014

Locally Optimized Product Quantization for Approximate Nearest Neighbor Search

Yannis Kalantidis; Yannis S. Avrithis

We present a simple vector quantizer that combines low distortion with fast search and apply it to approximate nearest neighbor (ANN) search in high dimensional spaces. Leveraging the very same data structure that is used to provide non-exhaustive search, i.e., inverted lists or a multi-index, the idea is to locally optimize an individual product quantizer (PQ) per cell and use it to encode residuals. Local optimization is over rotation and space decomposition, interestingly, we apply a parametric solution that assumes a normal distribution and is extremely fast to train. With a reasonable space and time overhead that is constant in the data size, we set a new state-of-the-art on several public datasets, including a billion-scale one.


international conference on multimedia retrieval | 2013

Getting the look: clothing recognition and segmentation for automatic product suggestions in everyday photos

Yannis Kalantidis; Lyndon Kennedy; Li-Jia Li

We present a scalable approach to automatically suggest relevant clothing products, given a single image without metadata. We formulate the problem as cross-scenario retrieval: the query is a real-world image, while the products from online shopping catalogs are usually presented in a clean environment. We divide our approach into two main stages: a) Starting from articulated pose estimation, we segment the person area and cluster promising image regions in order to detect the clothing classes present in the query image. b) We use image retrieval techniques to retrieve visually similar products from each of the detected classes. We achieve clothing detection performance comparable to the state-of-the-art on a very recent annotated dataset, while being more than 50 times faster. Finally, we present a large scale clothing suggestion scenario, where the product database contains over one million products.


acm multimedia | 2010

Retrieving landmark and non-landmark images from community photo collections

Yannis S. Avrithis; Yannis Kalantidis; Giorgos Tolias; Evaggelos Spyrou

State of the art data mining and image retrieval in community photo collections typically focus on popular subsets, e.g. images containing landmarks or associated to Wikipedia articles. We propose an image clustering scheme that, seen as vector quantization compresses a large corpus of images by grouping visually consistent ones while providing a guaranteed distortion bound. This allows us, for instance, to represent the visual content of all thousands of images depicting the Parthenon in just a few dozens of scene maps and still be able to retrieve any single, isolated, non-landmark image like a house or graffiti on a wall. Starting from a geo-tagged dataset, we first group images geographically and then visually, where each visual cluster is assumed to depict different views of the the same scene. We align all views to one reference image and construct a 2D scene map by preserving details from all images while discarding repeating visual features. Our indexing, retrieval and spatial matching scheme then operates directly on scene maps. We evaluate the precision of the proposed method on a challenging one-million urban image dataset.


international conference on multimedia retrieval | 2011

Scalable triangulation-based logo recognition

Yannis Kalantidis; Lluis Garcia Pueyo; Michele Trevisiol; Roelof van Zwol; Yannis S. Avrithis

We propose a scalable logo recognition approach that extends the common bag-of-words model and incorporates local geometry in the indexing process. Given a query image and a large logo database, the goal is to recognize the logo contained in the query, if any. We locally group features in triples using multi-scale Delaunay triangulation and represent triangles by signatures capturing both visual appearance and local geometry. Each class is represented by the union of such signatures over all instances in the class. We see large scale recognition as a sub-linear search problem where signatures of the query image are looked up in an inverted index structure of the class models. We evaluate our approach on a large-scale logo recognition dataset with more than four thousand classes.


Multimedia Tools and Applications | 2011

VIRaL: Visual Image Retrieval and Localization

Yannis Kalantidis; Giorgos Tolias; Yannis S. Avrithis; Marios Phinikettos; Evaggelos Spyrou; Phivos Mylonas; Stefanos D. Kollias

New applications are emerging every day exploiting the huge data volume in community photo collections. Most focus on popular subsets, e.g., images containing landmarks or associated to Wikipedia articles. In this work we are concerned with the problem of accurately finding the location where a photo is taken without needing any metadata, that is, solely by its visual content. We also recognize landmarks where applicable, automatically linking them to Wikipedia. We show that the time is right for automating the geo-tagging process, and we show how this can work at large scale. In doing so, we do exploit redundancy of content in popular locations—but unlike most existing solutions, we do not restrict to landmarks. In other words, we can compactly represent the visual content of all thousands of images depicting e.g., the Parthenon and still retrieve any single, isolated, non-landmark image like a house or a graffiti on a wall. Starting from an existing, geo-tagged dataset, we cluster images into sets of different views of the same scene. This is a very efficient, scalable, and fully automated mining process. We then align all views in a set to one reference image and construct a 2D scene map. Our indexing scheme operates directly on scene maps. We evaluate our solution on a challenging one million urban image dataset and provide public access to our service through our online application, VIRaL.


european conference on computer vision | 2012

Approximate gaussian mixtures for large scale vocabularies

Yannis S. Avrithis; Yannis Kalantidis

We introduce a clustering method that combines the flexibility of Gaussian mixtures with the scaling properties needed to construct visual vocabularies for image retrieval. It is a variant of expectation-maximization that can converge rapidly while dynamically estimating the number of components. We employ approximate nearest neighbor search to speed-up the E-step and exploit its iterative nature to make search incremental, boosting both speed and precision. We achieve superior performance in large scale retrieval, being as fast as the best known approximate k-means.


international conference on image processing | 2010

Image clustering through community detection on hybrid image similarity graphs

Symeon Papadopoulos; Christos Zigkolis; Giorgos Tolias; Yannis Kalantidis; Phivos Mylonas; Yiannis Kompatsiaris; Athena Vakali

The wide adoption of photo sharing applications such as Flickr


acm multimedia | 2010

Feature map hashing: sub-linear indexing of appearance and global geometry

Yannis S. Avrithis; Giorgos Tolias; Yannis Kalantidis

We present a new approach to image indexing and retrieval, which integrates appearance with global image geometry in the indexing process, while enjoying robustness against viewpoint change, photometric variations, occlusion, and background clutter. We exploit shape parameters of local features to estimate image alignment via a single correspondence. Then, for each feature, we construct a sparse spatial map of all remaining features, encoding their normalized position and appearance, typically vector quantized to visual word. An image is represented by a collection of such feature maps and RANSAC-like matching is reduced to a number of set intersections. Because the induced dissimilarity is still not a metric, we extend min-wise independent permutations to collections of sets and derive a similarity measure for feature map collections. We then exploit sparseness to build an inverted file whereby the retrieval process is sub-linear in the total number of images, ideally linear in the number of relevant ones. We achieve excellent performance on 10^4 images, with a query time in the order of milliseconds.


Computer Vision and Image Understanding | 2014

Towards large-scale geometry indexing by feature selection

Giorgos Tolias; Yannis Kalantidis; Yannis S. Avrithis; Stefanos D. Kollias

We present a new approach to image indexing and retrieval, which integrates appearance with global image geometry in the indexing process, while enjoying robustness against viewpoint change, photometric variations, occlusion, and background clutter. We exploit shape parameters of local features to estimate image alignment via a single correspondence. Then, for each feature, we construct a sparse spatial map of all remaining features, encoding their normalized position and appearance, typically vector quantized to visual word. An image is represented by a collection of such feature maps and RANSAC-like matching is reduced to a number of set intersections. The required index space is still quadratic in the number of features. To make it linear, we propose a novel feature selection model tailored to our feature map representation, replacing our earlier hashing approach. The resulting index space is comparable to baseline bag-of-words, scaling up to one million images while outperforming the state of the art on three publicly available datasets. To our knowledge, this is the first geometry indexing method to dispense with spatial verification at this scale, bringing query times down to milliseconds.


international conference on computer vision | 2015

Web-Scale Image Clustering Revisited

Yannis S. Avrithis; Yannis Kalantidis; Evangelos Anagnostopoulos; Ioannis Z. Emiris

Large scale duplicate detection, clustering and mining of documents or images has been conventionally treated with seed detection via hashing, followed by seed growing heuristics using fast search. Principled clustering methods, especially kernelized and spectral ones, have higher complexity and are difficult to scale above millions. Under the assumption of documents or images embedded in Euclidean space, we revisit recent advances in approximate k-means variants, and borrow their best ingredients to introduce a new one, inverted-quantized k-means (IQ-means). Key underlying concepts are quantization of data points and multi-index based inverted search from centroids to cells. Its quantization is a form of hashing and analogous to seed detection, while its updates are analogous to seed growing, yet principled in the sense of distortion minimization. We further design a dynamic variant that is able to determine the number of clusters k in a single run at nearly zero additional cost. Combined with powerful deep learned representations, we achieve clustering of a 100 million image collection on a single machine in less than one hour.

Collaboration


Dive into the Yannis Kalantidis's collaboration.

Top Co-Authors

Avatar

Yannis S. Avrithis

National Technical University of Athens

View shared research outputs
Top Co-Authors

Avatar

Evaggelos Spyrou

National Technical University of Athens

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Stefanos D. Kollias

National Technical University of Athens

View shared research outputs
Top Co-Authors

Avatar

Ilianna Kollia

National Technical University of Athens

View shared research outputs
Top Co-Authors

Avatar

Andreas Stafylopatis

National Technical University of Athens

View shared research outputs
Top Co-Authors

Avatar

Athena Vakali

Aristotle University of Thessaloniki

View shared research outputs
Top Co-Authors

Avatar

Christos Zigkolis

Aristotle University of Thessaloniki

View shared research outputs
Top Co-Authors

Avatar

Evangelos Anagnostopoulos

National and Kapodistrian University of Athens

View shared research outputs
Top Co-Authors

Avatar

Ioannis Z. Emiris

National and Kapodistrian University of Athens

View shared research outputs
Researchain Logo
Decentralizing Knowledge