Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Bart Thomee is active.

Publication


Featured researches published by Bart Thomee.


Communications of The ACM | 2016

YFCC100M: the new data in multimedia research

Bart Thomee; David A. Shamma; Gerald Friedland; Benjamin Elizalde; Karl Ni; Douglas N. Poland; Damian Borth; Li-Jia Li

This publicly available curated dataset of almost 100 million photos and videos is free and legal for all.This publicly available curated dataset of almost 100 million photos and videos is free and legal for all.


multimedia information retrieval | 2010

New trends and ideas in visual concept detection: the MIR flickr retrieval evaluation initiative

Mark J. Huiskes; Bart Thomee; Michael S. Lew

The MIR Flickr collection consists of 25000 high-quality photographic images of thousands of Flickr users, made available under the Creative Commons license. The database includes all the original user tags and EXIF metadata. Additionally, detailed and accurate annotations are provided for topics corresponding to the most prominent visual concepts in the user tag data. The rich metadata allow for a wide variety of image retrieval benchmarking scenarios. In this paper, we provide an overview of the various strategies that were devised for automatic visual concept detection using the MIR Flickr collection. In particular we discuss results from various experiments in combining social data and low-level content-based descriptors to improve the accuracy of visual concept classifiers. Additionally, we present retrieval results obtained by relevance feedback methods, demonstrating (i) how their performance can be enhanced using features based on visual concept classifiers, and (ii) how their performance, based on small samples, can be measured relative to their large sample classifier counterparts. Additionally, we identify a number of promising trends and ideas in visual concept detection. To keep the MIR Flickr collection up-to-date on these developments, we have formulated two new initiatives to extend the original image collection. First, the collection will be extended to one million Creative Commons Flickr images. Second, a number of state-of-the-art content-based descriptors will be made available for the entire collection.


International Journal of Multimedia Information Retrieval | 2012

Interactive search in image retrieval: a survey

Bart Thomee; Michael S. Lew

We are living in an Age of Information where the amount of accessible data from science and culture is almost limitless. However, this also means that finding an item of interest is increasingly difficult, a digital needle in the proverbial haystack. In this article, we focus on the topic of content-based image retrieval using interactive search techniques, i.e., how does one interactively find any kind of imagery from any source, regardless of whether it is photographic, MRI or X-ray? We highlight trends and ideas from over 170 recent research papers aiming to capture the wide spectrum of paradigms and methods in interactive search, including its subarea relevance feedback. Furthermore, we identify promising research directions and several grand challenges for the future.


acm multimedia | 2010

TOP-SURF: a visual words toolkit

Bart Thomee; E. Bakker; Michael S. Lew

TOP-SURF is an image descriptor that combines interest points with visual words, resulting in a high performance yet compact descriptor that is designed with a wide range of content-based image retrieval applications in mind. TOP-SURF offers the flexibility to vary descriptor size and supports very fast image matching. In addition to the source code for the visual word extraction and comparisons, we also provide a high level API and very large pre-computed codebooks targeting web image content for both research and teaching purposes.


cross language evaluation forum | 2013

ImageCLEF 2013: The Vision, the Data and the Open Challenges

Barbara Caputo; Henning Müller; Bart Thomee; Mauricio Villegas; Roberto Paredes; David Zellhöfer; Hervé Goëau; Alexis Joly; Pierre Bonnet; Jesús Martínez Gómez; Ismael García Varea; Miguel Cazorla

This paper presents an overview of the ImageCLEF 2013 lab. Since its first edition in 2003, ImageCLEF has become one of the key initiatives promoting the benchmark evaluation of algorithms for the cross-language annotation and retrieval of images in various domains, such as public and personal images, to data acquired by mobile robot platforms and botanic collections. Over the years, by providing new data collections and challenging tasks to the community of interest, the ImageCLEF lab has achieved an unique position in the multi lingual image annotation and retrieval research landscape. The 2013 edition consisted of three tasks: the photo annotation and retrieval task, the plant identification task and the robot vision task. Furthermore, the medical annotation task, that traditionally has been under the ImageCLEF umbrella and that this year celebrates its tenth anniversary, has been organized in conjunction with AMIA for the first time. The paper describes the tasks and the 2013 competition, giving an unifying perspective of the present activities of the lab while discussion the future challenges and opportunities.


acm multimedia | 2014

The Placing Task: A Large-Scale Geo-Estimation Challenge for Social-Media Videos and Images

Jaeyoung Choi; Bart Thomee; Gerald Friedland; Liangliang Cao; Karl Ni; Damian Borth; Benjamin Elizalde; Luke R. Gottlieb; Carmen J. Carrano; Roger A. Pearce; Douglas N. Poland

The Placing Task is a yearly challenge offered by the MediaEval Multimedia Benchmarking Initiative that requires participants to develop algorithms that automatically predict the geo-location of social media videos and images. We introduce a recent development of a new standardized web-scale geo-tagged dataset for Placing Task 2014, which contains 5.5 million photos and 35,000 videos. This standardized benchmark with a large persistent dataset allows research community to easily evaluate new algorithms and to analyze their performance with respect to the state-of-the-art approaches. We discuss the characteristics of this years Placing Task along with the description of the new dataset components and how they were collected.


multimedia information retrieval | 2008

Large scale image copy detection evaluation

Bart Thomee; Mark J. Huiskes; E. Bakker; Michael S. Lew

In this paper we provide a comparative study of content-based copy detection methods, which include research literature methods based on salient point matching (SURF), discrete cosine and wavelet transforms, color histograms, biologically motivated visual matching and other methods. In our evaluation we focus on large-scale applications, especially on performance in the context of search engines for web images. We assess the scalability of the tested methods by investigating the detection accuracy relative to descriptor size, description time per image and matching time per image. For testing, original images altered by a diverse set of realistic transformations are embedded in a collection of one million web images.


international world wide web conferences | 2013

Uncovering locally characterizing regions within geotagged data

Bart Thomee; Adam Rae

We propose a novel algorithm for uncovering the colloquial boundaries of locally characterizing regions present in collections of labeled geospatial data. We address the problem by first modeling the data using scale-space theory, allowing us to represent it simultaneously across different scales as a family of increasingly smoothed density distributions. We then derive region boundaries by applying localized label weighting and image processing techniques to the scale-space representation of each label. Important insights into the data can be acquired by visualizing the shape and size of the resulting boundaries for each label at multiple scales. We demonstrate our technique operating at scale by discovering the boundaries of the most geospatially salient tags associated with a large collection of georeferenced photos from Flickr and compare our characterizing regions that emerge from the data with those produced by a recent technique from the research literature.


multimedia information retrieval | 2004

Relevance feedback: perceptual learning and retrieval in bio-computing, photos, and video

Micha Haas; Joachim T. Rijsdam; Bart Thomee; Michael S. Lew

One of the most important characteristics about relevance feedback is that it ideally finds a set of human perceptually correlated results because the user is directly involved in the search process. In principle, relevance feedback is an iterative learning process where positive and negative examples accumulate as the user gives feedback on each new iteration of results. If we view relevance feedback as a learning problem then we can immediately grasp that there will be the associated problem of learning from a small training set. Towards a solution, we present MediaNet, which is an approach toward integrating additional knowledge sources into the relevance feedback process. The additional knowledge sources are used to shape the learning space when insufficient training samples are available. We also integrate genetic or evolutionary algorithms directly into the search process. Experiments are given on test collections in bio-computing, general photos and video


international world wide web conferences | 2015

Describing and Understanding Neighborhood Characteristics through Online Social Media

Mohamed Kafsi; Henriette Cramer; Bart Thomee; David A. Shamma

Geotagged data can be used to describe regions in the world and discover local themes. However, not all data produced within a region is necessarily specifically descriptive of that area. To surface the content that is characteristic for a region, we present the geographical hierarchy model (GHM), a probabilistic model based on the assumption that data observed in a region is a random mixture of content that pertains to different levels of a hierarchy. We apply the GHM to a dataset of 8 million Flickr photos in order to discriminate between content (i.e. tags) that specifically characterizes a region (e.g. neighborhood) and content that characterizes surrounding areas or more general themes. Knowledge of the discriminative and non-discriminative terms used throughout the hierarchy enables us to quantify the uniqueness of a given region and to compare similar but distant regions. Our evaluation demonstrates that our model improves upon traditional Naive Bayes classification by 47% and hierarchical TF-IDF by 27%. We further highlight the differences and commonalities with human reasoning about what is locally characteristic for a neighborhood, distilled from ten interviews and a survey that covered themes such as time, events, and prior regional knowledge.

Collaboration


Dive into the Bart Thomee's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Benjamin Elizalde

International Computer Science Institute

View shared research outputs
Top Co-Authors

Avatar

Douglas N. Poland

Lawrence Livermore National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Gerald Friedland

International Computer Science Institute

View shared research outputs
Top Co-Authors

Avatar

Karl Ni

Lawrence Livermore National Laboratory

View shared research outputs
Top Co-Authors

Avatar

Claudia Hauff

Delft University of Technology

View shared research outputs
Researchain Logo
Decentralizing Knowledge