Joan Mas
Autonomous University of Barcelona
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Joan Mas.
international conference on document analysis and recognition | 2013
Dimosthenis Karatzas; Faisal Shafait; Seiichi Uchida; Masakazu Iwamura; Lluís Gómez i Bigorda; Sergi Robles Mestre; Joan Mas; David Fernandez Mota; Jon Almazán; Lluís Pere de las Heras
This report presents the final results of the ICDAR 2013 Robust Reading Competition. The competition is structured in three Challenges addressing text extraction in different application domains, namely born-digital images, real scene images and real-scene videos. The Challenges are organised around specific tasks covering text localisation, text segmentation and word recognition. The competition took place in the first quarter of 2013, and received a total of 42 submissions over the different tasks offered. This report describes the datasets and ground truth specification, details the performance evaluation protocols used and presents the final results along with a brief summary of the participating methods.
international conference on document analysis and recognition | 2011
Dimosthenis Karatzas; S. Robles Mestre; Joan Mas; F. Nourbakhsh; Partha Pratim Roy
This paper presents the results of the first Challenge of ICDAR 2011 Robust Reading Competition. Challenge 1 is focused on the extraction of text from born-digital images, specifically from images found in Web pages and emails. The challenge was organized in terms of three tasks that look at different stages of the process: text localization, text segmentation and word recognition. In this paper we present the results of the challenge for all three tasks, and make an open call for continuous participation outside the context of ICDAR 2011.
international conference on document analysis and recognition | 2007
Joan Mas; Gemma Sánchez; Josep Lladós; Bart Lamiroy
This paper presents a syntactic recognition approach for on-line drawn graphical symbols. The proposed method consists in an incremental on-line predictive parser based on symbol descriptions by an adjacency grammar. The parser analyzes input strokes as they are drawn by the user and is able to get ahead which symbols are likely to be recognized when a partial subshape is drawn in an intermediate state. In addition, the parser takes into account two issues. First, symbol strokes are drawn in any order by the user and second, since it is an on-line framework, the system requires real-time response. The method has been applied to an on-line sketching interface for architectural symbols.
document analysis systems | 2004
Gemma Sánchez; Ernest Valveny; Josep Lladós; Joan Mas; Narcís Lozano
This paper proposes a general architecture to extract knowledge from graphic documents. The architecture consists of three major components. First, a set of modules able to extract descriptors that, combined with domain-dependent knowledge and recognition strategies, allow to interpret a given graphical document. Second, a representation model based on a graph structure that allows to hierarchically represent the information of the document at different abstraction levels. Finally, the third component implements a calligraphic interface that allows the feedback between the user and the system. The second part of the paper describes an application scenario of the above platform. The scenario is a system for the interpretation of sketches of architectural plans. This is a tool to convert sketches to a CAD representation or to edit a given plan by a sketchy interface. The application scenario combines different symbol recognition algorithms stated in terms of document descriptors to extract building elements such as doors, windows, walls and furniture.
Pattern Recognition | 2010
Joan Mas; Josep Lladós; Gemma Sánchez; Joaquim A. Jorge
This paper presents a syntactic approach based on Adjacency Grammars (AG) for sketch diagram modeling and understanding. Diagrams are a combination of graphical symbols arranged according to a set of spatial rules defined by a visual language. AG describe visual shapes by productions defined in terms of terminal and non-terminal symbols (graphical primitives and subshapes), and a set functions describing the spatial arrangements between symbols. Our approach to sketch diagram understanding provides three main contributions. First, since AG are linear grammars, there is a need to define shapes and relations inherently bidimensional using a sequential formalism. Second, our parsing approach uses an indexing structure based on a spatial tessellation. This serves to reduce the search space when finding candidates to produce a valid reduction. This allows order-free parsing of 2D visual sentences while keeping combinatorial explosion in check. Third, working with sketches requires a distortion model to cope with the natural variations of hand drawn strokes. To this end we extended the basic grammar with a distortion measure modeled on the allowable variation on spatial constraints associated with grammar productions. Finally, the paper reports on an experimental framework an interactive system for sketch analysis. User tests performed on two real scenarios show that our approach is usable in interactive settings.
iberian conference on pattern recognition and image analysis | 2005
Joan Mas; Gemma Sánchez; Josep Lladós
The recent advances in sketch-based applications and digital-pen protocols make visual languages useful tools for Human Computer Interaction. Graphical symbols are the core elements of a sketch and, hence a visual language. Thus, symbol recognition approaches are the basis for visual language parsing. In this paper we propose an adjacency grammar to represent graphical symbols in a sketchy framework. Adjacency grammars represent the visual syntax in terms of adjacency relations between primitives. Graphical symbols may be either diagram components or gestures. An on-line parsing method is also proposed. The performance of the recognition is evaluated using a benchmarking database of 5000 on-line symbols. Finally, an application framework for sketching architectural floor plans is described.
international conference on document analysis and recognition | 2011
Lluís-Pere de las Heras; Joan Mas; Gemma S´nchez; Ernest Valveny
Segmentation of architectural floor plans is a challenging task, mainly because of the large variability in the notation between different plans. In general, traditional techniques, usually based on analyzing and grouping structural primitives obtained by vectorization, are only able to handle a reduced range of similar notations. In this paper we propose an alternative patch-based segmentation approach working at pixel level, without need of vectorization. The image is divided into a set of patches and a set of features is extracted for every patch. Then, each patch is assigned to a visual word of a previously learned vocabulary and given a probability of belonging to each class of objects. Finally, a post-process assigns the final label for every pixel. This approach has been applied to the detection of walls on two datasets of architectural floor plans with different notations, achieving high accuracy rates.
graphics recognition | 2008
Joan Mas; Joaquim A. Jorge; Gemma Sánchez; Josep Lladós
While much work has been done in Structural and Syntactical Pattern Recognition applied to drawings, most approaches are non-interactive. However, the recent emergence of viable pen-computers makes it desirable to handle pen-input such as sketches and drawings interactively. This paper presents a syntax-directed approach to parse sketches based on Relational Adjacency Grammars, which describe spatial and topological relations among parts of a sketch. Our approach uses a 2D grid to avoid re-scanning all the previous input whenever new strokes entered into the system, thus speeding up parsing considerably. To evaluate the performance of our approach we have tested the system using non-trivial inputs analyzed with two different grammars, one to design user interfaces and the other to describe floor-plans. The results clearly show the effectiveness of our approach and demonstrate good scalability to larger drawings.
Proceedings of the First International Conference on Digital Access to Textual Cultural Heritage | 2014
Alicia Fornés; Josep Lladós; Joan Mas; Joana Maria Pujades; Anna Cabré
In this paper we present a crowdsourcing web-based application for extracting information from demographic handwritten document images. The proposed application integrates two points of view: the semantic information for demographic research, and the ground-truthing for document analysis research. Concretely, the application has the contents view, where the information is recorded into forms, and the labeling view, with the word labels for evaluating document analysis techniques. The crowdsourcing architecture allows to accelerate the information extraction (many users can work simultaneously), validate the information, and easily provide feedback to the users. We finally show how the proposed application can be extended to other kind of demographic historical manuscripts.
graphics recognition | 2011
Lluís-Pere de las Heras; Joan Mas; Gemma Sánchez; Ernest Valveny
Architectural floor plans exhibit a large variability in notation. Therefore, segmenting and identifying the elements of any kind of plan becomes a challenging task for approaches based on grouping structural primitives obtained by vectorization. Recently, a patch-based segmentation method working at pixel level and relying on the construction of a visual vocabulary has been proposed in [1], showing its adaptability to different notations by automatically learning the visual appearance of the elements in each different notation. This paper presents an evolution of that previous work, after analyzing and testing several alternatives for each of the different steps of the method: Firstly, an automatic plan-size normalization process is done. Secondly we evaluate different features to obtain the description of every patch. Thirdly, we train an SVM classifier to obtain the category of every patch instead of constructing a visual vocabulary. These variations of the method have been tested for wall detection on two datasets of architectural floor plans with different notations. After studying in deep each of the steps in the process pipeline, we are able to find the best system configuration, which highly outperforms the results on wall segmentation obtained by the original paper.