Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Mikel L. Forcada is active.

Publication


Featured researches published by Mikel L. Forcada.


Machine Translation | 2011

Apertium: a free/open-source platform for rule-based machine translation

Mikel L. Forcada; Mireia Ginestí-Rosell; Jacob Nordfalk; Jim O'Regan; Sergio Ortiz-Rojas; Juan Antonio Pérez-Ortiz; Felipe Sánchez-Martínez; Gema Ramírez-Sánchez; Francis M. Tyers

Apertium is a free/open-source platform for rule-based machine translation. It is being widely used to build machine translation systems for a variety of language pairs, especially in those cases (mainly with related-language pairs) where shallow transfer suffices to produce good quality translations, although it has also proven useful in assimilation scenarios with more distant pairs involved. This article summarises the Apertium platform: the translation engine, the encoding of linguistic data, and the tools developed around the platform. The present limitations of the platform and the challenges posed for the coming years are also discussed. Finally, evaluation results for some of the most active language pairs are presented. An appendix describes Apertium as a free/open-source project.


processing of the portuguese language | 2006

Open-Source portuguese–spanish machine translation

Carme Armentano-Oller; Rafael C. Carrasco; Antonio M. Corbí-Bellot; Mikel L. Forcada; Mireia Ginestí-Rosell; Sergio Ortiz-Rojas; Juan Antonio Pérez-Ortiz; Gema Ramírez-Sánchez; Felipe Sánchez-Martínez; Miriam A. Scalco

This paper describes the current status of development of an open-source shallow-transfer machine translation (MT) system for the [European] Portuguese


workshop on statistical machine translation | 2008

MaTrEx: The DCU MT System for WMT 2008

Sergio Penkale; Rejwanul Haque; Sandipan Dandapat; Pratyush Banerjee; Ankit Kumar Srivastava; Jinhua Du; Pavel Pecina; Sudip Kumar Naskar; Mikel L. Forcada; Andy Way

\leftrightarrow


Computational Linguistics | 2002

Incremental construction and maintenance of minimal finite-state automata

Rafael C. Carrasco; Mikel L. Forcada

Spanish language pair, developed using the OpenTrad Apertium MT toolbox (www.apertium.org). Apertium uses finite-state transducers for lexical processing, hidden Markov models for part-of-speech tagging, and finite-state-based chunking for structural transfer, and is based on a simple rationale: to produce fast, reasonably intelligible and easily correctable translations between related languages, it suffices to use a MT strategy which uses shallow parsing techniques to refine word-for-word MT. This paper briefly describes the MT engine, the formats it uses for linguistic data, and the compilers that convert these data into an efficient format used by the engine, and then goes on to describe in more detail the pilot Portuguese


Neural Computation | 1995

Learning the initial state of a second-order recurrent neural network during regular-language inference

Mikel L. Forcada; Rafael C. Carrasco

\leftrightarrow


Neural Computation | 2000

Stable Encoding of Finite-State Machines in Discrete-Time Recurrent Neural Nets with Sigmoid Units

Rafael C. Carrasco; Mikel L. Forcada; M. Ángeles Valdés-Muòoz; Ramón P. Ñeco

Spanish linguistic data.


Machine Translation | 2006

Automatic induction of bilingual resources from aligned parallel corpora: application to shallow-transfer machine translation

Helena de Medeiros Caseli; Maria das Graças Volpe Nunes; Mikel L. Forcada

In this paper, we give a description of the machine translation system developed at DCU that was used for our participation in the evaluation campaign of the Third Workshop on Statistical Machine Translation at ACL 2008. We describe the modular design of our data-driven MT system with particular focus on the components used in this participation. We also describe some of the significant modules which were unused in this task. We participated in the EuroParl task for the following translation directions: Spanish-English and French-English, in which we employed our hybrid EBMT-SMT architecture to translate. We also participated in the Czech-English News and News Commentary tasks which represented a previously untested language pair for our system. We report results on the provided development and test sets.


Journal of Artificial Intelligence Research | 2009

Inferring shallow-transfer machine translation rules from small parallel corpora

Felipe Sánchez-Martínez; Mikel L. Forcada

Daciuk et al. [Computational Linguistics 26(1):316 (2000)] describe a method for constructing incrementally minimal, deterministic, acyclic finite-state automata (dictionaries) from sets of strings. But acyclic finite-state automata have limitations: For instance, if one wants a linguistic application to accept all possible integer numbers or Internet addresses, the corresponding finite-state automaton has to be cyclic. In this article, we describe a simple and equally efficient method for modifying any minimal finite-state automaton (be it acyclic or not) so that a string is added to or removed from the language it accepts; both operations are very important when dictionary maintenance is performed and solve the dictionary construction problem addressed by Daciuk et al. as a special case. The algorithms proposed here may be straightforwardly derived from the customary textbook constructions for the intersection and the complementation of finite-state automata; the algorithms exploit the special properties of the automata resulting from the intersection operation when one of the finite-state automata accepts a single string.


The Prague Bulletin of Mathematical Linguistics | 2010

Combining Content-Based and URL-Based Heuristics to Harvest Aligned Bitexts from Multilingual Sites with Bitextor

Miquel Esplà-Gomis; Mikel L. Forcada

Recent work has shown that second-order recurrent neural networks (2ORNNs) may be used to infer regular languages. This paper presents a modified version of the real-time recurrent learning (RTRL) algorithm used to train 2ORNNs, that learns the initial state in addition to the weights. The results of this modification, which adds extra flexibility at a negligible cost in time complexity, suggest that it may be used to improve the learning of regular languages when the size of the network is small.


Journal of Chemical Physics | 1991

On liquid‐film thickness measurements with the atomic‐force microscope

Mikel L. Forcada; Mario M. Jakas; Alberto Gras‐Martí

There has been a lot of interest in the use of discrete-time recurrent neural nets (DTRNN) to learn finite-state tasks, with interesting results regarding the induction of simple finite-state machines from inputoutput strings. Parallel work has studied the computational power of DTRNN in connection with finite-state computation. This article describes a simple strategy to devise stable encodings of finite-state machines in computationally capable discrete-time recurrent neural architectures with sigmoid units and gives a detailed presentation on how this strategy may be applied to encode a general class of finite-state machines in a variety of commonly used first- and second-order recurrent neural networks. Unlike previous work that either imposed some restrictions to state values or used a detailed analysis based on fixed-point attractors, our approach applies to any positive, bounded, strictly growing, continuous activation function and uses simple bounding criteria based on a study of the conditions under which a proposed encoding scheme guarantees that the DTRNN is actually behaving as a finite-state machine.

Collaboration


Dive into the Mikel L. Forcada's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Andy Way

Dublin City University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge