Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Juhani Luotolahti is active.

Publication


Featured researches published by Juhani Luotolahti.


Proceedings of the CoNLL 2017 Shared Task: Multilingual Parsing from Raw#N# Text to Universal Dependencies | 2017

CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies

Daniel Zeman; Martin Popel; Milan Straka; Jan Hajic; Joakim Nivre; Filip Ginter; Juhani Luotolahti; Sampo Pyysalo; Slav Petrov; Martin Potthast; Francis M. Tyers; Elena Badmaeva; Memduh Gokirmak; Anna Nedoluzhko; Silvie Cinková; Jaroslava Hlaváčová; Václava Kettnerová; Zdenka Uresová; Jenna Kanerva; Stina Ojala; Anna Missilä; Christopher D. Manning; Sebastian Schuster; Siva Reddy; Dima Taji; Nizar Habash; Herman Leung; Marie-Catherine de Marneffe; Manuela Sanguinetti; Maria Simi

The Conference on Computational Natural Language Learning (CoNLL) features a shared task, in which participants train and test their learning systems on the same data sets. In 2017, the task was devoted to learning dependency parsers for a large number of languages, in a real-world setting without any gold-standard annotation on input. All test sets followed a unified annotation scheme, namely that of Universal Dependencies. In this paper, we define the task and evaluation methodology, describe how the data sets were prepared, report and analyze the main results, and provide a brief categorization of the different approaches of the participating systems.


north american chapter of the association for computational linguistics | 2015

SETS: Scalable and Efficient Tree Search in Dependency Graphs

Juhani Luotolahti; Jenna Kanerva; Sampo Pyysalo; Filip Ginter

We present a syntactic analysis query toolkit geared specifically towards massive dependency parsebanks and morphologically rich languages. The query language allows arbitrary tree queries, including negated branches, and is suitable for querying analyses with rich morphological annotation. Treebanks of over a million words can be comfortably queried on a low-end netbook, and a parsebank with over 100M words on a single consumer-grade server. We also introduce a web-based interface for interactive querying. All contributions are available under open licenses.


Proceedings of the CoNLL 2017 Shared Task: Multilingual Parsing from Raw Text to Universal Dependencies : August 3-4, 2017 Vancouver, Canada, 2017, ISBN 978-1-945626-70-8, págs. 119-125 | 2017

TurkuNLP: Delexicalized Pre-training of Word Embeddings for Dependency Parsing

Jenna Kanerva; Juhani Luotolahti; Filip Ginter

We present the TurkuNLP entry in the CoNLL 2017 Shared Task on Multilingual Parsing from Raw Text to Universal Dependencies. The system is based on the UDPipe parser with our focus being in ex- ploring various techniques to pre-train the word embeddings used by the parser in order to improve its performance especially on languages with small training sets. The system ranked 11th among the 33 participants overall, being 8th on the small treebanks, 10th on the large treebanks, 12th on the parallel test sets, and 26th on the sur- prise languages.


International Conference on Internet Science | 2018

Neural Network Hate Deletion: Developing a Machine Learning Model to Eliminate Hate from Online Comments

Joni Salminen; Juhani Luotolahti; Hind Almerekhi; Bernard J. Jansen; Soon-Gyo Jung

We propose a method for modifying hateful online comments to non-hateful comments without losing the understandability and original meaning of the comments. To accomplish this, we retrieve and classify 301,153 hateful and 1,041,490 non-hateful comments from Facebook and YouTube channels of a large international media organization that is a target of considerable online hate. We supplement this dataset by 10,000 Reddit comments manually labeled for hatefulness. Using these two datasets, we train a neural network to distinguish linguistic patterns. The model we develop, Neural Network Hate Deletion (NNHD), computes how hateful the sentences of a social media comment are and if they are above a given threshold, it deletes them using a language dependency tree. We evaluate the results by comparing crowd workers’ perceptions of hatefulness and understandability before and after transformation and find that our method reduces hatefulness without resulting in a significant loss of understandability. In some cases, removing hateful elements improves understandability by reducing the linguistic complexity of the comment. In addition, we find that NNHD can satisfactorily retain the original meaning on average but is not perfect in this regard. In terms of practical implications, NNHD could be used in social media platforms to suggest more neutral use of language to agitated online users.


Proceedings of the First Conference on Machine Translation: Volume 2, Shared Task Papers | 2016

Cross-Lingual Pronoun Prediction with Deep Recurrent Neural Networks.

Juhani Luotolahti; Jenna Kanerva; Filip Ginter

In this paper we present our winning system in the WMT16 Shared Task on CrossLingual Pronoun Prediction, where the objective is to predict a missing target language pronoun based on the target and source sentences. Our system is a deep recurrent neural network, which reads both the source language and target language context with a softmax layer making the final prediction. Our system achieves the best macro recall on all four language pairs. The margin to the next best system ranges between less than 1pp and almost 12pp depending on the language pair.


north american chapter of the association for computational linguistics | 2015

Turku: Semantic Dependency Parsing as a Sequence Classification

Jenna Kanerva; Juhani Luotolahti; Filip Ginter

This paper presents the University of Turku entry to the SemEval-2015 task on BroadCoverage Semantic Dependency Parsing. The system uses an existing transition-based parser as a sequence classifier to jointly predict all arguments of one candidate predicate at a time. Compared to our 2014 entry, the 2015 system gains about 3pp in terms of F-score for a fraction of the development time. Depending on the subtask, the difference between our entry and the winning system ranges between 1 and 5pp.


Baltic HLT | 2014

Syntactic N-gram Collection from a Large-Scale Corpus of Internet Finnish.

Jenna Kanerva; Juhani Luotolahti; Veronika Laippala; Filip Ginter


empirical methods in natural language processing | 2017

Cross-Lingual Pronoun Prediction with Deep Recurrent Neural Networks v2.0.

Juhani Luotolahti; Jenna Kanerva; Filip Ginter


Proceedings of the Third International Conference on Dependency Linguistics (Depling 2015) | 2015

Towards Universal Web Parsebanks

Juhani Luotolahti; Jenna Kanerva; Veronika Laippala; Sampo Pyysalo; Filip Ginter


Proceedings of the 21st Nordic Conference on Computational Linguistics, NoDaLiDa, 22-24 May 2017, Gothenburg, Sweden | 2017

Creating register sub-corpora for the Finnish Internet Parsebank.

Veronika Laippala; Juhani Luotolahti; Aki-Juhani Kyröläinen; Tapio Salakoski; Filip Ginter

Collaboration


Dive into the Juhani Luotolahti's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Sampo Pyysalo

University of Manchester

View shared research outputs
Top Co-Authors

Avatar

Daniel Zeman

Charles University in Prague

View shared research outputs
Top Co-Authors

Avatar

Jan Hajic

Charles University in Prague

View shared research outputs
Top Co-Authors

Avatar

Milan Straka

Charles University in Prague

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge