Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Cícero Nogueira dos Santos is active.

Publication


Featured researches published by Cícero Nogueira dos Santos.


international joint conference on natural language processing | 2015

Classifying Relations by Ranking with Convolutional Neural Networks

Cícero Nogueira dos Santos; Bing Xiang; Bowen Zhou

Relation classification is an important semantic processing task for which state-ofthe-art systems still rely on costly handcrafted features. In this work we tackle the relation classification task using a convolutional neural network that performs classification by ranking (CR-CNN). We propose a new pairwise ranking loss function that makes it easy to reduce the impact of artificial classes. We perform experiments using the the SemEval-2010 Task 8 dataset, which is designed for the task of classifying the relationship between two nominals marked in a sentence. Using CRCNN, we outperform the state-of-the-art for this dataset and achieve a F1 of 84.1 without using any costly handcrafted features. Additionally, our experimental results show that: (1) our approach is more effective than CNN followed by a softmax classifier; (2) omitting the representation of the artificial class Other improves both precision and recall; and (3) using only word embeddings as input features is enough to achieve state-of-the-art results if we consider only the text between the two target nominals.


conference on computational natural language learning | 2016

Abstractive Text Summarization Using Sequence-to-Sequence RNNs and Beyond

Ramesh Nallapati; Bowen Zhou; Cícero Nogueira dos Santos; Caglar Gulcehre; Bing Xiang

In this work, we model abstractive text summarization using Attentional Encoder-Decoder Recurrent Neural Networks, and show that they achieve state-of-the-art performance on two different corpora. We propose several novel models that address critical problems in summarization that are not adequately modeled by the basic architecture, such as modeling key-words, capturing the hierarchy of sentence-to-word structure, and emitting words that are rare or unseen at training time. Our work shows that many of our proposed models contribute to further improvement in performance. We also propose a new dataset consisting of multi-sentence summaries, and establish performance benchmarks for further research.


meeting of the association for computational linguistics | 2015

Boosting Named Entity Recognition with Neural Character Embeddings

Cícero Nogueira dos Santos; Victor Guimarães

Most state-of-the-art named entity recognition (NER) systems rely on handcrafted features and on the output of other NLP tasks such as part-of-speech (POS) tagging and text chunking. In this work we propose a language-independent NER system that uses automatically learned features only. Our approach is based on the CharWNN deep neural network, which uses word-level and character-level representations (embeddings) to perform sequential classification. We perform an extensive number of experiments using two annotated corpora in two different languages: HAREM I corpus, which contains texts in Portuguese; and the SPA CoNLL-2002 corpus, which contains texts in Spanish. Our experimental results shade light on the contribution of neural character embeddings for NER. Moreover, we demonstrate that the same neural network which has been successfully applied to POS tagging can also achieve state-of-the-art results for language-independet NER, using the same hyperparameters, and without any handcrafted features. For the HAREM I corpus, CharWNN outperforms the state-of-the-art system by 7.9 points in the F1-score for the total scenario (ten NE classes), and by 7.2 points in the F1 for the selective scenario (five NE classes).


international joint conference on natural language processing | 2015

Learning Hybrid Representations to Retrieve Semantically Equivalent Questions

Cícero Nogueira dos Santos; Luciano Barbosa; Dasha Bogdanova; Bianca Zadrozny

Retrieving similar questions in online QA (2) BOW-CNN is more robust than the pure CNN for long texts.


meeting of the association for computational linguistics | 2016

Improved Representation Learning for Question Answer Matching

Ming Tan; Cícero Nogueira dos Santos; Bing Xiang; Bowen Zhou

Passage-level question answer matching is a challenging task since it requires effective representations that capture the complex semantic relations between questions and answers. In this work, we propose a series of deep learning models to address passage answer selection. To match passage answers to questions accommodating their complex semantic relations, unlike most previous work that utilizes a single deep learning structure, we develop hybrid models that process the text using both convolutional and recurrent neural networks, combining the merits on extracting linguistic information from both structures. Additionally, we also develop a simple but effective attention mechanism for the purpose of constructing better answer representations according to the input question, which is imperative for better modeling long answer sequences. The results on two public benchmark datasets, InsuranceQA and TREC-QA, show that our proposed models outperform a variety of strong baselines.


processing of the portuguese language | 2008

Portuguese Part-of-Speech Tagging Using Entropy Guided Transformation Learning

Cícero Nogueira dos Santos; Ruy Luiz Milidiú; Raúl P. Rentería

Entropy Guided Transformation Learning (ETL) is a new machine learning strategy that combines the advantages of Decision Trees (DT) and Transformation Based Learning (TBL). In this work, we apply the ETL framework to Portuguese Part-of-Speech Taggging. We use two different corpora: Mac-Morpho and Tycho Brahae. ETL achieves the best results reported so far for Machine Learning based POS tagging of both corpora. ETL provides a new training strategy that accelerates transformation learning. For the Mac-Morpho corpus this corresponds to a factor of three speedup. ETL shows accuracies of 96.75% and 96.64% for Mac-Morpho and Tycho Brahae, respectively.


multi agent systems and agent based simulation | 2013

Large-Scale Multi-agent-Based Modeling and Simulation of Microblogging-Based Online Social Network

Maira Athanazio de Cerqueira Gatti; Paulo Rodrigo Cavalin; Samuel Martins Barbosa Neto; Claudio S. Pinhanez; Cícero Nogueira dos Santos; Daniel Lemes Gribel; Ana Paula Appel

Online Social Networks (OSN) are self-organized systems with emergent behavior from the individual interactions. Microblogging services in OSN, like Twitter and Facebook, became extremely popular and are being used to target marketing campaigns. Key known issues on this targeting is to be able to predict human behavior like posting, forwarding or replying a message with regard to topics and sentiments, and to analyze the emergent behavior of such actions. To tackle this problem we present a method to model and simulate interactive behavior in microblogging OSN taking into account the users sentiment. We make use of a stochastic multi-agent based approach and we explore Barack Obama’s Twitter network as an egocentric network to present the experimental simulation results. We demonstrate that with this engineering method it is possible to develop social media simulators using a bottom-up approach (micro level) to evaluate the emergent behavior (macro level) and our preliminary results show how to better tune the modeler and the sampling and text classification impact on the simulation model.


conference on computational natural language learning | 2015

Detecting Semantically Equivalent Questions in Online User Forums

Dasha Bogdanova; Cícero Nogueira dos Santos; Luciano Barbosa; Bianca Zadrozny

Two questions asking the same thing could be too different in terms of vocabulary and syntactic structure, which makes identifying their semantic equivalence challenging. This study aims to detect semantically equivalent questions in online user forums. We perform an extensive number of experiments using data from two different Stack Exchange forums. We compare standard machine learning methods such as Support Vector Machines (SVM) with a convolutional neural network (CNN). The proposed CNN generates distributed vector representations for pairs of questions and scores them using a similarity metric. We evaluate in-domain word embeddings versus the ones trained with Wikipedia, estimate the impact of the training set size, and evaluate some aspects of domain adaptation. Our experimental results show that the convolutional neural network with in-domain word embeddings achieves high performance even with limited training data.


Computational Linguistics | 2014

Latent trees for coreference resolution

Eraldo R. Fernandes; Cícero Nogueira dos Santos; Ruy Luiz Milidiú

We describe a structure learning system for unrestricted coreference resolution that explores two key modeling techniques: latent coreference trees and automatic entropy-guided feature induction. The latent tree modeling makes the learning problem computationally feasible because it incorporates a meaningful hidden structure. Additionally, using an automatic feature induction method, we can efficiently build enhanced nonlinear models using linear model learning algorithms. We present empirical results that highlight the contribution of each modeling technique used in the proposed system. Empirical evaluation is performed on the multilingual unrestricted coreference CoNLL-2012 Shared Task datasets, which comprise three languages: Arabic, Chinese and English. We apply the same system to all languages, except for minor adaptations to some language-dependent features such as nested mentions and specific static pronoun lists. A previous version of this system was submitted to the CoNLL-2012 Shared Task closed track, achieving an official score of 58.69, the best among the competitors. The unique enhancement added to the current system version is the inclusion of candidate arcs linking nested mentions for the Chinese language. By including such arcs, the score increases by almost 4.5 points for that language. The current system shows a score of 60.15, which corresponds to a 3.5% error reduction, and is the best performing system for each of the three languages.


foundations of computational intelligence | 2012

Entropy Guided Transformation Learning

Cícero Nogueira dos Santos; Ruy Luiz Milidiú

This chapter details the entropy guided transformation learning algorithm [8, 23]. ETL is an effective way to overcome the transformation based learning bottleneck: the construction of good template sets. In order to better motivate and describe ETL, we first provide an overview of the TBL algorithm in Sect. 2.1. Next, in Sect. 2.2, we explain why the manual construction of template sets is a bottleneck for TBL. Then, in Sect. 2.3, we detail the entropy guided template generation strategy employed by ETL. In Sect. 2.3, we also present strategies to handle high dimensional features and to include the current classification feature in the generated templates. In Sects. 2.4–2.6 we present some variations on the basic ETL strategy. Finally, in Sect. 2.7, we discuss some related works.

Collaboration


Dive into the Cícero Nogueira dos Santos's collaboration.

Researchain Logo
Decentralizing Knowledge