Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Somayajulu Sripada is active.

Publication


Featured researches published by Somayajulu Sripada.


Artificial Intelligence | 2005

Choosing words in computer-generated weather forecasts

Ehud Reiter; Somayajulu Sripada; Jim Hunter; Jin Yu; Ian P. Davy

One of the main challenges in automatically generating textual weather forecasts is choosing appropriate English words to communicate numeric weather data. A corpus-based analysis of how humans write forecasts showed that there were major differences in how individual writers performed this task, that is, in how they translated data into words. These differences included both different preferences between potential near-synonyms that could be used to express information, and also differences in the meanings that individual writers associated with specific words. Because we thought these differences could confuse readers, we built our SumTime-Mousam weather-forecast generator to use consistent data-to-word rules, which avoided words which were only used by a few people, and words which were interpreted differently by different people. An evaluation by forecast users suggested that they preferred SumTime-Mousams texts to human-generated texts, in part because of better word choice; this may be the first time that an evaluation has shown that nlg texts are better than human-authored texts.


Ai Communications | 2009

From data to text in the Neonatal Intensive Care Unit: Using NLG technology for decision support and information management

Albert Gatt; François Portet; Ehud Reiter; Jim Hunter; Saad Mahamood; Wendy Moncur; Somayajulu Sripada

Contemporary Neonatal Intensive Care Units collect vast amounts of patient data in various formats, making efficient processing of information by medical professionals difficult. Moreover, different stakeholders in the neonatal scenario, which include parents as well as staff occupying different roles, have different information requirements. This paper describes recent and ongoing work on building systems that automatically generate textual summaries of neonatal data. Our evaluation results show that the technology is viable and comparable in its effectiveness for decision support to existing presentation modalities. We discuss the lessons learned so far, as well as the major challenges involved in extending current technology to deal with a broader range of data types, and to improve the textual output in the form of more coherent summaries.


Journal of Artificial Intelligence Research | 2003

Acquiring correct knowledge for natural language generation

Ehud Reiter; Somayajulu Sripada; Roma Robertson

Natural language generation (NLG) systems are computer software systems that produce texts in English and other human languages, often from non-linguistic input data. NLG systems, like most ai systems, need substantial amounts of knowledge. However, our experience in two NLG projects suggests that it is difficult to acquire correct knowledge for NLG systems; indeed, every knowledge acquisition (KA) technique we tried had significant problems. In general terms, these problems were due to the complexity, novelty, and poorly understood nature of the tasks our systems attempted, and were worsened by the fact that people write so differently. This meant in particular that corpus-based KA approaches suffered because it was impossible to assemble a sizable corpus of high-quality consistent manually written texts in our domains; and structured expert-oriented KA techniques suffered because experts disagreed and because we could not get enough information about special and unusual cases to build robust systems. We believe that such problems are likely to affect many other NLG systems as well. In the long term, we hope that new KA techniques may emerge to help NLG system builders. In the shorter term, we believe that understanding how individual KA techniques can fail, and using a mixture of different KA techniques with different strengths and weaknesses, can help developers acquire NLG knowledge that is mostly correct.


Computational Linguistics | 2002

Squibs and discussions: human variation and lexical choice

Ehud Reiter; Somayajulu Sripada

Much natural language processing research implicitly assumes that word meanings are fixed in a language community, but in fact there is good evidence that different people probably associate slightly different meanings with words. We summarize some evidence for this claim from the literature and from an ongoing research project, and discuss its implications for natural language generation, especially for lexical choice, that is, choosing appropriate words for a generated text.


knowledge discovery and data mining | 2003

Generating English summaries of time series data using the Gricean maxims

Somayajulu Sripada; Ehud Reiter; Jim Hunter; Jin Yu

We are developing technology for generating English textual summaries of time-series data, in three domains: weather forecasts, gas-turbine sensor readings, and hospital intensive care data. Our weather-forecast generator is currently operational and being used daily by a meteorological company. We generate summaries in three steps: (a) selecting the most important trends and patterns to communicate; (b) mapping these patterns onto words and phrases; and (c) generating actual texts based on these words and phrases. In this paper we focus on the first step, (a), selecting the information to communicate, and describe how we perform this using modified versions of standard data analysis algorithms such as segmentation. The modifications arose out of empirical work with users and domain experts, and in fact can all be regarded as applications of the Gricean maxims of Quality, Quantity, Relevance, and Manner, which describe how a cooperative speaker should behave in order to help a hearer correctly interpret a text. The Gricean maxims are perhaps a key element of adapting data analysis algorithms for effective communication of information to human users, and should be considered by other researchers interested in communicating data to human users.


Artificial Intelligence in Medicine | 2012

Automatic generation of natural language nursing shift summaries in neonatal intensive care: BT-Nurse

Jim Hunter; Yvonne Freer; Albert Gatt; Ehud Reiter; Somayajulu Sripada; Cindy Sykes

INTRODUCTION Our objective was to determine whether and how a computer system could automatically generate helpful natural language nursing shift summaries solely from an electronic patient record system, in a neonatal intensive care unit (NICU). METHODS A system was developed which automatically generates partial NICU shift summaries (for the respiratory and cardiovascular systems), using data-to-text technology. It was evaluated for 2 months in the NICU at the Royal Infirmary of Edinburgh, under supervision. RESULTS In an on-ward evaluation, a substantial majority of the summaries was found by outgoing and incoming nurses to be understandable (90%), and a majority was found to be accurate (70%), and helpful (59%). The evaluation also served to identify some outstanding issues, especially with regard to extra content the nurses wanted to see in the computer-generated summaries. CONCLUSIONS It is technically possible automatically to generate limited natural language NICU shift summaries from an electronic patient record. However, it proved difficult to handle electronic data that was intended primarily for display to the medical staff, and considerable engineering effort would be required to create a deployable system from our proof-of-concept software.


artificial intelligence in medicine in europe | 2007

Automatic Generation of Textual Summaries from Neonatal Intensive Care Data

François Portet; Ehud Reiter; Jim Hunter; Somayajulu Sripada

Intensive care is becoming increasingly complex. If mistakes are to be avoided, there is a need for the large amount of clinical data to be presented effectively to the medical staff. Although the most common approach is to present the data graphically, it has been shown that textual summarisation can lead to improved decision making. As the first step in the BabyTalk project, a prototype is being developed which will generate a textual summary of 45 minutes of continuous physiological signals and discrete events (e.g.: equipment settings and drug administration). Its architecture brings together techniques from the different areas of signal analysis, medical reasoning, and natural language generation. Although the current system is still being improved, it is powerful enough to generate meaningful texts containing the most relevant information. This prototype will be extended to summarize several hours of data and to include clinical interpretation.


Twenty-second SGAI International Conference on Knowledge Based Systems and Applied Artificial Intelligence | 2003

Segmenting Time Series for Weather Forecasting

Somayajulu Sripada; Ehud Reiter; Jim Hunter; Jin Yu

We are investigating techniques for producing textual summaries of time series data. Deep reasoning techniques have proven impractical because we lack perfect knowledge about users and their tasks. Data analysis techniques such as segmentation are more attractive, but they have been developed for data mining, not for communication. We examine how segmentation should be modified to make it suitable for generating textual summaries. Our algorithm has been implemented in a weather forecast generation system.


natural language generation | 2001

A two-stage model for content determination

Somayajulu Sripada; Ehud Reiter; Jim Hunter; Jin Yu

In this paper we describe a two-stage model for content determination in systems that summarise time series data. The first stage involves building a qualitative overview of the data set, and the second involves using this overview, together with the actual data, to produce summaries of the time-series data. This model is based on our observations of how human experts summarise time-series data.


Journal of the American Medical Informatics Association | 2011

BT-Nurse: computer generation of natural language shift summaries from complex heterogeneous medical data

Jim Hunter; Yvonne Freer; Albert Gatt; Ehud Reiter; Somayajulu Sripada; Cindy Sykes; Dave Westwater

The BT-Nurse system uses data-to-text technology to automatically generate a natural language nursing shift summary in a neonatal intensive care unit (NICU). The summary is solely based on data held in an electronic patient record system, no additional data-entry is required. BT-Nurse was tested for two months in the Royal Infirmary of Edinburgh NICU. Nurses were asked to rate the understandability, accuracy, and helpfulness of the computer-generated summaries; they were also asked for free-text comments about the summaries. The nurses found the majority of the summaries to be understandable, accurate, and helpful (p<0.001 for all measures). However, nurses also pointed out many deficiencies, especially with regard to extra content they wanted to see in the computer-generated summaries. In conclusion, natural language NICU shift summaries can be automatically generated from an electronic patient record, but our proof-of-concept software needs considerable additional development work before it can be deployed.

Collaboration


Dive into the Somayajulu Sripada's collaboration.

Top Co-Authors

Avatar

Ehud Reiter

University of Aberdeen

View shared research outputs
Top Co-Authors

Avatar

Jim Hunter

University of Aberdeen

View shared research outputs
Top Co-Authors

Avatar

Jin Yu

University of Aberdeen

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

François Portet

Centre national de la recherche scientifique

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Cindy Sykes

Edinburgh Royal Infirmary

View shared research outputs
Researchain Logo
Decentralizing Knowledge