Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Richard D. Lawrence is active.

Publication


Featured researches published by Richard D. Lawrence.


knowledge discovery and data mining | 2009

Sentiment analysis of blogs by combining lexical knowledge with text classification

Prem Melville; Wojciech Gryc; Richard D. Lawrence

The explosion of user-generated content on the Web has led to new opportunities and significant challenges for companies, that are increasingly concerned about monitoring the discussion around their products. Tracking such discussion on weblogs, provides useful insight on how to improve products or market them more effectively. An important component of such analysis is to characterize the sentiment expressed in blogs about specific brands and products. Sentiment Analysis focuses on this task of automatically identifying whether a piece of text expresses a positive or negative opinion about the subject matter. Most previous work in this area uses prior lexical knowledge in terms of the sentiment-polarity of words. In contrast, some recent approaches treat the task as a text classification problem, where they learn to classify sentiment based only on labeled training data. In this paper, we present a unified framework in which one can use background lexical information in terms of word-class associations, and refine this information for specific domains using any available training examples. Empirical results on diverse domains show that our approach performs better than using background knowledge or training data in isolation, as well as alternative approaches to using lexical knowledge with text classification.


Data Mining and Knowledge Discovery | 2001

Personalization of Supermarket Product Recommendations

Richard D. Lawrence; George S. Almasi; Vladimir Kotlyar; Marisa S. Viveros; Sastry S. Duri

We describe a personalized recommender system designed to suggest new products to supermarket shoppers. The recommender functions in a pervasive computing environment, namely, a remote shopping system in which supermarket customers use Personal Digital Assistants (PDAs) to compose and transmit their orders to the store, which assembles them for subsequent pickup. The recommender is meant to provide an alternative source of new ideas for customers who now visit the store less frequently. Recommendations are generated by matching products to customers based on the expected appeal of the product and the previous spending of the customer. Associations mining in the product domain is used to determine relationships among product classes for use in characterizing the appeal of individual products. Clustering in the customer domain is used to identify groups of shoppers with similar spending histories. Cluster-specific lists of popular products are then used as input to the matching process.The recommender is currently being used in a pilot program with several hundred customers. Analysis of results to date have shown a 1.8% boost in program revenue as a result of purchases made directly from the list of recommended products. A substantial fraction of the accepted recommendations are from product classes new to the customer, indicating a degree of willingness to expand beyond present purchase patterns in response to reasonable suggestions.


Data Mining and Knowledge Discovery | 1999

A Scalable Parallel Algorithm for Self-Organizing Maps with Applicationsto Sparse Data Mining Problems

Richard D. Lawrence; George S. Almasi; Holly E. Rushmeier

We describe a scalable parallel implementation of the self organizing map (SOM) suitable for data-mining applications involving clustering or segmentation against large data sets such as those encountered in the analysis of customer spending patterns. The parallel algorithm is based on the batch SOM formulation in which the neural weights are updated at the end of each pass over the training data. The underlying serial algorithm is enhanced to take advantage of the sparseness often encountered in these data sets. Analysis of a realistic test problem shows that the batch SOM algorithm captures key features observed using the conventional on-line algorithm, with comparable convergence rates.Performance measurements on an SP2 parallel computer are given for two retail data sets and a publicly available set of census data.These results demonstrate essentially linear speedup for the parallel batch SOM algorithm, using both a memory-contained sparse formulation as well as a separate implementation in which the mining data is accessed directly from a parallel file system. We also present visualizations of the census data to illustrate the value of the clustering information obtained via the parallel SOM method.


Ibm Systems Journal | 2000

Java programming for high-performance numerical computing

José E. Moreira; Samuel P. Midkiff; Manish Gupta; Pedro V. Artigas; Marc Snir; Richard D. Lawrence

First proposed as a mechanism for enhancing Web content, the JavaTM language has taken off as a serious general-purpose programming language. Industry and academia alike have expressed great interest in using the Java language as a programming language for scientific and engineering computations. Applications in these domains are characterized by intensive numerical computing and often have very high performance requirements. In this paper we discuss programming techniques that lead to Java numerical codes with performance comparable to FORTRAN or C, the more traditional languages for this field. The techniques are centered around the use of a high-performance numerical library, written entirely in the Java language, and on compiler technology. The numerical library takes the form of the Array package for Java. Proper use of this package, and of other appropriate tools for compiling and running a Java application, results in code that is clean, portable, and fast. We illustrate the programming and performance issues through case studies in data mining and electromagnetism.


Ibm Systems Journal | 2002

Applications of flexible pricing in business-to-business electronic commerce

Martin Bichler; Jayant R. Kalagnanam; Kaan Katircioglu; Alan J. King; Richard D. Lawrence; Ho Soo Lee; Grace Y. Lin; Yingdong Lu

The increasingly dynamic nature of business-to-business electronic commerce has produced a recent shift away from fixed pricing and toward flexible pricing. Flexible pricing, as defined here, includes both differential pricing, in which different buyers may receive different prices based on expected valuations, and dynamic-pricing mechanisms, such as auctions, where prices and conditions are based on bids by market participants. In this paper we survey ongoing work in flexible pricing in the context of the supply chain, including revenue management, procurement, and supply-chain coordination. We review negotiation mechanisms for procurement, including optimization approaches to the evaluation of complex, multidimensional bids. We also discuss several applications of flexible pricing on the sell side, including pricing strategies for response to requests for quotes, dynamic pricing in a reverse logistics application, and pricing in the emerging area of hosted applications services. We conclude with a discussion of future research directions in this rapidly growing area.


Computers & Operations Research | 2007

A logistic regression framework for information technology outsourcing lifecycle management

Aleksandra Mojsilovic; Bonnie K. Ray; Richard D. Lawrence; Samer Takriti

We present a methodology for managing outsourcing projects from the vendors perspective, designed to maximize the value to both the vendor and its clients. The methodology is applicable across the outsourcing lifecycle, providing the capability to select and target new clients, manage the existing client portfolio and quantify the realized benefits to the client resulting from the outsourcing agreement. Specifically, we develop a statistical analysis framework to model client behavior at each stage of the outsourcing lifecycle, including: (1) a predictive model and tool for white space client targeting and selection-opportunity identification (2) a model and tool for client risk assessment and project portfolio management-client tracking, and (3) a systematic analysis of outsourcing results, impact analysis, to gain insights into potential benefits of IT outsourcing as a part of a successful management strategy. Our analysis is formulated in a logistic regression framework, modified to allow for non-linear input-output relationships, auxiliary variables, and small sample sizes. We provide examples to illustrate how the methodology has been successfully implemented for targeting, tracking, and assessing outsourcing clients within IBM global services division. Scope and purpose The predominant literature on IT outsourcing often examines various aspects of vendor-client relationship, strategies for successful outsourcing from the client perspective, and key sources of risk to the client, generally ignoring the risk to the vendor. However, in the rapidly changing market, a significant share of risks and responsibilities falls on vendor, as outsourcing contracts are often renegotiated, providers replaced, or services brought back in house. With the transformation of outsourcing engagements, the risk on the vendors side has increased substantially, driving the vendors financial and business performance and eventually impacting the value delivery to the client. As a result, only well-ran vendor firms with robust processes and tools that allow identification and active management of risk at all stages of the outsourcing lifecycle are able to deliver value to the client. This paper presents a framework and methodology for managing a portfolio of outsourcing projects from the vendors perspective, throughout the entire outsourcing lifecycle. We address three key stages of the outsourcing process: (1) opportunity identification and qualification (i.e. selection of the most likely new clients), (2) client portfolio risk management during engagement and delivery, and (3) quantification of benefits to the client throughout the life of the deal.


knowledge discovery and data mining | 2011

Multi-view transfer learning with a large margin approach

Dan Zhang; Jingrui He; Yan Liu; Luo Si; Richard D. Lawrence

Transfer learning has been proposed to address the problem of scarcity of labeled data in the target domain by leveraging the data from the source domain. In many real world applications, data is often represented from different perspectives, which correspond to multiple views. For example, a web page can be described by its contents and its associated links. However, most existing transfer learning methods fail to capture the multi-view {nature}, and might not be best suited for such applications. To better leverage both the labeled data from the source domain and the features from different views, {this paper proposes} a general framework: Multi-View Transfer Learning with a Large Margin Approach (MVTL-LM). On one hand, labeled data from the source domain is effectively utilized to construct a large margin classifier; on the other hand, the data from both domains is employed to impose consistencies among multiple views. As an instantiation of this framework, we propose an efficient optimization method, which is guaranteed to converge to ε precision in O(1/ε) steps. Furthermore, we analyze its error bound, which improves over existing results of related methods. An extensive set of experiments are conducted to demonstrate the advantages of our proposed method over state-of-the-art techniques.


parallel computing | 1994

The IBM external user interface for scalable parallel systems

Vasanth Bala; Jehoshua Bruck; Raymond M. Bryant; Robert Cypher; Peter de Jong; Pablo Elustondo; D. Frye; Alex Ho; Ching-Tien Ho; Gail Blumenfeld Irwin; Shlomo Kipnis; Richard D. Lawrence; Marc Snir

Abstract The IBM External User Interface (EUI) for scalable parallel systems is a parallel programming library designed for the IBM line of scalable parallel computers. The first computer in this line, the IBM 9076 SP1, was announced in February 1993. In essence, the EUI is a library of coordination and communication routines that can be invoked from within FORTRAN or C application programs. The EUI consists of four main components: task management routines, message passing routines, task group routines, and collective communication routines. This paper examines several aspects of the design and development of the EUI.


international conference on machine learning | 2009

Uncertainty sampling and transductive experimental design for active dual supervision

Vikas Sindhwani; Prem Melville; Richard D. Lawrence

Dual supervision refers to the general setting of learning from both labeled examples as well as labeled features. Labeled features are naturally available in tasks such as text classification where it is frequently possible to provide domain knowledge in the form of words that associate strongly with a class. In this paper, we consider the novel problem of active dual supervision, or, how to optimally query an example and feature labeling oracle to simultaneously collect two different forms of supervision, with the objective of building the best classifier in the most cost effective manner. We apply classical uncertainty and experimental design based active learning schemes to graph/kernel-based dual supervision models. Empirical studies confirm the potential of these schemes to significantly reduce the cost of acquiring labeled data for training high-quality models.


conference on information and knowledge management | 2012

Learning to rank for robust question answering

Arvind Agarwal; Hema Raghavan; Karthik Subbian; Prem Melville; Richard D. Lawrence; David Gondek; James Fan

This paper aims to solve the problem of improving the ranking of answer candidates for factoid based questions in a state-of-the-art Question Answering system. We first provide an extensive comparison of 5 ranking algorithms on two datasets -- from the Jeopardy quiz show and a medical domain. We then show the effectiveness of a cascading approach, where the ranking produced by one ranker is used as input to the next stage. The cascading approach shows sizeable gains on both datasets. We finally evaluate several rank aggregation techniques to combine these algorithms, and find that Supervised Kemeny aggregation is a robust technique that always beats the baseline ranking approach used by Watson for the Jeopardy competition. We further corroborate our results on TREC Question Answering datasets.

Collaboration


Dive into the Richard D. Lawrence's collaboration.

Top Co-Authors

Avatar

Yan Liu

University of Southern California

View shared research outputs
Top Co-Authors

Avatar

Jingrui He

Arizona State University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge