Xianzhi Wang
Harbin Institute of Technology
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Xianzhi Wang.
international acm sigir conference on research and development in information retrieval | 2015
Lina Yao; Quan Z. Sheng; Yongrui Qin; Xianzhi Wang; Ali Shemshadi; Qi He
Point-of-Interest (POI) recommendation is a new type of recommendation task that comes along with the prevalence of location-based social networks in recent years. Compared with traditional tasks, it focuses more on personalized, context-aware recommendation results to provide better user experience. To address this new challenge, we propose a Collaborative Filtering method based on Non-negative Tensor Factorization, a generalization of the Matrix Factorization approach that exploits a high-order tensor instead of traditional User-Location matrix to model multi-dimensional contextual information. The factorization of this tensor leads to a compact model of the data which is specially suitable for context-aware POI recommendations. In addition, we fuse users social relations as regularization terms of the factorization to improve the recommendation accuracy. Experimental results on real-world datasets demonstrate the effectiveness of our approach.
international conference on web services | 2013
Xianzhi Wang; Zhongjie Wang; Xiaofei Xu
As available services accumulate on the Internet, QoS-aware service selection (SSP) becomes an increasingly difficult task. Since Artificial Bee Colony algorithm (ABC) has been successful in solving many problems as a simpler implementation of swarm intelligence, its application to SSP is promising. However, ABC was initially designed for numerical optimization, and its effectiveness highly depends on what we call optimality continuity property of the solution space, i.e., similar variable values (or neighboring solutions) result in similar objective values (or evaluation results). We will show that SSP does not possess such property. We further propose an approximation approach based on greedy search strategies for ABC, to overcome this problem. In this approach, neighboring solutions are generated for a composition greedily based on the neighboring services of its component services. Two algorithms with different neighborhood measures are presented based on this approach. The resulting neighborhood structure of the proposed algorithms is analogical to that of continuous functions, so that the advantages of ABC can be fully leveraged in solving SSP. Also, they are pure online algorithms which are as simple as canonical ABC. The rationale of the proposed approach is discussed and the complexity of the algorithms is analyzed. Experiments conducted against canonical ABC indicate that the proposed algorithms can achieve better optimality within limited time.
conference on information and knowledge management | 2015
Xianzhi Wang; Quan Z. Sheng; Xiu Susie Fang; Lina Yao; Xiaofei Xu; Xue Li
Truth-finding is the fundamental technique for corroborating reports from multiple sources in both data integration and collective intelligent applications. Traditional truth-finding methods assume a single true value for each data item and therefore cannot deal will multiple true values (i.e., the multi-truth-finding problem). So far, the existing approaches handle the multi-truth-finding problem in the same way as the single-truth-finding problems. Unfortunately, the multi-truth-finding problem has its unique features, such as the involvement of sets of values in claims, different implications of inter-value mutual exclusion, and larger source profiles. Considering these features could provide new opportunities for obtaining more accurate truth-finding results. Based on this insight, we propose an integrated Bayesian approach to the multi-truth-finding problem, by taking these features into account. To improve the truth-finding efficiency, we reformulate the multi-truth-finding problem model based on the mappings between sources and (sets of) values. New mutual exclusive relations are defined to reflect the possible co-existence of multiple true values. A finer-grained copy detection method is also proposed to deal with sources with large profiles. The experimental results on three real-world datasets show the effectiveness of our approach.
international conference on web services | 2015
Lina Yao; Xianzhi Wang; Quan Z. Sheng; Wenjie Ruan; Wei Zhang
In this paper, we explore service recommendation and selection in the reusable composition context. The goal is to aid developers finding the most appropriate services in their composition tasks. We specifically focus on mashups, a domain that increasingly targets people without sophisticated programming knowledge. We propose a probabilistic matrix factorization approach with implicit correlation regularization to solve this problem. In particular, we advocate that the co-invocation of services in mashups is driven by both explicit textual similarity and implicit correlation of services, and therefore develop a latent variable model to uncover the latent connections between services by analyzing their co-invocation patterns. We crawled a real dataset from Programmable Web, and extensively evaluated the effectiveness of our proposed approach.
international conference on web services | 2011
Xianzhi Wang; Zhongjie Wang; Xiaofei Xu
Service composition has the capability of constructing coarse-grained solutions by dynamically aggregating a set of services to satisfy complex requirements, but it suffers from dramatic decrease on the efficiency of determining the best composition solution when large scale candidate services are available. Most current approaches look for the optimal composition solution by real-time computation, and the composition efficiency greatly depends on the adopted algorithms. To eliminate such deficiency, this paper proposes a semi-empirical composition approach which incorporates two stages, i.e., periodical clustering and real-time composition. The former partitions the candidate services and historical requirements into clusters based on similarity measurement, and then the probabilistic correspondences between service clusters and requirement clusters are identified by statistical analysis. The latter deals with a new requirement by firstly finding its most similar requirement cluster and the corresponding service clusters by leveraging Bayesian inference, then a set of concrete services are optimally selected from such reduced solution space and constitute the final composition solution. Instead of relying on solely historical data exploration or on pure real-time computation, our approach distinguishes from traditional methods by combining the two perspectives together. Experiments demonstrate the advantages of this approach.
international conference on web services | 2017
Chaoran Huang; Lina Yao; Xianzhi Wang; Boualem Benatallah; Quan Z. Sheng
Question answering (Q&A) communities have gained momentum recently as an effective means of knowledge sharing over the crowds, where many users are experts in the real-world and can make quality contributions in certain domains or technologies. Although the massive user-generated Q&A data present a valuable source of human knowledge, a related challenging issue is how to find those expert users effectively. In this paper, we propose a framework for finding such experts in a collaborative network. Accredited with recent works on distributed word representations, we are able to summarize text chunks from the semantics perspective and infer knowledge domains by clustering pre-trained word vectors. In particular, we exploit a graph-based clustering method for knowledge domain extraction and discern the shared latent factors using matrix factorization techniques. The proposed clustering method features requiring no post-processing of clustering indicators and the matrix factorization method is combined with the semantic similarity of the historical answers to conduct expertise ranking of users given a query. We use Stack Overflow, a website with a large group of users and a large number of posts on topics related to computer programming, to evaluate the proposed approach and conduct extensively experiments to show the effectiveness of our approach.
conference on information and knowledge management | 2015
Xianzhi Wang; Quan Z. Sheng; Xiu Susie Fang; Xue Li; Xiaofei Xu; Lina Yao
Many real-world applications rely on multiple data sources to provide information on their interested items. Due to the noises and uncertainty in data, given a specific item, the information from different sources may conflict. To make reliable decisions based on these data, it is important to identify the trustworthy information by resolving these conflicts, i.e., the truth discovery problem. Current solutions to this problem detect the veracity of each value jointly with the reliability of each source for each data item. In this way, the efficiency of truth discovery is strictly confined by the problem scale, which in turn limits truth discovery algorithms from being applicable on a large scale. To address this issue, we propose an approximate truth discovery approach, which divides sources and values into groups according to a user-specified approximation criterion. The groups are then used for efficient inter-value influence computation to improve the accuracy. Our approach is applicable to most existing truth discovery algorithms. Experiments on real-world datasets show that our approach improves the efficiency compared to existing algorithms while achieving similar or even better accuracy. The scalability is further demonstrated by experiments on large synthetic datasets.
ieee international conference on services computing | 2011
Xianzhi Wang; Zhongjie Wang; Xiaofei Xu
Service composition follows a three-party paradigm, i.e., a broker mediates between service providers and service consumers to properly select and compose a set of distributed services together so that requirements raised by consumers are satisfied by the composite service on demand. As the de facto provider of composite services, the broker charges the consumers, on the other hand, it awards cost to the providers whose services are involved in the composite services. Besides traditional quality-oriented optimization from the consumers point of view, the profit that a broker could earn from the composition is another objective to be optimized. But just as the quality optimization, service selection for profit optimization suffers from dramatic efficiency decline along with the growth in the number of candidate services. On the premise that the expected quality are guaranteed, this paper presents a divide and select approach for high-efficiency profit optimization, with price as heuristics. This approach can be applied to both static and dynamic pricing scenarios of service composition. Experiments demonstrate the feasibility.
IEEE Transactions on Mobile Computing | 2018
Lina Yao; Quan Z. Sheng; Xue Li; Tao Gu; Mingkui Tan; Xianzhi Wang; Sen Wang; Wenjie Ruan
Understanding and recognizing human activities is a fundamental research topic for a wide range of important applications such as fall detection and remote health monitoring and intervention. Despite active research in human activity recognition over the past years, existing approaches based on computer vision or wearable sensor technologies present several significant issues such as privacy (e.g., using video camera to monitor the elderly at home) and practicality (e.g., not possible for an older person with dementia to remember wearing devices). In this paper, we present a low-cost, unobtrusive, and robust system that supports independent living of older people. The system interprets what a person is doing by deciphering signal fluctuations using radio-frequency identification (RFID) technology and machine learning algorithms. To deal with noisy, streaming, and unstable RFID signals, we develop a compressive sensing, dictionary-based approach that can learn a set of compact and informative dictionaries of activities using an unsupervised subspace decomposition. In particular, we devise a number of approaches to explore the properties of sparse coefficients of the learned dictionaries for fully utilizing the embodied discriminative information on the activity recognition task. Our approach achieves efficient and robust activity recognition via a more compact and robust representation of activities. Extensive experiments conducted in a real-life residential environment demonstrate that our proposed system offers a good overall performance and shows the promising practical potential to underpin the applications for the independent living of the elderly.
world congress on services | 2010
Xianzhi Wang; Zhongjie Wang; Xiaofei Xu; Alice Liu; Dianhui Chu
Traditional service composition approaches focus on selecting and composing multiple service components together to fulfill one single requirement. But in most real-world scenarios, there are multiple requirements raised by multiple consumers and they form a discrete and uneven flow (i.e., a temporal sequence). Due to the limited number of available services and their limited capacities, how to ensure the equilibrium between the satisfaction degrees of these temporally sequential requirements becomes an important issue to be addressed. This paper proposes an equilibrium-oriented service composition approach taking into account both the limitedness of service capacity and utilization of historical data. The temporal sequential requirements are divided gradually along with the formation of length-flexible time-segments one by one. Based on this segmentation, service capacity is preserved proportionally for the estimated future requirements, and multiple requirements within one segment are ensured to get relatively equal chances of being satisfied with relatively equal quality. Experiments reveal improved sustainability and superior temporal stability of service quality compared with applying traditional methods to this scenario.