Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Yong Zhuang is active.

Publication


Featured researches published by Yong Zhuang.


conference on recommender systems | 2013

A fast parallel SGD for matrix factorization in shared memory systems

Yong Zhuang; Wei-Sheng Chin; Yu-Chin Juan; Chih-Jen Lin

Matrix factorization is known to be an effective method for recommender systems that are given only the ratings from users to items. Currently, stochastic gradient descent (SGD) is one of the most popular algorithms for matrix factorization. However, as a sequential approach, SGD is difficult to be parallelized for handling web-scale problems. In this paper, we develop a fast parallel SGD method, FPSGD, for shared memory systems. By dramatically reducing the cache-miss rate and carefully addressing the load balance of threads, FPSGD is more efficient than state-of-the-art parallel algorithms for matrix factorization.


conference on recommender systems | 2016

Field-aware Factorization Machines for CTR Prediction

Yu-Chin Juan; Yong Zhuang; Wei-Sheng Chin; Chih-Jen Lin

Click-through rate (CTR) prediction plays an important role in computational advertising. Models based on degree-2 polynomial mappings and factorization machines (FMs) are widely used for this task. Recently, a variant of FMs, field-aware factorization machines (FFMs), outperforms existing models in some world-wide CTR-prediction competitions. Based on our experiences in winning two of them, in this paper we establish FFMs as an effective method for classifying large sparse data including those from CTR prediction. First, we propose efficient implementations for training FFMs. Then we comprehensively analyze FFMs and compare this approach with competing models. Experiments show that FFMs are very useful for certain classification problems. Finally, we have released a package of FFMs for public use.


ACM Transactions on Intelligent Systems and Technology | 2015

A Fast Parallel Stochastic Gradient Method for Matrix Factorization in Shared Memory Systems

Wei-Sheng Chin; Yong Zhuang; Yu-Chin Juan; Chih-Jen Lin

Matrix factorization is known to be an effective method for recommender systems that are given only the ratings from users to items. Currently, stochastic gradient (SG) method is one of the most popular algorithms for matrix factorization. However, as a sequential approach, SG is difficult to be parallelized for handling web-scale problems. In this article, we develop a fast parallel SG method, FPSG, for shared memory systems. By dramatically reducing the cache-miss rate and carefully addressing the load balance of threads, FPSG is more efficient than state-of-the-art parallel algorithms for matrix factorization.


pacific-asia conference on knowledge discovery and data mining | 2015

A Learning-Rate Schedule for Stochastic Gradient Methods to Matrix Factorization

Wei-Sheng Chin; Yong Zhuang; Yu-Chin Juan; Chih-Jen Lin

Stochastic gradient methods are effective to solve matrix factorization problems. However, it is well known that the performance of stochastic gradient method highly depends on the learning rate schedule used; a good schedule can significantly boost the training process. In this paper, motivated from past works on convex optimization which assign a learning rate for each variable, we propose a new schedule for matrix factorization. The experiments demonstrate that the proposed schedule leads to faster convergence than existing ones. Our schedule uses the same parameter on all data sets included in our experiments; that is, the time spent on learning rate selection can be significantly reduced. By applying this schedule to a state-of-the-art matrix factorization package, the resulting implementation outperforms available parallel matrix factorization packages.


pacific-asia conference on knowledge discovery and data mining | 2015

Distributed Newton Methods for Regularized Logistic Regression

Yong Zhuang; Wei-Sheng Chin; Yu-Chin Juan; Chih-Jen Lin

Regularized logistic regression is a very useful classification method, but for large-scale data, its distributed training has not been investigated much. In this work, we propose a distributed Newton method for training logistic regression. Many interesting techniques are discussed for reducing the communication cost and speeding up the computation. Experiments show that the proposed method is competitive with or even faster than state-of-the-art approaches such as Alternating Direction Method of Multipliers (ADMM) and Vowpal Wabbit (VW). We have released an MPI-based implementation for public use.


knowledge discovery and data mining | 2013

Combination of feature engineering and ranking models for paper-author identification in KDD Cup 2013

Chun-Liang Li; Ting-Wei Lin; Cheng-Hao Tsai; Wei-Cheng Chang; Kuan-Hao Huang; Tzu-Ming Kuo; Shan-Wei Lin; Young-San Lin; Yu-Chen Lu; Chun-Pai Yang; Cheng-Xia Chang; Wei-Sheng Chin; Yu-Chin Juan; Hsiao-Yu Tung; Jui-Pin Wang; Cheng-Kuang Wei; Felix Wu; Tu-Chun Yin; Tong Yu; Yong Zhuang; Shou-De Lin; Hsuan-Tien Lin; Chih-Jen Lin

The track 1 problem in KDD Cup 2013 is to discriminate between papers confirmed by the given authors from the other deleted papers. This paper describes the winning solution of team National Taiwan University for track 1 of KDD Cup 2013. First, we conduct the feature engineering to transform the various provided text information into 97 features. Second, we train classification and ranking models using these features. Last, we combine our individual models to boost the performance by using results on the internal validation set and the official Valid set. Some effective post-processing techniques have also been proposed. Our solution achieves 0.98259 MAP score and ranks the first place on the private leaderboard of Test set.


conference on information and knowledge management | 2018

Naive Parallelization of Coordinate Descent Methods and an Application on Multi-core L1-regularized Classification

Yong Zhuang; Yu-Chin Juan; Guo-Xun Yuan; Chih-Jen Lin

It is well known that a direct parallelization of sequential optimization methods (e.g., coordinate descent and stochastic gradient methods) is often not effective. The reason is that at each iteration, the number of operations may be too small. We point out that this common understanding may not be true if the algorithm sequentially accesses the data in a feature-wise manner. For almost all real-world sparse sets we have examined, some features are much denser than others. Thus a direct parallelization of loops in a sequential method may result in excellent speedup. This approach possesses an advantage of retaining all convergence results because the algorithm is not changed at all. We apply this idea on coordinate descent (CD) methods, which are effective single-thread technique for L1-regularized classification. Further, an investigation on the shrinking technique commonly used to remove some features in the training process shows that this technique helps the parallelization of CD methods. Experiments indicate that a naive parallelization achieves better speedup than existing methods that laboriously modify the algorithm to achieve parallelism. Though a bit ironic, we conclude that the naive parallelization of the CD method is a highly competitive and robust multi-core implementation for L1-regularized classification.


Journal of Machine Learning Research | 2014

Effective string processing and matching for author disambiguation

Wei-Sheng Chin; Yong Zhuang; Yu-Chin Juan; Felix Wu; Hsiao-Yu Tung; Tong Yu; Jui-Pin Wang; Cheng-Xia Chang; Chun-Pai Yang; Wei-Cheng Chang; Kuan-Hao Huang; Tzu-Ming Kuo; Shan-Wei Lin; Young-San Lin; Yu-Chen Lu; Cheng-Kuang Wei; Tu-Chun Yin; Chun-Liang Li; Ting-Wei Lin; Cheng-Hao Tsai; Shou-De Lin; Hsuan-Tien Lin; Chih-Jen Lin


Journal of Machine Learning Research | 2016

LIBMF: a library for parallel matrix factorization in shared-memory systems

Wei-Sheng Chin; Bo-Wen Yuan; Meng-Yuan Yang; Yong Zhuang; Yu-Chin Juan; Chih-Jen Lin


Journal of Machine Learning Research | 2015

Combination of feature engineering and ranking models for paper-author identification in KDD cup 2013

Chun-Liang Li; Ting-Wei Lin; Cheng-Hao Tsai; Wei-Cheng Chang; Kuan-Hao Huang; Tzu-Ming Kuo; Shan-Wei Lin; Young-San Lin; Yu-Chen Lu; Chun-Pai Yang; Cheng-Xia Chang; Wei-Sheng Chin; Yu-Chin Juan; Hsiao-Yu Tung; Jui-Pin Wang; Cheng-Kuang Wei; Felix Wu; Tu-Chun Yin; Tong Yu; Yong Zhuang; Shou-De Lin; Hsuan-Tien Lin; Chih-Jen Lin

Collaboration


Dive into the Yong Zhuang's collaboration.

Top Co-Authors

Avatar

Chih-Jen Lin

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Yu-Chin Juan

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Wei-Sheng Chin

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Cheng-Hao Tsai

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Cheng-Kuang Wei

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Cheng-Xia Chang

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Chun-Liang Li

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Chun-Pai Yang

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Felix Wu

National Taiwan University

View shared research outputs
Top Co-Authors

Avatar

Hsiao-Yu Tung

National Taiwan University

View shared research outputs
Researchain Logo
Decentralizing Knowledge