Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Li-Zhi Liao is active.

Publication


Featured researches published by Li-Zhi Liao.


Mathematical Programming | 2002

A new inexact alternating directions method for monotone variational inequalities

Bingsheng He; Li-Zhi Liao; Deren Han; Hai Yang

Abstract.The alternating directions method (ADM) is an effective method for solving a class of variational inequalities (VI) when the proximal and penalty parameters in sub-VI problems are properly selected. In this paper, we propose a new ADM method which needs to solve two strongly monotone sub-VI problems in each iteration approximately and allows the parameters to vary from iteration to iteration. The convergence of the proposed ADM method is proved under quite mild assumptions and flexible parameter conditions.


Applied Mathematics and Optimization | 2001

New Conjugacy Conditions and Related Nonlinear Conjugate Gradient Methods

Yu-Hong Dai; Li-Zhi Liao

Abstract. Conjugate gradient methods are a class of important methods for unconstrained optimization, especially when the dimension is large. This paper proposes a new conjugacy condition, which considers an inexact line search scheme but reduces to the old one if the line search is exact. Based on the new conjugacy condition, two nonlinear conjugate gradient methods are constructed. Convergence analysis for the two methods is provided. Our numerical results show that one of the methods is very efficient for the given test problems.


Journal of Optimization Theory and Applications | 2002

Improvements of some projection methods for monotone nonlinear variational inequalities

Bingsheng He; Li-Zhi Liao

In this paper, we study the relationship of some projection-type methods for monotone nonlinear variational inequalities and investigate some improvements. If we refer to the Goldstein–Levitin–Polyak projection method as the explicit method, then the proximal point method is the corresponding implicit method. Consequently, the Korpelevich extragradient method can be viewed as a prediction-correction method, which uses the explicit method in the prediction step and the implicit method in the correction step. Based on the analysis in this paper, we propose a modified prediction-correction method by using better prediction and correction stepsizes. Preliminary numerical experiments indicate that the improvements are significant.


IEEE Transactions on Neural Networks | 2005

A novel neural network for variational inequalities with linear and nonlinear constraints

Xing-Bao Gao; Li-Zhi Liao; Liqun Qi

Variational inequality is a uniform approach for many important optimization and equilibrium problems. Based on the sufficient and necessary conditions of the solution, this paper presents a novel neural network model for solving variational inequalities with linear and nonlinear constraints. Three sufficient conditions are provided to ensure that the proposed network with an asymmetric mapping is stable in the sense of Lyapunov and converges to an exact solution of the original problem. Meanwhile, the proposed network with a gradient mapping is also proved to be stable in the sense of Lyapunov and to have a finite-time convergence under some mild conditions by using a new energy function. Compared with the existing neural networks, the new model can be applied to solve some nonmonotone problems, has no adjustable parameter, and has lower complexity. Thus, the structure of the proposed network is very simple. Since the proposed network can be used to solve a broad class of optimization problems, it has great application potential. The validity and transient behavior of the proposed neural network are demonstrated by several numerical examples.


SIAM Journal on Matrix Analysis and Applications | 1999

A Smoothing Newton Method for Extended Vertical Linear Complementarity Problems

Houduo Qi; Li-Zhi Liao

In this paper, we reformulate the extended vertical linear complementarity problem (EVLCP(m,q)) as a nonsmooth equation H(t,x)=0, where


Journal of Global Optimization archive | 2004

Neurodynamical Optimization

Li-Zhi Liao; Houduo Qi; Liqun Qi

H: \mbox{\smallBbb R}^{n+1} \to \mbox{\smallBbb R}^{n+1}


IEEE Transactions on Neural Networks | 2004

A neural network for a class of convex quadratic minimax problems with constraints

Xing-Bao Gao; Li-Zhi Liao; Weimin Xue

,


Journal of Global Optimization | 2001

Stability Analysis of Gradient-Based Neural Networks for Optimization Problems

Qiaoming Han; Li-Zhi Liao; Houduo Qi; Liqun Qi

t \in \mbox{\smallBbb R}


Journal of Optimization Theory and Applications | 2001

Decomposition method with a variable parameter for a class of monotone variational inequality problems

Shengli Wang; Li-Zhi Liao

is a parameter variable, and


Computational Optimization and Applications | 2000

A Smoothing Newton Method for General Nonlinear Complementarity Problems

Houduo Qi; Li-Zhi Liao

x \in \mbox{\smallBbb R}

Collaboration


Dive into the Li-Zhi Liao's collaboration.

Top Co-Authors

Avatar

Michael K. Ng

Hong Kong Baptist University

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Xing-Bao Gao

Shaanxi Normal University

View shared research outputs
Top Co-Authors

Avatar

Lei-Hong Zhang

Shanghai University of Finance and Economics

View shared research outputs
Top Co-Authors

Avatar

Delin Chu

National University of Singapore

View shared research outputs
Top Co-Authors

Avatar

Liqun Qi

Hong Kong Polytechnic University

View shared research outputs
Top Co-Authors

Avatar

Houduo Qi

University of Southampton

View shared research outputs
Top Co-Authors

Avatar

Duan Li

The Chinese University of Hong Kong

View shared research outputs
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Hon Wah Tam

Hong Kong Baptist University

View shared research outputs
Researchain Logo
Decentralizing Knowledge