Results 11 to 20 of about 19,938 (266)
The purpose of this paper is to introduce and analyze modified hybrid steepest-descent methods for a general system of variational inequalities (GSVI), with solutions being also zeros of an m-accretive operator A in the setting of real uniformly convex ...
Lu-Chuan Ceng, Ching-Feng Wen
doaj +1 more source
Long-Time Asymptotics of a Three-Component Coupled mKdV System
We present an application of the nonlinear steepest descent method to a three-component coupled mKdV system associated with a 4 × 4 matrix spectral problem. An integrable coupled mKdV hierarchy with three potentials is first generated. Based
Wen-Xiu Ma
doaj +1 more source
A Generalized Hybrid Steepest-Descent Method for Variational Inequalities in Banach Spaces
The hybrid steepest-descent method introduced by Yamada (2001) is an algorithmic solution to the variational inequality problem over the fixed point set of nonlinear mapping and applicable to a broad range of convexly constrained nonlinear inverse ...
Wong NC, Yao JC, Sahu DR
doaj +2 more sources
We establish the strong convergence of prediction-correction and relaxed hybrid steepest-descent method (PRH method) for variational inequalities under some suitable conditions that simplify the proof.
Haiwen Xu
doaj +1 more source
Overlooked Degree of Freedom in Steepest Descent Method: Steepest Descent Method Corresponding to Divergence-Free WKB Method [PDF]
1Research Organization of Science and Engineering, Ritsumeikan University, Kusatsu 525-8577, Japan 2Department of Applied Mathematics and Informatics, Ryukoku University, Otsu 520-2194, Japan 3Laboratoire de Genie Chimique, Bât 2R1 31062 Toulouse Cedex 9, Universite Paul Sabatier, France 4Department of Physical Sciences, Ritsumeikan University, Kusatsu
T. Hyouguchi, R. Seto, S. Adachi
openaire +1 more source
Bayesian optimization for computationally extensive probability distributions. [PDF]
An efficient method for finding a better maximizer of computationally extensive probability distributions is proposed on the basis of a Bayesian optimization technique.
Ryo Tamura, Koji Hukushima
doaj +1 more source
In this paper, it is aimed to computationally conduct a performance benchmarking for the steepest descent and the three well-known conjugate gradient methods (i.e., Fletcher-Reeves, Polak- Ribiere and Hestenes-Stiefel) along with six different step ...
Kadir Kiran
doaj
Steepest-Descent Approach to Triple Hierarchical Constrained Optimization Problems
We introduce and analyze a hybrid steepest-descent algorithm by combining Korpelevich’s extragradient method, the steepest-descent method, and the averaged mapping approach to the gradient-projection algorithm.
Lu-Chuan Ceng +3 more
doaj +1 more source
Hybrid Steepest Descent Method with Variable Parameters for General Variational Inequalities
We study the strong convergence of a hybrid steepest descent method with variable parameters for the general variational inequality (GVI) . Consequently, as an application, we obtain some results concerning the constrained generalized pseudoinverse. Our
Yu Yanrong, Chen Rudong
doaj +2 more sources
An Optimally Generalized Steepest-Descent Algorithm for Solving Ill-Posed Linear Systems
It is known that the steepest-descent method converges normally at the first few iterations, and then it slows down. We modify the original steplength and descent direction by an optimization argument with the new steplength as being a merit function to ...
Chein-Shan Liu
doaj +1 more source

