Results 31 to 40 of about 141,184 (170)

Design of Momentum Fractional Stochastic Gradient Descent for Recommender Systems

open access: yesIEEE Access, 2019
The demand for recommender systems in E-commerce industry has increased tremendously. Efficient recommender systems are being proposed by different E-business companies with the intention to give users accurate and most relevant recommendation of ...
Zeshan Aslam Khan   +4 more
doaj   +1 more source

Correspondence between neuroevolution and gradient descent

open access: yesNature Communications, 2021
Gradient-based and non-gradient-based methods for training neural networks are usually considered to be fundamentally different. The authors derive, and illustrate numerically, an analytic equivalence between the dynamics of neural network training under
Stephen Whitelam   +3 more
doaj   +1 more source

Stochastic gradient descent on GPUs [PDF]

open access: yesProceedings of the 8th Workshop on General Purpose Processing using GPUs, 2015
Irregular algorithms such as Stochastic Gradient Descent (SGD) can benefit from the massive parallelism available on GPUs. However, unlike in data-parallel algorithms, synchronization patterns in SGD are quite complex. Furthermore, scheduling for scale-free graphs is challenging.
Rashid Kaleem   +2 more
openaire   +1 more source

DEVELOPMENT OF R PACKAGE AND EXPERIMENTAL ANALYSIS ON PREDICTION OF THE CO2 COMPRESSIBILITY FACTOR USING GRADIENT DESCENT [PDF]

open access: yesJournal of Engineering Science and Technology, 2018
Nowadays, many variants of gradient descent (i.e., the methods included in machine learning for regression) have been proposed. Moreover, these algorithms have been widely used to deal with real-world problems.
LALA SEPTEM RIZA   +5 more
doaj  

Normalized stochastic gradient descent learning of general complex‐valued models

open access: yesElectronics Letters, 2021
The stochastic gradient descent (SGD) method is one of the most prominent first‐order iterative optimisation algorithms, enabling linear adaptive filters as well as general nonlinear learning schemes.
T. Paireder, C. Motz, M. Huemer
doaj   +1 more source

Komparasi Metode Optimasi Adam dan SGD dalam Skema Direct Inverse Control untuk Sistem Kendali Data Sikap dan Ketinggian Quadcopter

open access: yesJurnal Elkomika, 2022
ABSTRAK Terdapat banyak variable nonlinear dalam sistem kendali untuk quadcopter sehingga cukup rumit untuk mengatur dinamika penerbangan wahana ini. Untuk mengatasi masalah tersebut akan dikembangkan suatu skema sistem kendali Direct Inverse Control ...
MUHAMMAD SABILA HAQQI   +1 more
doaj   +1 more source

Damped Newton Stochastic Gradient Descent Method for Neural Networks Training

open access: yesMathematics, 2021
First-order methods such as stochastic gradient descent (SGD) have recently become popular optimization methods to train deep neural networks (DNNs) for good generalization; however, they need a long training time.
Jingcheng Zhou   +3 more
doaj   +1 more source

A Synchronized Gradient Descent Algorithm Based on Distributed Coding [PDF]

open access: yesJisuanji gongcheng, 2021
The Asynchronized Stochastic Gradient Descent(ASGD) algorithm based on data parallelization require frequent gradient data exchanges between distributed computing nodes,which affects the execution efficiency of the algorithm.This paper proposes a ...
LI Bowen, XIE Zaipeng, MAO Yingchi, XU Yuanyuan, ZHU Xiaorui, ZHANG Ji
doaj   +1 more source

Counterexamples for Noise Models of Stochastic Gradients

open access: yesExamples and Counterexamples, 2023
Stochastic Gradient Descent (SGD) is a widely used, foundational algorithm in data science and machine learning. As a result, analyses of SGD abound making use of a variety of assumptions, especially on the noise behavior of the stochastic gradients ...
Vivak Patel
doaj   +1 more source

Randomized Stochastic Gradient Descent Ascent

open access: yes, 2021
An increasing number of machine learning problems, such as robust or adversarial variants of existing algorithms, require minimizing a loss function that is itself defined as a maximum. Carrying a loop of stochastic gradient ascent (SGA) steps on the (inner) maximization problem, followed by an SGD step on the (outer) minimization, is known as Epoch ...
Sebbouh, Othmane   +2 more
openaire   +3 more sources

Home - About - Disclaimer - Privacy