This work investigates the integration of multiplicative calculus into gradient descent algorithms, including Adaptive Gradient algorithm (AdaGrad), Root Mean Squared Propagation (RMSProp), Nesterov Accelerated Gradient (NAG), and Momentum, to optimize exponential-quadratic-logarithmic composite functions with the positivity constrained. This research, conducted across five scenarios within the Constrained and Unconstrained Testing Environment (CUTEst), compares these multiplicative methods with their classical counterparts under a variety of constraints environments such as bounded, quadratic, and other types, and unconstrained environments. The results demonstrate the significant superiority of multiplicative-based algorithms, especially in unconstrained and bounded constrained scenarios, and demonstrate their potential for complex optimization tasks. Statistical analysis supports the observed performance advantages, indicating significant opportunities for optimization strate-gies in positive domains.
Gradient Descent Optimization Multiplicative calculus Machine Learning Composite Functions Non-newtonian Calculus
Birincil Dil | İngilizce |
---|---|
Konular | Performans Değerlendirmesi |
Bölüm | Makaleler |
Yazarlar | |
Erken Görünüm Tarihi | 28 Haziran 2024 |
Yayımlanma Tarihi | 30 Haziran 2024 |
Gönderilme Tarihi | 12 Nisan 2024 |
Kabul Tarihi | 16 Haziran 2024 |
Yayımlandığı Sayı | Yıl 2024 |