This work investigates the integration of multiplicative calculus into gradient descent algorithms, including Adaptive Gradient algorithm (AdaGrad), Root Mean Squared Propagation (RMSProp), Nesterov Accelerated Gradient (NAG), and Momentum, to optimize exponential-quadratic-logarithmic composite functions with the positivity constrained. This research, conducted across five scenarios within the Constrained and Unconstrained Testing Environment (CUTEst), compares these multiplicative methods with their classical counterparts under a variety of constraints environments such as bounded, quadratic, and other types, and unconstrained environments. The results demonstrate the significant superiority of multiplicative-based algorithms, especially in unconstrained and bounded constrained scenarios, and demonstrate their potential for complex optimization tasks. Statistical analysis supports the observed performance advantages, indicating significant opportunities for optimization strate-gies in positive domains.
Gradient Descent Optimization Multiplicative calculus Machine Learning Composite Functions Non-newtonian Calculus
Primary Language | English |
---|---|
Subjects | Performance Evaluation |
Journal Section | Articles |
Authors | |
Early Pub Date | June 28, 2024 |
Publication Date | June 30, 2024 |
Submission Date | April 12, 2024 |
Acceptance Date | June 16, 2024 |
Published in Issue | Year 2024 Volume: 10 Issue: 1 |