{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,16]],"date-time":"2025-05-16T04:08:06Z","timestamp":1747368486823,"version":"3.40.5"},"reference-count":0,"publisher":"University of Florida George A Smathers Libraries","license":[{"start":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T00:00:00Z","timestamp":1747180800000},"content-version":"unspecified","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc\/4.0"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["FLAIRS"],"abstract":"<jats:p>Optimizing neural networks often encounters challenges such as saddle points, plateaus, and ill-conditioned curvature, limiting the effectiveness of standard optimizers like Adam, Nadam, and RMSProp. To address these limitations, we propose the Curvature-Adaptive Learning Rate (CALR) optimizer, a novel method that leverages local curvature estimates to dynamically adjust learning rates. CALR, along with its variants incorporating gradient clipping and cosine annealing schedules, offers enhanced robustness and faster convergence across diverse optimization tasks. Theoretical analysis confirms CALR\u2019s convergence properties, while empirical evaluations on benchmark functions\u2014Rosenbrock, Himmelblau, and Saddle Point\u2014highlight its efficiency in complex optimization landscapes. Furthermore, CALR demonstrates superior performance on neural network training tasks using MNIST and CIFAR-10 datasets, achieving faster convergence, lower loss, and better generalization compared to traditional optimizers. These results establish CALR as a promising optimization strategy for challenging neural network training problems.<\/jats:p>","DOI":"10.32473\/flairs.38.1.138986","type":"journal-article","created":{"date-parts":[[2025,5,15]],"date-time":"2025-05-15T15:21:05Z","timestamp":1747322465000},"source":"Crossref","is-referenced-by-count":0,"title":["Curvature-Adaptive Learning Rate Optimizer: Theoretical Insights and Empirical Evaluation on Neural Network Training"],"prefix":"10.32473","volume":"38","author":[{"given":"Kehelwala Dewage Gayan","family":"Maduranga","sequence":"first","affiliation":[]}],"member":"17357","published-online":{"date-parts":[[2025,5,14]]},"container-title":["The International FLAIRS Conference Proceedings"],"original-title":[],"link":[{"URL":"https:\/\/journals.flvc.org\/FLAIRS\/article\/download\/138986\/144070","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/journals.flvc.org\/FLAIRS\/article\/download\/138986\/144070","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,15]],"date-time":"2025-05-15T15:21:05Z","timestamp":1747322465000},"score":1,"resource":{"primary":{"URL":"https:\/\/journals.flvc.org\/FLAIRS\/article\/view\/138986"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,14]]},"references-count":0,"URL":"https:\/\/doi.org\/10.32473\/flairs.38.1.138986","relation":{},"ISSN":["2334-0762","2334-0754"],"issn-type":[{"value":"2334-0762","type":"electronic"},{"value":"2334-0754","type":"print"}],"subject":[],"published":{"date-parts":[[2025,5,14]]}}}