000 01648nam a22001697a 4500
082 _a006.31
_bM476D
100 _aMazumder, Sudipta
_945441
245 _aDesigning Optimizer for Deep Learning
_cby Sudipta Mazumder
260 _aIIT Jodhpur
_bDepartment of Computer Science and Technology
_c2023
300 _avii,10p.
_bHB
500 _aOptimizers form the backbone of any convolutional network as they are responsible for making the functions converge faster. The optimizers do this by modifying the weights and learning rate of the algorithm, which reduces the loss and improves the accuracy. A lot of optimizers have gained traction over the years, out of which SGD and Adam take the cake. Adam has taken the lead as it helps to reduce the dying gradient problem of SGD. However, we still have scope for improvement. With this paper, we aim to introduce a new algorithm that surpasses the performance of Adam by calculating the angular gradients (cosine and tangent angles) at consecutive steps. This algorithm uses the gradient of the current step, the previous step, and the step previous to that. As we present more information to the optimizer’s algorithm, the algorithm has more information, making it better poised to make more accurate predictions at faster convergence rates. We have tested this approach on benchmark datasets and compared it with other state-of-the-art optimizers, and have obtained superior results in almost every approach.
650 _aDepartment of Computer Science and Technology
_945442
650 _aOptimizers
_945443
650 _aMTech Theses
_945444
700 _aPaul, Angshuman
_945445
942 _cTH
999 _c16588
_d16588