Conformable fractional calculus will be a promising area of research for information processing as natural language and material modelling due to its ease of implementation. In this paper, we propose a fractional gradient descent method for the backpropagation training of neural networks. In particular, the conformable fractional calculus is employed to evaluate the fractional differential gradient function instead of the classical differential gradient function. The results obtained on a large dataset with this approach provide a new optimized, faster and simpler implemented algorithm than the conventional one.
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
Tel.: +1 703 830 6300
Fax: +1 703 830 2300 firstname.lastname@example.org
(Corporate matters and books only) IOS Press c/o Accucoms US, Inc.
For North America Sales and Customer Service
West Point Commons
Lansdale PA 19446
Tel.: +1 866 855 8967
Fax: +1 215 660 5042 email@example.com