Academia.eduAcademia.edu

Outline

Differential Neural Networks (DNN)

IEEE Access

https://doi.org/10.1109/ACCESS.2020.3019307

Abstract

In this work, we propose an artificial neural network topology to estimate the derivative of a function. This topology is called a differential neural network because it allows the estimation of the derivative of any of the network outputs with respect to any of its inputs. The main advantage of a differential neural network is that it uses some of the weights of a multilayer neural network. Therefore, a differential neural network does not need to be trained. First, a multilayer neural network is trained to find the best set of weights that minimize an error function. Second, the weights of the trained network and its neuron activations are used to build a differential neural network. Consequently, a multilayer artificial neural can produce a specific output, and simultaneously, estimate the derivative of any of its outputs with respect to any of its inputs. Several computer simulations were carried out to validate the performance of the proposed method. The computer simulation results showed that differential neural networks are capable of estimating with good accuracy the derivative of a function. The method was developed for an artificial neural network with two layers; however, the method can be extended to more than two layers. Similarly, the analysis in this study is presented for two common activation functions. Nonetheless, other activation functions can be used as long as the derivative of the activation function can be computed. INDEX TERMS Differential neural network, artificial intelligence, neural network structure, derivative estimation, multilayer network.

References (34)

  1. S. Marsland, Machine Learning: An Algorithmic Perspective (Chapman & Hall/CRC Machine Learning & Pattern Recognition), 2nd ed. Boca Raton, FL, USA: CRC Press, 2015, pp. 190-210.
  2. M. I. Jordan and T. M. Mitchell, ''Machine learning: Trends, perspectives, and prospects,'' Science, vol. 349, no. 6245, pp. 255-260, Jul. 2015.
  3. C. Yin, Y. Zhu, J. Fei, and X. He, ''A deep learning approach for intru- sion detection using recurrent neural networks,'' IEEE Access, vol. 5, pp. 21954-21961, 2017.
  4. W. Liu, M. Zhang, Z. Luo, and Y. Cai, ''An ensemble deep learning method for vehicle type classification on visual traffic surveillance sensors,'' IEEE Access, vol. 5, pp. 24417-24425, 2017.
  5. B. Jin, Z. Jing, and H. Zhao, ''Incremental and decremental extreme learning machine based on generalized inverse,'' IEEE Access, vol. 5, pp. 20852-20865, 2017.
  6. S. Ledesma, M.-A. Ibarra-Manzano, E. Cabal-Yepez,
  7. D.-L. Almanza-Ojeda, and J.-G. Avina-Cervantes, ''Analysis of data sets with learning conflicts for machine learning,'' IEEE Access, vol. 6, pp. 45062-45070, 2018.
  8. R. J. Martin, S. L. Swapna, and S. Sujatha, ''Adopting machine learning models for data analytics-A technical note,'' Int. J. Comput. Sci. Eng., vol. 6, no. 10, pp. 359-364, Oct. 2018.
  9. Y. LeCun, Y. Bengio, and G. Hinton, ''Deep learning,'' Nature, vol. 521, pp. 436-444, May 2015.
  10. V. Mnih, K. Kavukcuoglu, and D. Silver, ''Human-level control through deep reinforcement learning,'' Nature, vol. 518, pp. 529-533, 2015.
  11. Liu, Z. Wang, X. Liu, N. Zeng, Y. Liu, and F. E. Alsaadi, ''A survey of deep neural network architectures and their applications,'' Neurocomput- ing, vol. 234, pp. 11-26, Apr. 2017.
  12. K. H. Jin, M. T. McCann, E. Froustey, and M. Unser, ''Deep convolutional neural network for inverse problems in imaging,'' IEEE Trans. Image Process., vol. 26, no. 9, pp. 4509-4522, Sep. 2017.
  13. M. Umar, F. Amin, H. A. Wahab, and D. Baleanu, ''Unsupervised con- strained neural network modeling of boundary value corneal model for eye surgery,'' Appl. Soft Comput., vol. 85, Dec. 2019, Art. no. 105826.
  14. M. A. Z. Raja, J. A. Khan, and I. M. Qureshi, ''A new stochastic approach for solution of Riccati differential equation of fractional order,'' Ann. Math. Artif. Intell., vol. 60, nos. 3-4, pp. 229-250, Dec. 2010.
  15. S. Mall and S. Chakraverty, ''Application of Legendre neural network for solving ordinary differential equations,'' Appl. Soft Comput., vol. 43, pp. 347-356, Jun. 2016.
  16. M. A. Z. Raja, M. A. Manzar, and R. Samar, ''An efficient computational intelligence approach for solving fractional order Riccati equations using ANN and SQP,'' Appl. Math. Model., vol. 39, nos. 10-11, pp. 3075-3093, Jun. 2015.
  17. A. Jafarian, S. Measoomy Nia, A. K. Golmankhaneh, and D. Baleanu, ''On artificial neural networks approach with new cost functions,'' Appl. Math. Comput., vol. 339, pp. 546-555, Dec. 2018.
  18. A. J. Meade and A. A. Fernandez, ''The numerical solution of linear ordinary differential equations by feedforward neural networks,'' Math. Comput. Model., vol. 19, no. 12, pp. 1-25, Jun. 1994.
  19. I. E. Lagaris, A. Likas, and D. I. Fotiadis, ''Artificial neural networks for solving ordinary and partial differential equations,'' IEEE Trans. Neural Netw., vol. 9, no. 5, pp. 987-1000, Sep. 1998.
  20. S. Mall and S. Chakraverty, ''Comparison of artificial neural network architecture in solving ordinary differential equations,'' Adv. Artif. Neural Syst., vol. 2013, pp. 1-12, Jan. 2013.
  21. M. Pakdaman, A. Ahmadian, S. Effati, S. Salahshour, and D. Baleanu, ''Solving differential equations of fractional order using an optimization technique based on training artificial neural network,'' Appl. Math. Com- put., vol. 293, pp. 81-95, Jan. 2017.
  22. F. Rostami and A. Jafarian, ''A new artificial neural network structure for solving high-order linear fractional differential equations,'' Int. J. Comput. Math., vol. 95, no. 3, pp. 528-539, Mar. 2018.
  23. S. J. Russell and P. Norvig, Artificial Intelligence: A Modern Approach, 4th ed. London, U.K.: Pearson, 2020, pp. 122-204.
  24. T. Masters, Deep Belief Nets in C++ and CUDA C: Restricted Boltzmann Machines and Supervised Feedforward Networks, vol. 1. Ithaca, NY, USA: Apress, 2017, pp. 11-25.
  25. T. Masters, Deep Belief Nets in C++ and CUDA C: Convolutional Nets, vol. 3. Ithaca, NY, USA: Apress, 2018, pp. 28-37.
  26. J. D. Kelleher, B. M. Namee, and A. D'Arcy, Fundamentals of Machine Learning for Predictive Data Analytics: Algorithms, Worked Examples, and Case Studies, 2nd ed. Cambridge, MA, USA: MIT Press, 2020.
  27. C. C. Aggarwal, Neural Networks and Deep Learning: A Textbook, 1st ed. Cham, Switzerland: Springer, 2018, pp. 53-103.
  28. I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning (Adaptive Computation and Machine Learning Series). Cambridge, MA, USA: MIT Press, 2016, pp. 267-309.
  29. S. Ledesma, M.-A. Ibarra-Manzano, M.-G. Garcia-Hernandez, and D.-L. Almanza-Ojeda, ''Neural lab a simulator for artificial neural net- works,'' in Proc. IEEE Comput. Conf., Jul. 2017, pp. 716-721.
  30. W. H. Press, S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery, Numer- ical Recipes in C++: The Art of Scientific Computing, 3rd ed. Cambridge, U.K.: Cambridge Univ. Press, 2007, pp. 549-555.
  31. S. Ledesma, M. Torres, D. Hernandez, G. Avina, and G. Garcia, ''Tem- perature cycling on simulated annealing for neural network learning,'' in Proc. MICAI. Berlin, Germany: Springer, 2007, pp. 161-171.
  32. S. Ledesma, G. Avina, and R. Sanchez, Practical Considerations for Sim- ulated Annealing Implementation (Simulated Annealing Book). London, U.K.: IntechOpen, 2008, pp. 401-420.
  33. K. L. Du and M. N. S. Swamy, Neural Networks and Statistical Learning, 2nd ed. London, U.K.: Springer, 2019, pp. 143-164.
  34. SERGIO LEDESMA received the M.S. degree from the University of Guanajuato, Mexico, while working on the setup of Internet, and the Ph.D. degree from the Stevens Institute of Technol- ogy, Hoboken, NJ, in 2001. After graduating, he worked for Barclays Bank as part of the IT- HR Group. He has worked as a Software Engineer for several years, and he is also the Creator of the Software Neural Laboratory, Wintempla, and TexLab. He is currently a Research Professor with the University of Guanajuato. He is on a sabbatical stay with the University of Ottawa, Canada. His research interests include artificial intelligence and software engineering. DORA-LUZ ALMANZA-OJEDA received the B.S. degree in electronics engineering and the M.S. degree in electrical engineering from the University of Guanajuato, Salamanca, Mexico, in 2003 and 2005, respectively, and the Ph.D. degree from Paul Sabatier University, Toulouse, France, in 2011. She is currently an Assistant Pro- fessor with the Electronics Engineering Depart- ment, University of Guanajuato. Her research interests include embedded vision for controlling autonomous robots and real time systems. MARIO-ALBERTO IBARRA-MANZANO (Member, IEEE) received the B.Eng. degree in communication and electronic engineering and the M.Eng. degree in electric from the University of Guanajuato, Salamanca, Mexico, in 2003 and 2006, respectively, and the Ph.D. degree (Hons.) from the Institut National des Sciences Appliquées, Toulouse, France, in 2011. He is currently an Assistant Professor with the Electronics Engineering Department, Universidad de Guanajuato. His research interests include digital design on FPGA for image processing applied on autonomous robots and real time systems.