Finding the Optimal Topology of an Approximating Neural Network
Mathematics
https://doi.org/10.3390/MATH1101021Abstract
A large number of researchers spend a lot of time searching for the most efficient neural network to solve a given problem. The procedure of configuration, training, testing, and comparison for expected performance is applied to each experimental neural network. The configuration parameters—training methods, transfer functions, number of hidden layers, number of neurons, number of epochs, and tolerable error—have multiple possible values. Setting guidelines for appropriate parameter values would shorten the time required to create an efficient neural network, facilitate researchers, and provide a tool to improve the performance of automated neural network search methods. The task considered in this paper is related to the determination of upper bounds for the number of hidden layers and the number of neurons in them for approximating artificial neural networks trained with algorithms using the Jacobi matrix in the error function. The derived formulas for the upper limits of the numb...
References (45)
- Rozycki, P.; Kolbusz, J.; Wilamowski, B.M. Dedicated deep neural network architectures and methods for their training. In Proceedings of the IEEE 19th International Conference on Intelligent Engineering Systems (INES), Bratislava, Slovakia, 3-5 September 2015.
- Torres, J.J.; Munoz, M.A.; Marro, J.; Garrido, P.L. Influence of topology on the performance of a neural network. Neurocomputing 2004, 58-60, 229-234.
- Emmert-Streib, F. Influence of the neural network topology on the learning dynamics. Neurocomputing 2006, 69, 1179-1182.
- Lv, C.; Xing, Y.; Zhang, J.; Na, X.; Li, Y.; Liu, T.; Cao, D.; Wang, F.-Y. Levenberg-Marquardt Backpropagation Training of Mul- tilayer Neural Networks for State Estimation of a Safety-Critical Cyber-Physical System. IEEE Trans. Ind. Inform. 2018, 14, 3436- 3446.
- Sapna, S.; Tamilarasi, A.; Kumar, M.; Backpropagation learning algorithm based on Levenberg Marquardt algorithm. Comput. Sci. Inf. Technol. 2012, 7, 393-398.
- Botev, A.; Ritter, H.; Barber, D. Practical Gauss-Newton Optimisation for Deep Learning. In Proceedings of the 34-th Interna- tional Conference on Machine Learning PMLR 70, Sydney, Australia, 6-11 August 2017; pp. 557-565.
- Gratton, S.; Lawless, A.; Nichols, N. Approximate Gauss-Newton Methods for Nonlinear Least Squares Problems. SIAM J. Optim. 2017, 18, 106-132.
- Okut, H. Bayesian Regularized Neural Networks for Small n Big p Data. In Artificial Neural Networks-Models and Applications; IntechOpen: London, UK, 2016.
- Gouravaraju, S.; Narayan, J.; Sauer, R.; Gautam, S. A Bayesian regularization-backpropagation neural network model for peel- ing computations, J. Adhes. 2023, 99, 92-115.
- Chel, H.; Majumder, A.; Nandi, D. Scaled Conjugate Gradient Algorithm in Neural Network Based Approach for Handwritten Text Recognition. Commun. Comput. Inf. Sci. 2011, 204, 196-210.
- Babani, L.; Jadhav, S.; Chaudhari, B. Scaled Conjugate Gradient Based Adaptive ANN Control for SVM-DTC Induction Motor Drive. In Artificial Intelligence Applications and Innovations. AIAI 2016; Iliadis, L., Maglogiannis, I., Eds; Springer: Cham, Switzer- land, 2016; Volume 475, pp. 384-395.
- Goldfarb, D.; Ren, Y.; Bahamou, A. Practical Quasi-Newton Methods for Training Deep Neural Networks. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, Canada, 6-12 December 2020.
- Wills, A.; Schön, T.; Jidling, C. A fast quasi-Newton-type method for large-scale stochastic optimization. IFAC-Papers 2020, 53, 1249-1254.
- Hunter, D.; Yu, H.; Pukish III, M.S.; Kolbusz, J.; Wilamowski, B.M. Selection of proper neural network sizes and architectures: A comparative study. IEEE Trans. Ind. Inform. 2012, 8, 228-240.
- Kuri, A. The Best Neural Network Architecture. In Proceedings of the Mexican International Congress on Artificila Intelligence, Monterrey, NL, Mexico, 24-29 October 2014.
- Khalil, K.; Eldash, O.; Kumar, A.; Bayoumi, M. An Efficient Approach for Neural Network Architecture. In proceedings of the 25th IEEE International Conference on Electronics, Circuits and Systems (ICECS), Bordeaux, France, 9-12 December 2018.
- Jinchuan, K.; Xinzhe, L. Empirical analysis of optimal hidden neurons in neural network modeling for stock prediction. In Proceedings of the Pacific-Asia Workshop on Computational Intelligence and Industrial Application, Wuhan, China, 19-20 December 2008; Volume 2, pp. 828-832.
- Alvarez, J.M.; Salzmann, M. Learning the Number of Neurons in Deep Networks. Adv. Neural Inf. Process. Syst. 2016, 29, 1-9.
- Li, J.Y.; Chow, T.W.S.; Yu, Y.L. Estimation theory and optimization algorithm for the number of hidden units in the higher- order feedforward neural network. In Proceedings of the IEEE International Conference on Neural Networks, Perth, WA, Australia, 27 November-1 December 1995; pp. 1229-1233.
- Tamura, S.; Tateishi, M. Capabilities of a four-layered feedforward neural network: Four layers versus three, In Proceedings of the IEEE Transactions on Neural Networks, March 1997, 8, 251-255.
- Sheela, K.G.; Deepa, S.N. Review on Methods to Fix Number of Hidden Neurons in Neural Networks. Math. Probl. Eng. 2013, 2013, 425740.
- Madhiarasan, M.; Deepa, S.N. Comparative analysis on hidden neurons estimation in multi layer perceptron neural networks for wind speed forecasting. Artif. Intell. Rev. 2016, 48, 449-471.
- Xu, S.; Chen, L. A novel approach for determining the optimal number of hidden layer neurons for FNN's and its application in data mining. In Proceedings of the 5th International Conference on Information Technology and Applications, 23-26 July 2008; pp. 683-686.
- Shibata, K.; Ikeda, Y. Effect of number of hidden neurons on learning in large-scale layered neural networks. In Proceedings of the ICROS-SICE International Joint Conference, Fukuoka International Congress Center, Japan, Cairns, Queensland, Australia, 18-21 August 2009; pp. 5008-5013.
- Yotov, K.; Hadzhikolev, E.; Hazdhikoleva, S. Determining the Number of Neurons in Artificial Neural Networks for Approximation, Trained with Algorithms Using the Jacobi Matrix. TEM J. 2020, 9, 1320-1329.
- Shen, H.-Y.; Wang, Z.; Gao, C.-Y. Determining the number of BP neural network hidden layer units. J. Tianjin Univ. Technol. 2008, 24, 13-15.
- Ibnu Choldun, R.M.; Santoso, J.; Surendro, K. Determining the number of hidden layers in neural network by using principal component analysis. Adv. Intell. Syst. Comput. 2020, 1038, 490-500.
- Stathakis, D. How many hidden layers and nodes? Int. J. Remote Sens. 2009, 30, 2133-2147.
- Hanay, Y.S.; Arakawa, S.; Murata, M. Network topology selection with multistate neural memories. Expert Syst. Appl. 2015, 42, 3219-3226.
- Perzina, R. Self-learning Genetic Algorithm for Neural Network Topology Optimization. Smart Innov. Syst. Technol. 2015, 38, 179-188. 10.1007/978-3-319-19728-9_15.
- Vizitiu, I.-C.; Popescu, F. GANN system to optimize both topology and neural weights of a feedforward neural network. In Proceedings of the 8th International Conference on Communications, Sintok, Kedah, Malaysia, 1-3 October, 2022.
- White, D.; Ligomenides, P. GANNet: A genetic algorithm for optimizing topology and weights in neural network design. Lect. Notes Comput. Sci. 1995, 686, 322-327.
- Arena, P.; Caponetto, R.; Fortuna, L.; Xibilia, M.G. (n.d.). Genetic algorithms to select optimal neural network topology. In Proceedings of the 35th Midwest Symposium on Circuits and Systems, Washington, DC, USA, 9-12 August 1992.
- Leon, F. Optimizing neural network topology using Shapley value. In Proceedings of the 18th International Conference on System Theory, Control and Computing (ICSTCC), Sinaia, Romania, 17-19 October 2014.
- Kuroda, K.; Hasegawa, M. Method for Estimating Neural Network Topology Based on SPIKE-Distance. Lect. Notes Comput. Sci. 2016, 9886, 91-98.
- Curteanu, S.; Leon, F.; Furtuna, R.; Dragoi, E.N.; Curteanu, N. Comparison between different methods for developing neural network topology applied to a complex polymerization process. In Proceedings of the 2010 International Joint Conference on Neural Networks (IJCNN). Barcelona, Spain, 18-23 July 2010.
- Guliyev, N.J.; Ismailov, V.E. On the approximation by single hidden layer feedforward neural networks with fixed weights. Neural Netw. 2018, 98, 296-304.
- Thomas, A.J.; Petridis, M.; Walters, S.D.; Gheytassi, S.M.; Morgan, R.E. Two Hidden Layers are Usually Better than One. Com- mun. Comput. Inf. Sci. 2017, 744, 279-290.
- Nakama, T. Comparisons of Single-and Multiple-Hidden-Layer Neural Networks. Lect. Notes Comput. Sci. 2011, 6675, 270-279.
- Ibnu, C.R.M.; Santoso, J.; Surendro, K. Determining the Neural Network Topology. In Proceedings of the 8th International Conference on Software and Computer Applications-ICSCA. 2019, 19, 357-362.
- MatLab. Available online: https://www.mathworks.com/products/matlab.html (accessed on 8 October 2022).
- Börlin, N. Nonlinear Optimization Least Squares Problems-The Gauss-Newton method. Available online: https://www8.cs.umu.se/kurser/5DA001/HT07/lectures/lsq-handouts (accessed on 30 December 2022).
- Cartis, C. Mathematical Institute, University of Oxford. Linear and nonlinear least-squares problems; the Gauss-Newton method. Available online: https://courses-archive.maths.ox.ac.uk/node/view_material/4898. (accessed on 8 October 2022).
- Madsen, K.; Nielsen, H.B.; Tingleff, O. Methods for Non-Linear Least Squares Problems, 2004. Available online: http://www2.imm.dtu.dk/pubdb/edoc/imm3215.pdf. (accessed on 8 October 2022).
- Disclaimer/Publisher's Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.