A Comparative Analysis of the Bayesian Regularization and Levenberg–Marquardt Training Algorithms in Neural Networks for Small Datasets: A Metrics Prediction of Neolithic Laminar Artefacts
<p>Technological measurements.</p> "> Figure 2
<p>Laminar artifacts from Nahal Reuel (<b>a</b>–<b>d</b>), Yiftahel (<b>e</b>,<b>f</b>), and Motza (<b>g</b>–<b>j</b>). Undamaged laminar artifacts (<b>a</b>,<b>c</b>,<b>e</b>,<b>g</b>,<b>i</b>) and laminar artifacts with missing distal ends (<b>b</b>,<b>d</b>,<b>f</b>,<b>h</b>,<b>j</b>).</p> "> Figure 3
<p>Example of a neural network with an architecture based on two hidden layers.</p> "> Figure 4
<p>Example of perceptron i in a hidden layer k R(s) activation function.</p> "> Figure 5
<p>Example of a sigmoid function.</p> "> Figure 6
<p>Best validation performance based on LR = 0.1. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The regression curve was carried out on training data.</p> "> Figure 7
<p>Error histogram during training validation and test processes based on LR = 0.1. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>).</p> "> Figure 7 Cont.
<p>Error histogram during training validation and test processes based on LR = 0.1. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>).</p> "> Figure 8
<p>Regression lines during training, validation, test, and merged processes based on LR = 0.1. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The analysis was carried out on 70% of the total data.</p> "> Figure 8 Cont.
<p>Regression lines during training, validation, test, and merged processes based on LR = 0.1. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The analysis was carried out on 70% of the total data.</p> "> Figure 9
<p>Regression lines during training prediction (<b>a</b>) and final test (<b>c</b>), both using a Bayesian regularization training algorithm based on LR = 0.1. Regression line during training prediction (<b>b</b>) and final test (<b>d</b>), both using a Levenberg–Marquardt regularization training algorithm based on LR = 0.1.</p> "> Figure 10
<p>Best validation performance based on LR = 0.3. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The regression curve was carried out on training data.</p> "> Figure 11
<p>Error histogram during training validation and test processes based on LR = 0.3. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>).</p> "> Figure 12
<p>Regression lines during training, validation, test, and merged processes based on LR = 0.3. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The analysis was carried out on 70% of the total data.</p> "> Figure 13
<p>Regression lines during training prediction (<b>a</b>) and final test (<b>c</b>), both using a Bayesian regularization training algorithm based on LR = 0.3. Regression lines during training prediction (<b>b</b>) and final test (<b>d</b>), both using a Levenberg–Marquardt regularization training algorithm based on LR = 0.3.</p> "> Figure 14
<p>Best validation performance based on LR = 0.4. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The regression line was carried out on training data.</p> "> Figure 15
<p>Error histogram during training validation and test processes based on LR = 0.4. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>).</p> "> Figure 16
<p>Regression lines during training, validation, test, and merged processes based on LR = 0.4. Bayesian regularization algorithm (<b>top</b>). Levenberg–Marquardt regularization algorithm (<b>bottom</b>). The analysis was carried out on 70% of the total data.</p> "> Figure 17
<p>Regression lines during training prediction (<b>a</b>) and final test (<b>c</b>), both using a Bayesian regularization training algorithm based on LR = 0.4. Regression lines during training prediction (<b>b</b>) and final test (<b>d</b>), both using a Levenberg–Marquardt regularization training algorithm based on LR = 0.4.</p> ">
Abstract
:1. Introduction
2. Material and Methods
- net.trainParam.epochs: This parameter specifies the number of epochs, which is the number of times the entire training set is presented to the neural network for training. A complete era is when the network has seen and learned from all the training data once.
- net.trainParam.lr: This parameter represents the learning rate (LR). It indicates how many times the neural network should update the weights based on the error during each iteration of the optimization algorithm. A higher learning rate may accelerate the training, but it may also cause oscillations or converging difficulties.
- net.trainParam.max_fail: This parameter represents the maximum number of consecutive failures allowed during training. If the network error stops decreasing for several epochs specified by this parameter, the training may stop before reaching the maximum number of epochs to avoid overfitting problems.
- net.trainParam.min_grad: This parameter specifies the minimum gradient threshold. During training, if the weight gradient of the neural network becomes lower than this value, the optimization algorithm may consider that the network has reached a convergence condition.
- net.trainParam.goal: This parameter is the training performance goal. It represents the average error that you want to achieve during training. Once the average neural network error reaches or approaches this value, the training ends.
- The learning rate (LR) should be initially set to small values and then adjusted as the training proceeds to avoid oscillations and divergence.
- The number of epochs (epochs) should be sufficient to allow the network to learn from the data, but not too large to avoid overfitting.
- max_fail can be adjusted according to the training stability and overfitting measurement.
- min_grad and goal are often set according to the desired accuracy and tolerance for convergence.
3. Results
3.1. Bayesian vs. Levenberg–Marquardt Based on LR = 0.1
3.2. Bayesian vs. Levenberg–Marquardt Based on LR = 0.3
3.3. Bayesian vs. Levenberg–Marquardt Based on LR = 0.4
4. Conclusions
Supplementary Materials
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Guyot, A.; Lennon, M.; Lorho, T.; Hubert-Moy, L. Combined Detection and Segmentation of Archeological Structures from LiDAR Data Using a Deep Learning Approach. J. Comput. Appl. Archaeol. 2021, 4, 1. [Google Scholar] [CrossRef]
- Scotland, A.; Øivind, T.; David, C.; Waldeland, U.A. Using deep neural networks on airborne laser scanning data: Results from a case study of semi-automatic mapping of archaeological topography. Archaeol. Prospect. 2018, 26, 165–175. [Google Scholar] [CrossRef]
- Caspari, G.; Crespo, B. Convolutional neural networks for archaeological site detection—Finding “princely” tombs. J. Archaeol. Sci. 2019, 110, 104998. [Google Scholar] [CrossRef]
- Davis, D.S.; Caspari, G.; Lipo, C.P.; Sanger, M.C. Deep Learning Reveals Extent of Archaic Native American Shell-Ring Building Practices. J. Archaeol. Sci. 2021, 132, 105433. [Google Scholar] [CrossRef]
- Küçükdemirci, M.; Sarris, A. Deep learning based automated analysis of archaeo-geophysical images. Archaeol. Prospect. 2020, 27, 107–118. [Google Scholar] [CrossRef]
- Trier, Ø.D.; Reksten, J.H.; Løseth, K. Automated mapping of cultural heritage in Norway from airborne lidar data using faster R-CNN. Int. J. Appl. Earth Obs. Geoinf. 2021, 95, 102241. [Google Scholar] [CrossRef]
- Cole, K.E.; Yaworsky, P.M.; Hart, I.A. Evaluating statistical models for establishing morphometric taxonomic identifications and a new approach using Random Forest. J. Archaeol. Sci. 2022, 143, 105610. [Google Scholar] [CrossRef]
- Eberl, M.; Bell, C.S.; Spencer-Smith, J.; Raj, M.; Sarubbi, A.; Johnson, P.S.; Rieth, A.E.; Chaudhry, U.; Aguila, R.E.; McBride, M. Machine Learning–Based Identification of Lithic Microdebitage. Adv. Archaeol. Pract. 2023, 11, 152–163. [Google Scholar] [CrossRef]
- Gualandi, M.L.; Gattiglia, G.; Anichini, F. An Open System for Collection and Automatic Recognition of Pottery through Neural Network Algorithms. Heritage 2021, 4, 140–159. [Google Scholar] [CrossRef]
- Saipraneeth, G.; Jyotindra, N.; Roger, A.S.; Sachin, S.G. A Bayesian regularization-backpropagation neural network model for peeling computations. J. Adhes. 2021, 99, 92–115. [Google Scholar] [CrossRef]
- Barber, D. Bayesian Reasoning and Machine Learning; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
- Sapna, S.; Tamilarasi, A.; Pravin, M.K. Backpropagation Learning Algorithm Based on Levenberg Marquardt Algorithm. Comput. Sci. Inf. Technol. 2012, 2, 393–398. [Google Scholar] [CrossRef]
- Qian, N. On the momentum term in gradient descent learning algorithms. Neural Netw. 1999, 12, 145–151. [Google Scholar] [CrossRef] [PubMed]
- Arimura, M. The Neolithic Lithic Industry at Tell Ain El-Kerkh; Archaeopress Archaeology: Oxford, UK, 2020. [Google Scholar]
- Shea, J.J. Stone Tools in the Palaeolithic and Neolithic Near East: A Guide; Cambridge University Press: New York, NY, USA, 2013. [Google Scholar]
- Barzilai, O. Social Complexity in the Southern Levantine PPNB as Reflected through Lithic Studies: The Bidirectional Blade Industries; BAR International Series; Archaeopress: Oxford, UK, 2010. [Google Scholar]
- Nobile, E.; Conati, C.B. The Standardisation of the PPNB Lithic Industry from Er-Rahib. Orig. Rev. Prehistory Protohistory Anc. Civiliz. 2022, 46, 7–28. [Google Scholar]
- Sariev, E.; Germano, G. Bayesian regularized artificial neural networks for the estimation of the probability of default. Quant. Financ. 2020, 20, 311–328. [Google Scholar] [CrossRef]
- Reynaldi, A.; Lukas, S.; Margaretha, H. Backpropagation and Levenberg-Marquardt Algorithm for Training Finite Element Neural Network. In Proceedings of the 2012 Sixth UKSim/AMSS European Symposium on Computer Modeling and Simulation, Valletta, Malta, 14–16 November 2012. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Troiano, M.; Nobile, E.; Mangini, F.; Mastrogiuseppe, M.; Conati Barbaro, C.; Frezza, F. A Comparative Analysis of the Bayesian Regularization and Levenberg–Marquardt Training Algorithms in Neural Networks for Small Datasets: A Metrics Prediction of Neolithic Laminar Artefacts. Information 2024, 15, 270. https://doi.org/10.3390/info15050270
Troiano M, Nobile E, Mangini F, Mastrogiuseppe M, Conati Barbaro C, Frezza F. A Comparative Analysis of the Bayesian Regularization and Levenberg–Marquardt Training Algorithms in Neural Networks for Small Datasets: A Metrics Prediction of Neolithic Laminar Artefacts. Information. 2024; 15(5):270. https://doi.org/10.3390/info15050270
Chicago/Turabian StyleTroiano, Maurizio, Eugenio Nobile, Fabio Mangini, Marco Mastrogiuseppe, Cecilia Conati Barbaro, and Fabrizio Frezza. 2024. "A Comparative Analysis of the Bayesian Regularization and Levenberg–Marquardt Training Algorithms in Neural Networks for Small Datasets: A Metrics Prediction of Neolithic Laminar Artefacts" Information 15, no. 5: 270. https://doi.org/10.3390/info15050270
APA StyleTroiano, M., Nobile, E., Mangini, F., Mastrogiuseppe, M., Conati Barbaro, C., & Frezza, F. (2024). A Comparative Analysis of the Bayesian Regularization and Levenberg–Marquardt Training Algorithms in Neural Networks for Small Datasets: A Metrics Prediction of Neolithic Laminar Artefacts. Information, 15(5), 270. https://doi.org/10.3390/info15050270