Improving Image Recognition Performance through an Improved Dual Convolutional Neural Network with Recurrent Integration and Residual Model
Keywords:
Convolutional Neural Network, Dual-optimized CNN, Recurrent Neural Network, ShortCut3-ResNet moduleAbstract
This observe proposes a new Convolutional Neural Network (CNN) framework for image category, the usage of both CNN and Recurrent Neural Networks (RNNs) for superior characteristic mastering. The technique combines RNNs in CNNs to offer neighborhood and temporal correlations can be extracted. Additionally, a new "ShortCut3-ResNet" module, triggered by the ultimate ResNet connections, facilitates easy float of data over layers. Moreover, the twin optimization model optimizes cooperatively on the convolutional and fully related tiers. This correction is evaluated the use of the CIFAR-10 statistics set. Experiments display the efficiency of the proposed method, achieving better overall performance in comparison to existing methods in phrases of accuracy and sample length The study also investigates the effect of activation characteristic, sampling strategies, pooling strategies, and dual optimization so, and provides precious insights for optimizing CNN overall performance.
Downloads
References
D. L. Hansen, B. Shneiderman, and M. A. Smith, Analyzing Social Media Networks with NodeXL: Insights from a Connected World. 2019. doi: 10.1016/B978-0-12-817756-3.09988-X.
M. Alkhawlani, M. Elmogy, and H. El Bakry, “Text-based, Content-based, and Semantic-based Image Retrievals: A Survey,” International conference on Machine Vision, Image processing and Pattern Analysis, vol. 04, no. 01, 2015.
L. Alzubaidi et al., “Review of deep learning: concepts, CNN architectures, challenges, applications, future directions,” J Big Data, vol. 8, no. 1, 2021, doi: 10.1186/s40537-021-00444-8.
J. Ma, X. Jiang, A. Fan, J. Jiang, and J. Yan, “Image Matching from Handcrafted to Deep Features: A Survey,” Int J Comput Vis, vol. 129, no. 1, 2021, doi: 10.1007/s11263-020-01359-2.
J. Gu et al., “Recent advances in convolutional neural networks,” Pattern Recognit, vol. 77, 2018, doi: 10.1016/j.patcog.2017.10.013.
K. Choudhary et al., “Recent advances and applications of deep learning methods in materials science,” npj Computational Materials, vol. 8, no. 1. 2022. doi: 10.1038/s41524-022-00734-6.
Persello et al., “Deep Learning and Earth Observation to Support the Sustainable Development Goals: Current approaches, open challenges, and future opportunities,” IEEE Geosci Remote Sens Mag, vol. 10, no. 2, 2022, doi: 10.1109/MGRS.2021.3136100.
Tsantekidis, N. Passalis, and A. Tefas, “Recurrent neural networks,” in Deep Learning for Robot Perception and Cognition, 2022. doi: 10.1016/B978-0-32-385787-1.00010-5.
“Integration of Natural Language Processing and Augmented Reality: ChatGPT Meets Apple Vision Pro,” International Research Journal of Modernization in Engineering Technology and Science, 2023, doi: 10.56726/irjmets43783.
H. Sadr, M. M. Pedram, and M. Teshnehlab, “A Robust Sentiment Analysis Method Based on Sequential Combination of Convolutional and Recursive Neural Networks,” Neural Process Lett, vol. 50, no. 3, 2019, doi: 10.1007/s11063-019-10049-1.
P. Matlani and M. Shrivastava, “Hybrid deep VGG-net convolutional classifier for video smoke detection,” CMES - Computer Modeling in Engineering and Sciences, vol. 119, no. 3, 2019, doi: 10.32604/cmes.2019.04985.
K. Kassylkassova, Z. Yessengaliyeva, G. Urazboev, and A. Kassylkassova, “OPTIMIZATION METHOD FOR INTEGRATION OF CONVOLUTIONAL AND RECURRENT NEURAL NETWORK,” Eurasian Journal of Mathematical and Computer Applications, vol. 11, no. 2, 2023, doi: 10.32523/2306-6172-2023-11-2-40-56.
Khan, A. Sohail, U. Zahoora, and A. S. Qureshi, “A survey of the recent architectures of deep convolutional neural networks,” Artif Intell Rev, vol. 53, no. 8, 2020, doi: 10.1007/s10462-020-09825-6.
Agrawal and N. Mittal, “Using CNN for facial expression recognition: a study of the effects of kernel size and number of filters on accuracy,” Visual Computer, vol. 36, no. 2, 2020, doi: 10.1007/s00371-019-01630-9.
Z. Chen, K. Gryllias, and W. Li, “Mechanical fault diagnosis using Convolutional Neural Networks and Extreme Learning Machine,” Mech Syst Signal Process, vol. 133, 2019, doi: 10.1016/j.ymssp.2019.106272.
S. H. S. Basha, S. R. Dubey, V. Pulabaigari, and S. Mukherjee, “Impact of fully connected layers on performance of convolutional neural networks for image classification,” Neurocomputing, vol. 378, 2020, doi: 10.1016/j.neucom.2019.10.008.
McNeely-White, J. R. Beveridge, and B. A. Draper, “Inception and ResNet features are (almost) equivalent,” Cogn Syst Res, vol. 59, 2020, doi: 10.1016/j.cogsys.2019.10.004.
Y. Zhang, Y. Qu, C. Li, Y. Lei, and J. Fan, “Ontology-driven hierarchical sparse coding for large-scale image classification,” Neurocomputing, vol. 360, 2019, doi: 10.1016/j.neucom.2019.05.059.
L. Carroll et al., “Multi-locus DNA metabarcoding of zooplankton communities and scat reveal trophic interactions of a generalist predator,” Sci Rep, vol. 9, no. 1, 2019, doi: 10.1038/s41598-018-36478-x.
Hu, M. Zhou, P. Yan, K. Bian, and R. Dai, “PCanet: A common solution for laser-induced fluorescence spectral classification,” IEEE Access, vol. 7, 2019, doi: 10.1109/ACCESS.2019.2933453.
Downloads
Published
How to Cite
Issue
Section
License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.