Multimodal Deep Learning Architecture to Evaluate Emotion Recognition in Tea Packing

Authors

  • Xue Yang Universiti Utara Malaysia, changloon, 06010, Malaysia and Jingdezhen Ceramic University, Jingdezhen, Jiangxi, 333000, China
  • Adzrool Idzwan Bin Ismail Universiti Utara Malaysia, changloon, 06010, Malaysia

Keywords:

Multi Modal, LSTM, Deep Learning, Emotion Recognition, Virtual Reality, Tea Packing

Abstract

The packaging of consumer goods, including tea, is increasingly evolving beyond its conventional role as a mere container. Virtual reality (VR) has found innovative applications in the field of packaging, transforming the way products are designed, marketed, and experienced. This research paper proposed a novel tea packaging by integrating virtual reality (VR), emotion recognition technology, and a novel Multimodal Fusion Deep LSTM (MFD-LSTM) model, creating a dynamic and interactive tea packaging experience that engages all the senses. The core contribution of this study revolves around the fusion of VR technology, emotion recognition, and the MFD-LSTM model. This synergy enables tea packaging to become a dynamic medium for conveying brand narratives and invoking emotional responses in consumers. The MFD-LSTM model, capable of processing multiple sensory inputs simultaneously, offers real-time recognition of consumer emotions, which, in turn, influences the unfolding VR experience. It is his research advocates for the widespread adoption of the interactive tea packaging experience model, which harnesses VR, emotion recognition, and the MFD-LSTM model to create a multisensory and emotionally resonant connection between tea brands and consumers. The proposed MFD-LSTM model effectively evaluates the emotions and increases the performance towards packing. Through analysis, it is concluded that the proposed MFD-LSTM model is effective in the packing scenario.

Downloads

Download data is not yet available.

References

Ezzameli, K., & Mahersia, H. (2023). Emotion recognition from unimodal to multimodal analysis: A review. Information Fusion, 101847.

Zhu, L., Pergola, G., Gui, L., Zhou, D., & He, Y. (2021). Topic-driven and knowledge-aware transformer for dialogue emotion detection. arXiv preprint arXiv:2106.01071.

Abdelkawy, H. K. M. (2021). Hybrid approaches for context recognition in Ambient Assisted Living systems: application to emotion recognition and human activity recognition and anticipation (Doctoral dissertation, Université Paris-Est Créteil Val-de-Marne-Paris 12).

Ahmad, Z., Rabbani, S., Zafar, M. R., Ishaque, S., Krishnan, S., & Khan, N. (2023). Multi-level stress assessment from ecg in a virtual reality environment using multimodal fusion. IEEE Sensors Journal.

Xia, B., Sakamoto, H., Wang, X., & Yamasaki, T. (2022). Packaging design analysis by predicting user preference and semantic attribute. ITE Transactions on Media Technology and Applications, 10(3), 120-129.

Ryumin, D., Kagirov, I., Axyonov, A., Pavlyuk, N., Saveliev, A., Kipyatkova, I., ... & Karpov, A. (2020). A multimodal user interface for an assistive robotic shopping cart. Electronics, 9(12), 2093.

Zhu, D., & Liu, G. (2022). Deep neural network model-assisted reconstruction and optimization of chinese characters in product packaging graphic patterns and visual styling design. Scientific Programming, 2022.

Noorian, S. S., Psyllidis, A., & Bozzon, A. (2019). ST-Sem: A Multimodal Method for Points-of-Interest Classification Using Street-Level Imagery. In Web Engineering: 19th International Conference, ICWE 2019, Daejeon, South Korea, June 11–14, 2019, Proceedings 19 (pp. 32-46). Springer International Publishing.

Liu, C., Huang, H., & Yang, P. (2023). Multi-task learning from multimodal single-cell omics with Matilda. Nucleic Acids Research, 51(8), e45-e45.

Sundaram, S., Kellnhofer, P., Li, Y., Zhu, J. Y., Torralba, A., & Matusik, W. (2019). Learning the signatures of the human grasp using a scalable tactile glove. Nature, 569(7758), 698-702.

Yang, M., Wu, C., Guo, Y., Jiang, R., Zhou, F., Zhang, J., & Yang, Z. (2023). Transformer-based deep learning model and video dataset for unsafe action identification in construction projects. Automation in Construction, 146, 104703.

Murshed, M. S., Murphy, C., Hou, D., Khan, N., Ananthanarayanan, G., & Hussain, F. (2021). Machine learning at the network edge: A survey. ACM Computing Surveys (CSUR), 54(8), 1-37.

Kumar, A., & Garg, G. (2019). Empirical study of shallow and deep learning models for sarcasm detection using context in benchmark datasets. Journal of ambient intelligence and humanized computing, 1-16.

Shoeibi, A., Khodatars, M., Ghassemi, N., Jafari, M., Moridian, P., Alizadehsani, R., ... & Acharya, U. R. (2021). Epileptic seizures detection using deep learning techniques: A review. International Journal of Environmental Research and Public Health, 18(11), 5780.

Ma, P., Zhang, Z., Jia, X., Peng, X., Zhang, Z., Tarwa, K., ... & Wang, Q. (2022). Neural network in food analytics. Critical Reviews in Food Science and Nutrition, 1-19.

Magassouba, A., Sugiura, K., & Kawai, H. (2020). A multimodal target-source classifier with attention branches to understand ambiguous instructions for fetching daily objects. IEEE Robotics and Automation Letters, 5(2), 532-539.

Yu, H., Liu, J., Chen, C., Heidari, A. A., Zhang, Q., Chen, H., ... & Turabieh, H. (2021). Corn leaf diseases diagnosis based on K-means clustering and deep learning. IEEE Access, 9, 143824-143835.

Shaikh, T. A., Rasool, T., & Lone, F. R. (2022). Towards leveraging the role of machine learning and artificial intelligence in precision agriculture and smart farming. Computers and Electronics in Agriculture, 198, 107119.

Shin, D., He, S., Lee, G. M., Whinston, A. B., Cetintas, S., & Lee, K. C. (2020). Enhancing social media analysis with visual data analytics: A deep learning approach (pp. 1459-1492). SSRN.

Zhang, H., Ouyang, C., Zuo, D., Zhou, H., Li, G., Huang, Z., & Yang, K. (2023). UVM++: A Large-scale Benchmark for Beverage Recognition in Intelligent Vending Machine. IEEE Transactions on Consumer Electronics.

Thakur, D., Saini, J. K., & Srinivasan, S. (2023). DeepThink IoT: The Strength of Deep Learning in Internet of Things. Artificial Intelligence Review, 1-68.

Downloads

Published

30.11.2023

How to Cite

Yang, X. ., & Ismail, A. I. B. . (2023). Multimodal Deep Learning Architecture to Evaluate Emotion Recognition in Tea Packing. International Journal of Intelligent Systems and Applications in Engineering, 12(6s), 521–532. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/3993

Issue

Section

Research Article