A Novel Deep Learning Technique for Detection of Violent Content in Videos

Authors

  • Asmita Poojari NITTE (Deemed to be University), Dept. of Computer Science and Engineering, NMAM Institute of Technology, Nitte - 574110, Karnataka, India
  • Pallavi K. N. NITTE (Deemed to be University), Dept. of Computer Science and Engineering, NMAM Institute of Technology, Nitte - 574110, Karnataka, India
  • McEnroe Ryan Dsilva Cloud Associate, Niveus Solutions, Udupi-576101, Karnataka, India
  • Jagadevi N. Kalshetty Dept. of Computer Science and Engineering, Nitte Meenakshi Institute of Technology, Bengaluru-560064, Karnataka , India

Keywords:

Content moderation, video classifier, specificity, violent video

Abstract

With a data of 2.5 quintillion bytes worth of data that is being generated daily, regulation of media uploaded on the social media sites such as Facebook, Instagram, and Reddit is a challenge. In addition to social media platforms, there are also private messaging platforms like WhatsApp and Microsoft Teams which are used by private companies for information exchange, team collaborations, and team conversations which must be content regulated. Content moderation is performed by people (moderators) who have to manually classify content into safe and not safe for work. The exposure of human content moderators to harmful and violent-content over the internet makes moderation less  desirable. In this study, we suggest and create a Machine Learning Model that can recognise violent video content and classify them as violent            and non-violent. Audio and video are the two parameters we use as inputs. The input video along with the audio is initially processed, if the audio is classified as violent, then the video is marked and classified as violent video. The associated video is put into a video classifier where it is further classed as violent or non-violent if the audio classifier deems the audio to be non-violent. Performance indicators like precision, accuracy, sensitivity, and specificity are used to demonstrate the performance of the suggested model.

Downloads

Download data is not yet available.

References

Yakaiah Potharaju, Manjunathachari Kamsali, Chennakesava Reddy Kesavari. “Classification of Ontological Violence Content Detection through Audio Features and Supervised Learning” International Journal of Intelligent Engineering and Systems, Vol.12, No.3, 2019.

Theodoros Giannakopoulos, Dimitris Kosmopoulos, Andreas Aristidou , S. Theodoridis. “Violence Content Classification Using Audio Features” Advances in Artificial Intelligence, 4th Helenic Conference on AI, SETN 2006, Heraklion, Crete, Greece, May 18-20, 2006, Proceedings

H. Wang, L. Yang, X. Wu and J. He, "A review of bloody violence in video classification," 2017 International Conference on the Frontiers and Advances in Data Science (FADS)

Accattoli, Simone & Sernani, Paolo & Falcionelli, Nicola & Mekuria, Dagmawi & Dragoni, Aldo Franco. (2020). ”Violence Detection in Videos by Combining 3D Convolutional Neural Networks and Support Vector Machines”. Applied Artificial Intelligence, February 2020.

Vosta, Soheil, and Kin-Choong Yow. "A CNN- RNN Combined Structure for Real-World Violence Detection in Surveillance Cameras", Applied Sciences, 2022.

Bruno Peixoto, et.al. “Multimodal Violence Detection in Videos”, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020.

Hospedales, T.; Gong, S.; Xiang, T. Video behaviour mining using a dynamic topic model. Int. J. Comput Vis. 2012, 98, 303–323

Sulman, N.; Sanocki, T.; Goldgof, D.; Kasturi, R. How effective is human video surveillance performance? In Proceedings of the 2008 19th IEEE International Conference on Pattern Recognition, ICPR, Tampa, FL, USA, 8–11 December 2008; pp. 1–3.

Nguyen, T.N.; Meunier, J. Anomaly detection in video sequence with appearance-motion correspondence. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, ICCV, Seoul, Korea, 27 October–2 November 2019; pp. 1273–1283

Tian, B.; Morris, B.T.; Tang, M.; Liu, Y.; Yao, Y.; Gou, C.; Shen, D.; Tang, S. Hierarchical and networked vehicle surveillance in its: A survey. IEEE Trans. Intell. Transp. Syst. 2017, 18, 25–48.

Yu, J.; Yow, K.C.; Jeon, M. Joint representation learning of appearance and motion for abnormal event detection. Mach. Vision Appl. 2018, 29, 1157–1170.

Varadarajan, J.; Odobez, J.M. Topic models for scene analysis and abnormality detection. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops, Kyoto, Japan, 27 September–4 October 2009; pp. 1338–1345.

Sodemann, A.A.; Ross, M.P.; Borghetti, B.J. A review of anomaly detection in automated surveillance. IEEE Trans. Syst. Man, Cybern. Part C (Appl. Rev.) 2012, 42, 1257–1272.

Zweng, A.; Kampel, M. Unexpected human behavior recognition in image sequences using multiple features. In Proceedings of the 2010 20th International Conference on Pattern Recognition, ICPR, Istanbul, Turkey, 23–26 August 2010; pp. 368–371.

Jodoin, P.M.; Konrad, J.; Saligrama, V. Modeling background activity for behavior subtraction. In Proceedings of the 2008 Second ACM/IEEE International Conference on Distributed Smart Cameras, Trento, Italy, 9–11 September 2008; pp. 1–10.

Dong, Q.; Wu, Y.; Hu, Z. Pointwise motion image (PMI): A novel motion representation and its applications to abnormality detection and behavior recognition. IEEE Trans. Circuits Syst. Video Technol. 2009, 19, 407–416.

Mecocci, A.; Pannozzo, M.; Fumarola, A. Automatic detection of anomalous behavioural events for advanced real-time video surveillance. In Proceedings of the 3rd International Workshop on Scientific Use of Submarine Cables and Related Technologies, Lugano, Switzerland, 31 July 2003; pp. 187–192.

Li, H.P.; Hu, Z.Y.; Wu, Y.H.; Wu, F.C. Behavior modeling and abnormality detection based on semi- supervised learning method. Ruan Jian Xue Bao (J. Softw.) 2007, 18, 527–537.

Yao, B.; Wang, L.; Zhu, S.C. Learning a scene contextual model for tracking and abnormality detection. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Anchorage, AK, USA, 23– 28 June 2008; pp. 1–8.

Yin, J.; Yang, Q.; Pan, J.J. Sensor-based abnormal human-activity detection. IEEE Trans. Knowl. Data Eng. 2008, 20, 1082–1090.

Benezeth, Y.; Jodoin, P.M.; Saligrama, V.; Rosenberger, C. Abnormal events detection based on spatio-temporal co-occurences. In Proceedings of the 2009 IEEE conference on computer vision and pattern recognition CVPR, Miami, FL, USA, 20–25 June 2009; pp. 2458–2465.

Begum, S. . S. ., Prasanth, K. D. ., Reddy, K. L. ., Kumar, K. S. ., & Nagasree, K. J. . (2023). RDNN for Classification and Prediction of Rock or Mine in Underwater Acoustics. International Journal on Recent and Innovation Trends in Computing and Communication, 11(3), 98–104. https://doi.org/10.17762/ijritcc.v11i3.6326

Wilson, T., Johnson, M., Gonzalez, L., Rodriguez, L., & Silva, A. Machine Learning Techniques for Engineering Workforce Management. Kuwait Journal of Machine Learning, 1(2). Retrieved from http://kuwaitjournals.com/index.php/kjml/article/view/120

Downloads

Published

21.09.2023

How to Cite

Poojari , A. ., K. N. , P. ., Dsilva , M. R. ., & Kalshetty, J. N. . (2023). A Novel Deep Learning Technique for Detection of Violent Content in Videos. International Journal of Intelligent Systems and Applications in Engineering, 11(4), 638–644. Retrieved from https://ijisae.org/index.php/IJISAE/article/view/3598

Issue

Section

Research Article