Violence Detection With Two-Stream Neural Network Based on C3D

4Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.

Abstract

In recent years, violence detection has gradually turned into an important research area in computer vision with many proposed models with high accuracy. However, there is unsatisfactory generalization ability of these methods over different datasets. In this paper, the authors propose a violence detection method based on C3D two-stream network for spatiotemporal features. First, the authors preprocess the video data of RGB stream and optical stream respectively. Second, the authors feed the data into two C3D networks to extract features from the RGB flow and the optical flow respectively. Third, the authors fuse the features extracted by the two networks to obtain a final prediction result. To testify to the performance of the proposed model, four different datasets (two public datasets and two self-built datasets) are selected in this paper. The experimental results show that the model has good generalization ability compared to state-of-the-art methods since it not only has good ability on large-scale datasets but also performs well on small-scale datasets.

Cite

CITATION STYLE

APA

Lu, zanzan, Xia, X., Wu, H., & Yang, C. (2021). Violence Detection With Two-Stream Neural Network Based on C3D. International Journal of Cognitive Informatics and Natural Intelligence, 15(4). https://doi.org/10.4018/IJCINI.287601

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free