You are here

Spatial-Temporal Words Learning for Crowd Behavior Recognition

Journal Name:

Publication Year:

Abstract (2. Language): 
In this paper we introduce a method to detect abnormal behavior in crowd scenes using spatial-temporal words, in which Spatial Temporal Interest Points (STIPs) are extracted as crowd behavior movement features. For this purpose, three methods for STIPs extraction are compared and analyzed, which are Harris corner, Gabor wavelet and Hessian Matrix. Then we select Hessian matrix which can get scale-invariant STIPs. Gradient histogram, optical flow histogram and spatio-temporal Haar feature are used to build descriptors for STIPs. In normal behavior modeling Bag-of-words strategy is used, the keywords of which are produced by GMM based on EM estimation. Then training samples are divided into several clips which are described in probability vectors using keywords, we combine all vectors as a normal behavior codebook. In recognition phase, we calculate the similarity distance between the coding vector of the test samples and the codebook, the abnormal behavior can be detected when the minimal distance exceeds a threshold. We verify the algorithm in UMN and UCF datasets, which shows that the proposed algorithm has effective identification for crowd abnormal behavior, and it has good robustness against scale variant and illumination changing
1-6

REFERENCES

References: 

[1] R. Mehran, A. Oyama, and M. Shah. Abnormal crowd behavior
detection using social force model. IEEE Computer Society Conference
on Computer Vision and Pattern Recognition (CVPR), 2009, pages 935–
942.
[2] Shandong Wu, Brian E. Moore, Mubarak Shah. Chaotic invariants of
Lagrangian particle trajectories for anomaly detection in crowded
scenes. IEEE Computer Society Conference on Computer Vision and
Pattern Recognition (CVPR), 2010, pp.2054-2060.
[3] Vijay Mahadevan, Weixin Li, Viral Bhalodia, Nuno Vasconcelos.
Anomaly Detectionin Crowded Scenes, IEEE Conf. on Computer Vision
and Pattern Recognition, San Francisco, 2010.
[4] A. B. Chan and N. Vasconcelos. Modeling, clustering, and segmenting
video with mixtures of dynamic textures. PAMI, 30(5):909–926, May
2008.
[5] Ramin Mehran1, Brian E. Moore2, and Mubarak Shah. A Streakline
Representation of Flow in Crowded Scenes, ECCV 2010, Part III, LNCS
6313, pp. 439–452.
[6] Ronald Poppe. A survey on vision-based human action recognition[J].
Image and Vision Computing, 2010, 28:976–990.
[7] Ivan Laptev, Tony Lindeberg. Space-Time Interest Points. International
Conference on Computer Vision(ICCV), Nice, France, 2003, pp.I:432-439.
[8] Dollar P,Rabaud V,Cotterll G,Belongie S. Behavior recognition via
sparse spatio-temporal features. Proceedings of the 2nd Joint IEEE
International Workshop on Visual Surveil-lance and Performance
Evaluation of Tracking and Surveillance, Beijing,China,2005:65-72.
[9] G. WIllems, T. Tuytelaars, and L. V. Gool. An efficient dense and scale-invariant spatio-temporal interest point detector. Proceedings of the
European Conference on Computer Vision (ECCV), 2008, pages 650-663
10] Xinyi Cui, Yazhou Liu, Shiguang Shan, Xilin Chen, Wen Gao. 3D Haar-Like Features for Pedestrian Detection. IEEE International Conference
on Multimedia and Expo, July 2007, pp.1263-1266.

Thank you for copying data from http://www.arastirmax.com