ABSTRACT An essential part of any activity recognition system claiming be truly real-time is the ability to perform feature extraction in real-time. We present, in this paper, a quite simple and computationally tractable approach for real-time human activity recognition that is based on simple statistical features. These features are simple and relatively small, accordingly they are easy and fast to be calculated, and further form a relatively low-dimensional feature space in which classification can be carried out robustly. On the Weizmann publicly benchmark dataset, promising results (i.e. 97.8%) have been achieved, showing the effectiveness of the proposed approach compared to the-state-of-the-art. Furthermore, the approach is quite fast and thus can provide timing guarantees to real-time applications.
Cite this paper
S. Sadek, A. Al-Hamadi, B. Michaelis and U. Sayed, "A Fast Statistical Approach for Human Activity Recognition," International Journal of Intelligence Science, Vol. 2 No. 1, 2012, pp. 9-15. doi: 10.4236/ijis.2012.21002.
 S. Sadek, A. Al-Hamadi, B. Michaelis and U. Sayed, “Human Actionrecognition: A Novel Scheme Using Fuzzy Log-Polar Histogram and Temporal Self-Similarity,” EURASIP Journal on Advances in Signal Processing, 2011.
 Y. G. Jiang, C. W. Ngo and J. Yang, “Towards Optimal bag-Offeatures for Object Categorization and Semantic Video Retrieval,” ACM International Conferences on Image and Video Retrieval, Vol. 8, 2007, pp. 494-501.
 R. Cutler and L. S. Davis, “Robust Real-Time Periodic Motion Detection, Analysis, and Applications,” IEEE Transactions on PAMI, Vol. 22, 2000, pp. 781-796.
 S. Sadek, A. Al-Hamadi, B. Michaelis and U. Sayed, “An Efficient Method for Real-Time Activity Recognition,” Proceedings of the International Conference on Soft Computing and Pattern Recognition, Paris, 2010, pp. 7- 10. doi:10.1109/SOCPAR.2010.5686433
 C. Thuran and V. Hlavaˇc, “Pose Primitive Based Human Action Recognition in Videos or Still Images,” Conference on Computer Vision and Pattern Recognition, 2008.
 S. Sadek, A. Al-Hamadi, B. Michaelis and U. Sayed, “Human Activity Recognition: A Scheme Using Multiple cues,” Proceedings of the International Symposium on Visual Computing, Las Vegas, 2010, pp. 574-583.
 W.-L. Lu, K. Okuma and J. J. Little, “Tracking and Recognizing Actions of Multiple Hockey Players Using the Boosted Particle Filter,” Image and Vision Computing, Vol. 27, 2009, pp. 189-205.
 S. Sadek, A. Al-Hamadi, M. Elmezain, B. Michaelis and U. Sayed, “Human Activity Recognition Using Temporal Shape Moments,” IEEE International Symposium on Signal Processing and Information Technology, Luxor, 2010, pp. 79-84. doi:10.1109/ISSPIT.2010.5711729
 S. Sadek, A. Al-Hamadi, B. Michaelis and U. Sayed, “Towards Robust Human Action Retrieval in Video,” Proceedings of the British Machine Vision Conference, Aberystwyth, 2010.
 P. Doll’ar, V. Rabaud, G. Cottrell and S. Belongie, “Behavior Recognition via Sparse Spatio-Temporal Featu- Res,” IEEE International Workshop on VS-PETS, 2005.
 J. Liu and M. Shah, “Learning Human Actions via Information Maximization,” Conference on Computer Vision and Pattern Recognition, Alaska, 2008.
 A. F. Bobick and J. W. Davis, “The Recognition of Human Movement Using Temporal Templates,” IEEE Transactions on PAMI, Vol. 23, No. 3, 2001, pp. 257-267.
 E. Shechtman and M. Irani, “Space-Time Behavior Based Correlation,” Conference on Computer Vision and Pattern Recognition, Vol. 1, 2005, pp. 405-412.
 M. D. Rodriguez, J. Ahmed and M. Shah, “Action MACH: A Spatio-Temporal Maximum Average Correlation Height Filter for Action Recognition,” Conference on Computer Vision and Pattern Recognition, Alaska, 2008.
 H. Jhuang, T. Serre, L. Wolf, and T. Poggio, “A biologically inspired system for action recognition,” International Conference on Computer Vision, Sophia Antipolis, 2007, pp. 257-267.
 K. Schindler and L. V. Gool, “Action Snippets: How Many Frames Does Action Recognition Require?” Conference on Computer Vision and Pattern Recognition, Alaska, 2008.
 B. Laxton, J. Lim and D. Kriegman, “Leveraging Temporal, Contextual and Ordering Constraints for Recognizing Complex Activities in Video,” Conference on Computer Vision and Pattern Recognition, Alaska, 2007, pp. 1-8. doi:10.1109/CVPR.2007.383074
 N. Olivera, A. Garg and E. Horvitz, “Layered Representations for Learning and Inferring Office Activity from Multiple Sensory Channels,” Computer Vision and Image Understanding, Vol. 96, No. 2, 2004, pp. 163-180.
 D. M. Blei and J. D. Lafferty, “Correlated Topic Models,” Advances in Neural Information Processing Systems, Vol. 18, 2006, pp. 147-154.
 V. N. Vapnik, “The Nature of Statistical Learning Theory,” Springer-Verlag, New York, 1995.
 M. Blank, L. Gorelick, E. Shechtman, M. Irani and R. Basri, “Actions as Space-Time Shapes,” International Conference on Computer Vision, Sophia Antipolis, 2005, pp. 1395-1402.
 M. Bregonzio, S. Gong and T. Xiang, “Recognising Action as Clouds of Space-Time Interest Points,” Conference on Computer Vision and Pattern Recognition, Alaska, 2009.
 Z. Zhang, Y. Hu, S. Chan and L.-T. Chia, “Motion Context: A New Representation for Human Action Recognition,” European Conference on Computer Vision, Crete, 2008, pp. 817-829.
 J. Niebles, H. Wang and L. Fei-Fei, “Unsupervised Learning of Human Action Categories Using Spatial-Temporal Words,” International Journal of Computer Vision, Vol. 79, No. 3, 2008, pp. 299-318.
 A. Fathi and G. Mori, “Action Recognition by Learning Midlevel Motion Features,” Conference on Computer Vision and Pattern Recognition, Alaska, 2008.