Role of Feature Position in recognition of Human Vehicle Interaction

  • Hina Arshad Institute of Space Technology, Islamabad, Pakistan
  • Khurram Khurshid Institute of Space Technology, Islamabad, Pakistan
  • M Haroon yousaf University of Engineering & Technology, Taxila, Pakistan
Keywords: Keywords, Human-Vehicle-Interaction-Recognition, Spatial Positions, Bag of Words (BoW), Spatiotemporal Features

Abstract

This paper proposes an improvement in the Bag of Words (BoW) model by introducing spatial position of features in interaction representation stage for Human Vehicle Interaction Recognition. The spatial positions of features are incorporated along with feature descriptor to obtain the structural information required to correctly classify different kind of human vehicle interaction. BoW lacks structural information as well as the relationship between spatiotemporal features. This renders this BoW approach, in its basic form, ineffective for human vehicle interaction recognition. The proposed approach enhances the BoW representation by emphasizing the role of spatial feature positions for interaction and incorporates the spatial and temporal relationship between the features. This approach has been tested on the state of the art dataset and achieved accuracy of 92.8%. It has been found that our approach out performs the Bag of Words approach and other state of the art work.

Author Biographies

Hina Arshad, Institute of Space Technology, Islamabad, Pakistan
Research Scholar at Institute of Space Technology, Islamabad, Pakistan
Khurram Khurshid, Institute of Space Technology, Islamabad, Pakistan
Assistant Professor in Electrical Engineering Deapartment at Institute of Space Technology, Islamabad, Pakistan
M Haroon yousaf, University of Engineering & Technology, Taxila, Pakistan
Assistant Professor in Computer Engineering Deapartment at University of Engineering & Technology, Taxila, Pakistan

References

REFERENCES

Bregonzio, M., Xiang, T. and Gong, S., 2012. Fusing appearance and distribution information of interest points for action recognition, Pattern Recognition, vol. 45, no. 3, and pp. 1220–1234.

Brendel, W. and Todorovic, S., 2011. Learning spatiotemporal graphs of human activities." In Computer Vision (ICCV), 2011 IEEE International Conference on, pp. 778-785.

Dollár, P., Rabaud, V., Cottrell, G. and Belongie, S., 2005. Behavior recognition via sparse spatio-temporal features, in Proceedings of the 2nd Joint IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance (VS-PETS '05), pp. 65–72

Lazebnik, S., Schmid, C. and Ponce, J., 2006. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on (Vol. 2, pp. 2169-2178). IEEE.X.

Efros, A.A., Berg, A.C., Mori, G. and Malik, J., 2003 Recognizing action at a distance. In: International Conference on Computer Vision.

Intille, S.S. and Bobick, A.F., 2001. Recognizing planned, multiperson action. Computer Vision and Image Understanding 81, no. 3 (2001): 414-445.

Lazebnik, S., Schmid, C. and Ponce, J., 2006. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on (Vol. 2, pp. 2169-2178). IEEE.X.

Niebles, J.C., Chen, C.W. and Fei-Fei, L., 2010. Modeling temporal structure of decomposable motion segments for activity classification. In Computer Vision–ECCV 2010, pp. 392-405. Springer Berlin Heidelberg.

Niebles, J.C., Wang, H. and Fei-Fei, L., 2008. Unsupervised learning of human action categories using spatial-temporal words. In Proc. of BMVC.

Reddy, K.K., Cuntoor, N., Perera, A. and Hoogs, A., 2012. Human action recognition in large-scale datasets using histogram of spatiotemporal gradients." In Advanced Video and Signal-Based Surveillance (AVSS), 2012 IEEE Ninth International Conference on, pp. 106-111

Raptis, M. and Sigal, L., 2013.Poselet key-framing: A model for human activity recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2650-2657

Schüldt, C., Laptev, I. and Caputo, B., 2004. Recognizing human actions: local SVM approach. In: International Conference on Pattern Recognition

Tang, K., Fei-Fei, L. and Koller, D., 2012. Learning latent temporal structure for complex event detection. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pp. 1250-1257

Tian, Y., Sukthankar, R. and Shah, M., 2013.Spatiotemporal deformable part models for action detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2642-2649.

Vahdat, A., Gao, B., Ranjbar, M. and Mori, G., 2011. A discriminative key pose sequence model for recognizing human interactions. In Computer Vision Workshops (ICCV Workshops), 2011 IEEE International Conference on, pp. 1729-1736

Wang, Y., Mori, G., 2013. Hidden part models for human action recognition: Probabilistic vs. max-margin. IEEE Transactions on Pattern Analysis and Machine Intelligence 33 (7), 1310–1323

Wu, X., Xu, D., Duan, L. and Luo, J., 2011 Action recognition using context and appearance distribution features,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '11), pp. 489–496

Yamato, J., Ohya, J. and Ishii, K., 1992. Recognizing human action in time-sequential images using hidden markov model. In Computer Vision and Pattern Recognition, 1992. Proceedings CVPR’92. IEEE Computer Society Conference on, pp. 379-385. IEEE

Published
2017-08-01
Section
Electrical Engineering