Open Access Open Access  Restricted Access Subscription or Fee Access

A Methodology for Extracting Standing Human Bodies from Video Stream

R. Dheepamuthuvalli, V. Sakthivel

Abstract


Imaging of human body segments is demanding task which supports many applications such as understanding of scenes and recognition of activities. A bottom-up technology for extracting human bodies automatically from single image, in case of almost upright position, is the available technique in cluttered environments. The dimension, position and face color are used for localizing human body, model construction of upper and lower body as per anthropometric constraints and skin color calculation. A highest level pose can be extracted by combining different levels of segmentation granularity. Jointly estimating the foreground and background during the body part search phase gives rise to the segments of human body, that alleviates the need for shape matching exactly. We have now proposed a system in which human body is extracted from video and from complex pose scenes in motion vectors. In the future, we wish to deal with more Complex poses, without depending on strong pose prior. More masks can be incorporated to deal with problems such as missing of extreme regions like hair, gloves and shoes, in the search of these parts, but care should be taken in keeping the computational complexity from rising excessively. 40 images (43 persons) from dataset of INRIA persons and 163 images from the dataset of ‘lab1’ are used for measuring the performance of our algorithm, where the accuracies measured are 89.53% and 97.68% respectively. Qualitative and quantitative experimental results prove than our technique outperforms state-of-the-art interactive and hybrid top-down/bottom-up approaches.


Keywords


Trajectories, Denoising, human Detection, Multi- Action Detection, Body Parts Based Classification

Full Text:

PDF

References


M. Andriluka, S. Roth, and B. Schiele, “Pictorial structures revisited: People detection and articulated pose estimation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recog., 2009, pp. 1014–1021.

J. Malik, S. Belongie, T. leung, and J. Shi, “Contour and texture analysis for image segmentation,” Int. J. Comput. Vis., vol. 43, no. 1, 2001, pp. 7-27.

V. Ferrari, M. Marin-Jimenez, and A. Zisserman, “Progressive search space reduction for human pose estimation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recog., 2008, pp. 1–8.

M. P. Kumar, A. Zisserman, and P. H. Torr, “Efficient discriminative learning of parts-based models,” in Proc. IEEE 12th Int. Conf. Comput. Vis. 2009, pp. 552–559.

V. Delaitre, I. Laptev, and J. Sivic, “Recognizing human actions in still images: A study of bag-of-features and part-based representations,” in Proc. IEEE Brit. Mach. Vis. Conf., 2010.

M. Eichner and V. Ferrari, “Better appearance models for pictorial structures,” in proc. Brit. Mach. Vis. Conf., 2009.

B. Yao and L. Fei-Fei, “Grouplet: A structured image representation for recognizing human and object interactions,” in Proc. IEEE Conf. Comput.Vis. Pattern Recog., 2010, pp. 9–16.

P. Buehler, M. Everingham, D. P. Huttenlocher, and A. Zisserman, “Long term arm and hand tracking for continuous sign language TV broadcasts,”in Proc. 19th Brit. Mach. Vis. Conf., 2008, pp. 1105–1114.

A. Farhadi and D. Forsyth, “Aligning ASL for statistical translation using a discriminative word model,” in Proc. IEEE Comput. Soc. Conf. Comput.Vis. Pattern Recog., 2006, pp. 1471–1476.

L. Zhao and L. S. Davis, “Iterative figure-ground discrimination,” in Proc.17th Int. Conf. Pattern Recog., 2004, pp. 67–70.

L. Grady, “Random walks for image segmentation,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 11, Nov. 2006, pp. 1768–1783.

C. Rother, V. Kolmogorov, and A. Blake, “Grabcut: Interactive foregroundextraction using iterated graph cuts,” ACM Trans. Graph., vol. 23, no. 3, Aug 2004 pp. 309–314.

V. Gulshan, C. Rother, A. Criminisi, A. Blake, and A. Zisserman,“Geodesic star convexity for interactive image segmentation,” in Proc.IEEE Conf. Comput. Vis. Pattern Recog., 2010, pp. 3129–3136.

Y. Y. Boykov and M.-P.Jolly, “Interactive graph cuts for optimal boundary& region segmentation of objects in ND images,” in Proc. IEEE 8th Int.Conf.Comput. Vis., 2001, pp. 105–112.

M. P. Kumar, P. H. S. Ton, and A. Zisserman, “Obj cut,” in Proc. IEEEComput.Soci. Conf. Comput. Vision Pattern Recog., 2005, pp. 18–25.

M.Y. Liu, O. Tuzel, S. Ramalingam, and R. Chellappa, “Entropy rate superpixel segmentation,” in Proc. IEEE Conf. Comput. Vis. Pattern Recog., 2011, pp. 2097–2104.

L. Huang, S. Tang, Y. Zhang, S. Lian, and S. Lin, “Robust human body segmentation based on part appearance and spatial constraint,” Neurocomputing, vol. 118, 2013, pp. 191–202.

P. F. Felzenszwalb and D. P. Huttenlocher, “Pictorial structures for objectrecognition,” Int. J. Comput.Vis., vol. 61, no. 1, 2005. pp. 55–79.

D. Ramanan, “Learning to parse images of articulated bodies,” Adv. Neur.Inf. Process.Sys., 2006,pp. 1129–1136.


Refbacks

  • There are currently no refbacks.


Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.