E03404025032

Page 1

The International Journal Of Engineering And Science (IJES) || Volume || 3 || Issue || 4 || Pages || 25-32 || 2014 || ISSN (e): 2319 – 1813 ISSN (p): 2319 – 1805

Action Recognition from Web Data 1,

Prof. Ms. R.R. Tuteja, 2, Ms. Shakeeba S. Khan

1, 2,

Department of Computer Science & Engg, PRMIT&R Amravati

--------------------------------------------------------------ABSTRACT------------------------------------------------------This paper proposes a generic approach of understanding human actions in uncontrolled video. The idea is to use images collected from the Web to learn representations of actions and use this knowledge to automatically annotate actions in videos. We use LDA to obtain a more compact and discriminative feature representation and binary SVMs for classification. Our approach is unsupervised in the sense that it requires no human intervention other than the text querying. We present experimental evidence that using action images collected from the Web. To our best knowledge, this is one of the first studies that try to recognize actions from web images.

Keywords - Annotate actions, Event Recognition, Generic database, Histogram of Oriented Gradients, tagging videos. -------------------------------------------------------------------------------------------------------------------------------------------Date of Submission: 14 April 2014 Date of Publication: 25 April 2014 --------------------------------------------------------------------------------------------------------------------------------------------

I. INTRODUCTION Human actions are Interaction with environment on specific purpose. Most research in human action recognition to date has focused on videos taken in controlled environments working with limited action vocabularies. Motion is a very important cue for recognizing actions. However, real world videos rarely exhibit such consistent and relatively simple settings. Instead, there is a wide range of environments where the actions can possibly take place, together with a large variety of possible actions that can be observed. Towards a more generic action recognition system, we propose to “learn” action representations from the Web and while doing this, improve the precision of the retrieved action images. Recent experiments show that action recognition based on key poses from single video frames is possible. But if the system is recognizing actions from real world videos this method require training with very large amounts of videos. Finding enough labeled video data that covers a diverse set of poses is quite challenging. Where else Web is a rich source of information, with many action images taken under various conditions and these are roughly annotated; i.e., the surrounding text is a clue used by search engines about the content of these images. Our apprehension is that one can use such a collection of images to learn certain pose instances of an action. Thus our work tries to join two lines of research “Internet vision” and “action recognition” together and makes it possible for one to benefit from the other. For our aim we need shape descriptors that are able to model the variations caused by high articulations. Our approach starts with employing a pose extractor, and then representing the pose via distribution of its rectangular regions. By using classification and feature reduction techniques, we test our representation via supervised and unsupervised settings.

II. WORKING OF THE SYSTEM The system first gathers images by simply querying the name of the action on a web image search engine like Google or Gigablast. Based on the assumption that the set of retrieved images contains relevant images of the queried action, we construct a dataset of action images in an incremental manner. This yields a large image set, which includes images of actions taken from multiple viewpoints in a range of environments, performed by people who have varying body proportions and different clothing. The images mostly present the “key poses” since these images try to convey the action with a single pose. Following figure 1 describe our system.

www.theijes.com

The IJES

Page 25


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.