Action Recognition Using Particle Flow Fields

Action Recognition Using Particle Flow Fields PDF Author: Kishore K. Reddy
Publisher:
ISBN:
Category :
Languages : en
Pages : 102

Get Book Here

Book Description
In recent years, research in human action recognition has advanced on multiple fronts to address various types of actions including simple, isolated actions in staged data (e.g., KTH dataset), complex actions (e.g., Hollywood dataset), and naturally occurring actions in surveillance videos (e.g, VIRAT dataset). Several techniques including those based on gradient, flow, and interest-points, have been developed for their recognition. Most perform very well in standard action recognition datasets, but fail to produce similar results in more complex, large-scale datasets. Action recognition on large categories of unconstrained videos taken from the web is a very challenging problem compared to datasets like KTH (six actions), IXMAS (thirteen actions), and Weizmann (ten actions). Challenges such as camera motion, different viewpoints, huge interclass variations, cluttered background, occlusions, bad illumination conditions, and poor quality of web videos cause the majority of the state-of-the-art action recognition approaches to fail. An increasing number of categories and the inclusion of actions with high confusion also increase the difficulty of the problem. The approach taken to solve this action recognition problem depends primarily on the dataset and the possibility of detecting and tracking the object of interest. In this dissertation, a new method for video representation is proposed and three new approaches to perform action recognition in different scenarios using varying prerequisites are presented. The prerequisites have decreasing levels of difficulty to obtain: 1) Scenario requires human detection and tracking to perform action recognition; 2) Scenario requires background and foreground separation to perform action recognition; and 3) No pre-processing is required for action recognition. First, we propose a new video representation using optical flow and particle advection. The proposed "Particle Flow Field" (PFF) representation has been used to generate motion descriptors and tested in a Bag of Video Words (BoVW) framework on the KTH dataset. We show that particle flow fields has better performance than other low-level video representations, such as 2D-Gradients, 3D-Gradients and optical flow. Second, we analyze the performance of the state-of-the-art technique based on the histogram of oriented 3D-Gradients in spatio temporal volumes, where human detection and tracking are required. We use the proposed particle flow field and show superior results compared to the histogram of oriented 3D-Gradients in spatio temporal volumes. The proposed method, when used for human action recognition, just needs human detection and does not necessarily require human tracking and figure centric bounding boxes. It has been tested on KTH (six actions), Weizmann (ten actions), and IXMAS (thirteen actions, 4 different views) action recognition datasets. Third, we propose using the scene context information obtained from moving and stationary pixels in the key frames, in conjunction with motion descriptors obtained using Bag of Words framework, to solve the action recognition problem on a large (50 actions) dataset with videos from the web. We perform a combination of early and late fusion on multiple features to handle the huge number of categories. We demonstrate that scene context is a very important feature for performing action recognition on huge datasets. The proposed method needs separation of moving and stationary pixels, and does not require any kind of video stabilization, person detection, or tracking and pruning of features. Our approach obtains good performance on a huge number of action categories. It has been tested on the UCF50 dataset with 50 action categories, which is an extension of the UCF YouTube Action (UCF11) Dataset containing 11 action categories. We also tested our approach on the KTH and HMDB51 datasets for comparison. Finally, we focus on solving practice problems in representing actions by bag of spatio temporal features (i.e. cuboids), which has proven valuable for action recognition in recent literature. We observed that the visual vocabulary based (bag of video words) method suffers from many drawbacks in practice, such as: (i) It requires an intensive training stage to obtain good performance; (ii) it is sensitive to the vocabulary size; (iii) it is unable to cope with incremental recognition problems; (iv) it is unable to recognize simultaneous multiple actions; (v) it is unable to perform recognition frame by frame. In order to overcome these drawbacks, we propose a framework to index large scale motion features using Sphere/Rectangle-tree (SR-tree) for incremental action detection and recognition. The recognition comprises of the following two steps: 1) recognizing the local features by non-parametric nearest neighbor (NN), and 2) using a simple voting strategy to label the action. It can also provide localization of the action. Since it does not require feature quantization it can efficiently grow the feature-tree by adding features from new training actions or categories. Our method provides an effective way for practical incremental action recognition. Furthermore, it can handle large scale datasets because the SR-tree is a disk-based data structure. We tested our approach on two publicly available datasets, the KTH dataset and the IXMAS multi-view dataset, and achieved promising results.

Action Recognition Using Particle Flow Fields

Action Recognition Using Particle Flow Fields PDF Author: Kishore K. Reddy
Publisher:
ISBN:
Category :
Languages : en
Pages : 102

Get Book Here

Book Description
In recent years, research in human action recognition has advanced on multiple fronts to address various types of actions including simple, isolated actions in staged data (e.g., KTH dataset), complex actions (e.g., Hollywood dataset), and naturally occurring actions in surveillance videos (e.g, VIRAT dataset). Several techniques including those based on gradient, flow, and interest-points, have been developed for their recognition. Most perform very well in standard action recognition datasets, but fail to produce similar results in more complex, large-scale datasets. Action recognition on large categories of unconstrained videos taken from the web is a very challenging problem compared to datasets like KTH (six actions), IXMAS (thirteen actions), and Weizmann (ten actions). Challenges such as camera motion, different viewpoints, huge interclass variations, cluttered background, occlusions, bad illumination conditions, and poor quality of web videos cause the majority of the state-of-the-art action recognition approaches to fail. An increasing number of categories and the inclusion of actions with high confusion also increase the difficulty of the problem. The approach taken to solve this action recognition problem depends primarily on the dataset and the possibility of detecting and tracking the object of interest. In this dissertation, a new method for video representation is proposed and three new approaches to perform action recognition in different scenarios using varying prerequisites are presented. The prerequisites have decreasing levels of difficulty to obtain: 1) Scenario requires human detection and tracking to perform action recognition; 2) Scenario requires background and foreground separation to perform action recognition; and 3) No pre-processing is required for action recognition. First, we propose a new video representation using optical flow and particle advection. The proposed "Particle Flow Field" (PFF) representation has been used to generate motion descriptors and tested in a Bag of Video Words (BoVW) framework on the KTH dataset. We show that particle flow fields has better performance than other low-level video representations, such as 2D-Gradients, 3D-Gradients and optical flow. Second, we analyze the performance of the state-of-the-art technique based on the histogram of oriented 3D-Gradients in spatio temporal volumes, where human detection and tracking are required. We use the proposed particle flow field and show superior results compared to the histogram of oriented 3D-Gradients in spatio temporal volumes. The proposed method, when used for human action recognition, just needs human detection and does not necessarily require human tracking and figure centric bounding boxes. It has been tested on KTH (six actions), Weizmann (ten actions), and IXMAS (thirteen actions, 4 different views) action recognition datasets. Third, we propose using the scene context information obtained from moving and stationary pixels in the key frames, in conjunction with motion descriptors obtained using Bag of Words framework, to solve the action recognition problem on a large (50 actions) dataset with videos from the web. We perform a combination of early and late fusion on multiple features to handle the huge number of categories. We demonstrate that scene context is a very important feature for performing action recognition on huge datasets. The proposed method needs separation of moving and stationary pixels, and does not require any kind of video stabilization, person detection, or tracking and pruning of features. Our approach obtains good performance on a huge number of action categories. It has been tested on the UCF50 dataset with 50 action categories, which is an extension of the UCF YouTube Action (UCF11) Dataset containing 11 action categories. We also tested our approach on the KTH and HMDB51 datasets for comparison. Finally, we focus on solving practice problems in representing actions by bag of spatio temporal features (i.e. cuboids), which has proven valuable for action recognition in recent literature. We observed that the visual vocabulary based (bag of video words) method suffers from many drawbacks in practice, such as: (i) It requires an intensive training stage to obtain good performance; (ii) it is sensitive to the vocabulary size; (iii) it is unable to cope with incremental recognition problems; (iv) it is unable to recognize simultaneous multiple actions; (v) it is unable to perform recognition frame by frame. In order to overcome these drawbacks, we propose a framework to index large scale motion features using Sphere/Rectangle-tree (SR-tree) for incremental action detection and recognition. The recognition comprises of the following two steps: 1) recognizing the local features by non-parametric nearest neighbor (NN), and 2) using a simple voting strategy to label the action. It can also provide localization of the action. Since it does not require feature quantization it can efficiently grow the feature-tree by adding features from new training actions or categories. Our method provides an effective way for practical incremental action recognition. Furthermore, it can handle large scale datasets because the SR-tree is a disk-based data structure. We tested our approach on two publicly available datasets, the KTH dataset and the IXMAS multi-view dataset, and achieved promising results.

Method of Particle Trajectory Recognition in Particle Flows of High Particle Concentration Using a Candidate Trajectory Tree Process with Variable Search Areas

Method of Particle Trajectory Recognition in Particle Flows of High Particle Concentration Using a Candidate Trajectory Tree Process with Variable Search Areas PDF Author:
Publisher:
ISBN:
Category :
Languages : en
Pages :

Get Book Here

Book Description
The application relates to particle trajectory recognition from a Centroid Population comprised of Centroids having an (x, y, t) or (x, y, f) coordinate. The method is applicable to visualization and measurement of particle flow fields of high particle. In one embodiment, the centroids are generated from particle images recorded on camera frames. The application encompasses digital computer systems and distribution mediums implementing the method disclosed and is particularly applicable to recognizing trajectories of particles in particle flows of high particle concentration. The method accomplishes trajectory recognition by forming Candidate Trajectory Trees and repeated searches at varying Search Velocities, such that initial search areas are set to a minimum size in order to recognize only the slowest, least accelerating particles which produce higher local concentrations. When a trajectory is recognized, the centroids in that trajectory are removed from consideration in future searches.

Special and Gifted Education: Concepts, Methodologies, Tools, and Applications

Special and Gifted Education: Concepts, Methodologies, Tools, and Applications PDF Author: Management Association, Information Resources
Publisher: IGI Global
ISBN: 1522500359
Category : Education
Languages : en
Pages : 2473

Get Book Here

Book Description
Diverse learners with exceptional needs require a specialized curriculum that will help them to develop socially and intellectually in a way that traditional pedagogical practice is unable to fulfill. As educational technologies and theoretical approaches to learning continue to advance, so do the opportunities for exceptional children. Special and Gifted Education: Concepts, Methodologies, Tools, and Applications is an exhaustive compilation of emerging research, theoretical concepts, and real-world examples of the ways in which the education of special needs and exceptional children is evolving. Emphasizing pedagogical innovation and new ways of looking at contemporary educational practice, this multi-volume reference work is ideal for inclusion in academic libraries for use by pre-service and in-service teachers, graduate-level students, researchers, and educational software designers and developers.

Assistive Technologies for Physical and Cognitive Disabilities

Assistive Technologies for Physical and Cognitive Disabilities PDF Author: Theng, Lau Bee
Publisher: IGI Global
ISBN: 1466673745
Category : Health & Fitness
Languages : en
Pages : 341

Get Book Here

Book Description
Research on assistive technologies is undergoing many developments in its effectiveness in helping those with varying impairments. New technologies are constantly being created, researched, and implemented for those who need these technological aides in daily life. Assistive Technologies for Physical and Cognitive Disabilities combines worldwide cases on people with physical and cognitive disabilities with the latest applications in assistive technologies. This reference work brings different researchers together under one title to discuss current findings, developments, and ongoing research in the area of rehabilitative technology. This reference book is of critical use to professionals, researchers, healthcare practitioners, caretakers, academicians, and students.

Computer Vision and Action Recognition

Computer Vision and Action Recognition PDF Author: Md. Atiqur Rahman Ahad
Publisher: Springer Science & Business Media
ISBN: 9491216201
Category : Computers
Languages : en
Pages : 228

Get Book Here

Book Description
Human action analyses and recognition are challenging problems due to large variations in human motion and appearance, camera viewpoint and environment settings. The field of action and activity representation and recognition is relatively old, yet not well-understood by the students and research community. Some important but common motion recognition problems are even now unsolved properly by the computer vision community. However, in the last decade, a number of good approaches are proposed and evaluated subsequently by many researchers. Among those methods, some methods get significant attention from many researchers in the computer vision field due to their better robustness and performance. This book will cover gap of information and materials on comprehensive outlook – through various strategies from the scratch to the state-of-the-art on computer vision regarding action recognition approaches. This book will target the students and researchers who have knowledge on image processing at a basic level and would like to explore more on this area and do research. The step by step methodologies will encourage one to move forward for a comprehensive knowledge on computer vision for recognizing various human actions.

Pattern Recognition and Machine Vision

Pattern Recognition and Machine Vision PDF Author: Patrick Shen-Pei Wang
Publisher: River Publishers
ISBN: 8792329365
Category : Computers
Languages : en
Pages : 481

Get Book Here

Book Description
In recent years, there has been a growing interest in the fields of pattern recognition and machine vision in academia and industries. New theories have been developed with new technology and systems designs in both hardware and software. They are widely applied to our daily life to solve real problems in diverse areas such as science, engineering, agriculture, e-commerce, education, robotics, government, medicine, games and animation, medical imaging analysis and diagnosis, military, and national security. The foundation of this field can be traced back to the late Prof. King-Sun Fu, one of the founding fathers of pattern recognition, who, with visionary insight, founded the International Association for Pattern Recognition in 1978. Almost 30 years later, the world has witnessed this field's rapid growth and development. It is probably true to say that most people are affected by or use applications of pattern recognition in daily life. Today, on the eve of 25th anniversary of the unfortunate and untimely passing of Prof. Fu, we are proud to produce this collection works from world renowned professionals and experts in pattern recognition and machine vision in honor and memory of the late Prof. King-Sun Fu. We hope this book will help further promote not only fundamental principles, systems, and technologies but also the vast range of applications that help in solving problems in daily life.

Computer Vision – ECCV 2018 Workshops

Computer Vision – ECCV 2018 Workshops PDF Author: Laura Leal-Taixé
Publisher: Springer
ISBN: 3030110249
Category : Computers
Languages : en
Pages : 763

Get Book Here

Book Description
The six-volume set comprising the LNCS volumes 11129-11134 constitutes the refereed proceedings of the workshops that took place in conjunction with the 15th European Conference on Computer Vision, ECCV 2018, held in Munich, Germany, in September 2018.43 workshops from 74 workshops proposals were selected for inclusion in the proceedings. The workshop topics present a good orchestration of new trends and traditional issues, built bridges into neighboring fields, and discuss fundamental technologies and novel applications.

Human Behavior Understanding

Human Behavior Understanding PDF Author: Albert Ali Salah
Publisher: Springer
ISBN: 3642147151
Category : Computers
Languages : en
Pages : 183

Get Book Here

Book Description
It was a great pleasure to organize the First International Workshop on Human Behavior Understanding (HBU), which took place as a satellite workshop to International Conference on Pattern Recognition (ICPR) on August 22, 2010, in Istanbul, Turkey. This workshop arose from the natural marriage of pattern recognitionwiththerapidlyadvancingareaofhumanbehavioranalysis.Ouraim was to gather researchersdealing with the problem of modeling human behavior under its multiple facets (expression of emotions, display of relational attitudes, performance of individual or joint actions, etc.), with particular attention to pattern recognition approaches that involve multiple modalities and those that model actual dynamics of behavior. The contiguity with ICPR, one of the most important events in the p- tern recognition and machine learning communities, is expected to foster cro- pollination with other areas, for example temporal pattern mining or time - ries analysis, which share their important methodological aspects with human behavior understanding. Furthermore, the presence of this workshop at ICPR was meant to attract researchers, in particular PhD students and postd- toral researchers, to work on the questions of human behavior understanding that is likely to play a major role in future technologies (ambient intelligence, human–robot interaction, arti?cial social intelligence, etc.), as witnessed by a number of researche?orts aimed at collecting and annotating large sets of multi sensor data,collected from observingpeople in naturaland often technologically challenging conditions.

Computer Vision in Human-Computer Interaction

Computer Vision in Human-Computer Interaction PDF Author: Thomas S. Huang
Publisher: Springer Science & Business Media
ISBN: 3540342028
Category : Computers
Languages : en
Pages : 130

Get Book Here

Book Description
This book constitutes the refereed proceedings of the International Workshop on Human-Computer Interaction, HCI/ECCV 2006. The 11 revised full papers presented were carefully reviewed and selected from 27 submissions. The papers address a wide range of theoretical and application issues in human-computer interaction ranging from face analysis, gesture and emotion recognition, and event detection to various applications in those fields.

Optical Flow and Trajectory Estimation Methods

Optical Flow and Trajectory Estimation Methods PDF Author: Joel Gibson
Publisher: Springer
ISBN: 3319449419
Category : Computers
Languages : en
Pages : 57

Get Book Here

Book Description
This brief focuses on two main problems in the domain of optical flow and trajectory estimation: (i) The problem of finding convex optimization methods to apply sparsity to optical flow; and (ii) The problem of how to extend sparsity to improve trajectories in a computationally tractable way. Beginning with a review of optical flow fundamentals, it discusses the commonly used flow estimation strategies and the advantages or shortcomings of each. The brief also introduces the concepts associated with sparsity including dictionaries and low rank matrices. Next, it provides context for optical flow and trajectory methods including algorithms, data sets, and performance measurement. The second half of the brief covers sparse regularization of total variation optical flow and robust low rank trajectories. The authors describe a new approach that uses partially-overlapping patches to accelerate the calculation and is implemented in a coarse-to-fine strategy. Experimental results show that combining total variation and a sparse constraint from a learned dictionary is more effective than employing total variation alone. The brief is targeted at researchers and practitioners in the fields of engineering and computer science. It caters particularly to new researchers looking for cutting edge topics in optical flow as well as veterans of optical flow wishing to learn of the latest advances in multi-frame methods. /div