A framework for human action detection via extraction of multimodal features.

This work discusses the application of an Artificial Intelligence technique called data extraction and a process-based ontology in constructing experimental qualitative models for video retrieval and detection. We present a framework architecture that uses multimodality features as the knowledge re...

Full description

Bibliographic Details
Main Author: N. A., Lili
Format: Article
Language:English
English
Published: 2009
Subjects:
Online Access:http://psasir.upm.edu.my/id/eprint/12694/
http://psasir.upm.edu.my/id/eprint/12694/1/A%20framework%20for%20human%20action%20detection%20via%20extraction%20of%20multimodal%20features.pdf
_version_ 1848841904676929536
author N. A., Lili
author_facet N. A., Lili
author_sort N. A., Lili
building UPM Institutional Repository
collection Online Access
description This work discusses the application of an Artificial Intelligence technique called data extraction and a process-based ontology in constructing experimental qualitative models for video retrieval and detection. We present a framework architecture that uses multimodality features as the knowledge representation scheme to model the behaviors of a number of human actions in the video scenes. The main focus of this paper placed on the design of two main components (model classifier and inference engine) for a tool abbreviated as VASD (Video Action Scene Detector) for retrieving and detecting human actions from video scenes. The discussion starts by presenting the workflow of the retrieving and detection process and the automated model classifier construction logic. We then move on to demonstrate how the constructed classifiers can be used with multimodality features for detecting human actions. Finally, behavioral explanation manifestation is discussed. The simulator is implemented in bilingual; Math Lab and C++ are at the backend supplying data and theories while Java handles all front-end GUI and action pattern updating. To compare the usefulness of the proposed framework, several experiments were conducted and the results were obtained by using visual features only (77.89% for precision; 72.10% for recall), audio features only (62.52% for precision; 48.93% for recall) and combined audiovisual (90.35% for precision; 90.65% for recall).
first_indexed 2025-11-15T07:50:39Z
format Article
id upm-12694
institution Universiti Putra Malaysia
institution_category Local University
language English
English
last_indexed 2025-11-15T07:50:39Z
publishDate 2009
recordtype eprints
repository_type Digital Repository
spelling upm-126942015-12-22T02:13:51Z http://psasir.upm.edu.my/id/eprint/12694/ A framework for human action detection via extraction of multimodal features. N. A., Lili This work discusses the application of an Artificial Intelligence technique called data extraction and a process-based ontology in constructing experimental qualitative models for video retrieval and detection. We present a framework architecture that uses multimodality features as the knowledge representation scheme to model the behaviors of a number of human actions in the video scenes. The main focus of this paper placed on the design of two main components (model classifier and inference engine) for a tool abbreviated as VASD (Video Action Scene Detector) for retrieving and detecting human actions from video scenes. The discussion starts by presenting the workflow of the retrieving and detection process and the automated model classifier construction logic. We then move on to demonstrate how the constructed classifiers can be used with multimodality features for detecting human actions. Finally, behavioral explanation manifestation is discussed. The simulator is implemented in bilingual; Math Lab and C++ are at the backend supplying data and theories while Java handles all front-end GUI and action pattern updating. To compare the usefulness of the proposed framework, several experiments were conducted and the results were obtained by using visual features only (77.89% for precision; 72.10% for recall), audio features only (62.52% for precision; 48.93% for recall) and combined audiovisual (90.35% for precision; 90.65% for recall). 2009 Article PeerReviewed application/pdf en http://psasir.upm.edu.my/id/eprint/12694/1/A%20framework%20for%20human%20action%20detection%20via%20extraction%20of%20multimodal%20features.pdf N. A., Lili (2009) A framework for human action detection via extraction of multimodal features. International Journal of Image Processing, 3 (2). pp. 73-79. ISSN 1985-2304 Human locomotion-Computer simulation. Motion perception (Vision). Computer vision. English
spellingShingle Human locomotion-Computer simulation.
Motion perception (Vision).
Computer vision.
N. A., Lili
A framework for human action detection via extraction of multimodal features.
title A framework for human action detection via extraction of multimodal features.
title_full A framework for human action detection via extraction of multimodal features.
title_fullStr A framework for human action detection via extraction of multimodal features.
title_full_unstemmed A framework for human action detection via extraction of multimodal features.
title_short A framework for human action detection via extraction of multimodal features.
title_sort framework for human action detection via extraction of multimodal features.
topic Human locomotion-Computer simulation.
Motion perception (Vision).
Computer vision.
url http://psasir.upm.edu.my/id/eprint/12694/
http://psasir.upm.edu.my/id/eprint/12694/1/A%20framework%20for%20human%20action%20detection%20via%20extraction%20of%20multimodal%20features.pdf