Abstract

Recognizing human actions in video has gradually attracted much attention in computer vision community, however, it also faces many realistic challenges caused by background clutter, viewpoint changes, variation of actors appearance. These challenges reflect the difficulty of obtaining a clean and discriminative video representation for classification. Recently, VLAD (Vector of Locally Aggregated Descriptors) has shown to be a simple and efficient encoding scheme to obtain discriminative video representations. However, VLAD uses only the nearest visual word in codebook to aggregate each descriptor feature no matter whether it is appropriate or not. Inspired by visual word ambiguity and salience encoding in image classification, we propose Uncertain VLAD (UVLAD) encoding scheme which aggregates each local descriptor feature by considering multiple nearest visual words. The proposed UVLAD scheme ensures each descriptor to be aggregated or discarded appropriately. We evaluate our method on two different benchmark datasets: KTH, and YouTube. Results from experiments show that our encoding scheme outperforms the state-of-arts methods in most cases.

Keywords:
Discriminative model Codebook Computer science Artificial intelligence Encoding (memory) Pattern recognition (psychology) Feature (linguistics) Ambiguity Bag-of-words model in computer vision Salience (neuroscience) Visual Word Representation (politics) Aggregate (composite) Image (mathematics) Image retrieval

Metrics

2
Cited By
0.00
FWCI (Field Weighted Citation Impact)
22
Refs
0.07
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Human Pose and Action Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Video Surveillance and Tracking Methods
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
© 2026 ScienceGate Book Chapters — All rights reserved.