Show/Hide Menu
Hide/Show Apps
Logout
Türkçe
Türkçe
Search
Search
Login
Login
OpenMETU
OpenMETU
About
About
Open Science Policy
Open Science Policy
Open Access Guideline
Open Access Guideline
Postgraduate Thesis Guideline
Postgraduate Thesis Guideline
Communities & Collections
Communities & Collections
Help
Help
Frequently Asked Questions
Frequently Asked Questions
Guides
Guides
Thesis submission
Thesis submission
MS without thesis term project submission
MS without thesis term project submission
Publication submission with DOI
Publication submission with DOI
Publication submission
Publication submission
Supporting Information
Supporting Information
General Information
General Information
Copyright, Embargo and License
Copyright, Embargo and License
Contact us
Contact us
Egocentric activity recognition using two-stage decision fusion
Date
2024-01-01
Author
Arabacı, Mehmet Ali
Sürer, Elif
Temizel, Alptekin
Metadata
Show full item record
This work is licensed under a
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License
.
Item Usage Stats
29
views
0
downloads
Cite This
The widespread adoption of wearable devices equipped with advanced sensor technologies has fueled the rapid growth of egocentric video capture, known as First Person Vision (FPV). Unlike traditional third-person videos, FPV exhibits distinct characteristics such as significant ego-motions and frequent scene changes, rendering conventional vision-based methods ineffective. This paper introduces a novel audio-visual decision fusion framework for egocentric activity recognition (EAR) that addresses these challenges. The proposed framework employs a two-stage decision fusion pipeline with explicit weight learning, integrating both audio and visual cues to enhance overall recognition performance. Additionally, a new publicly available dataset, the Egocentric Outdoor Activity Dataset, comprising 1392 video clips featuring 30 diverse outdoor activities, is also introduced to facilitate comparative evaluations of EAR algorithms and spur further research in the field. Experimental results demonstrate that the integration of audio and visual information significantly improves activity recognition performance, outperforming single modality approaches and equally weighted decisions from multiple modalities.
Subject Keywords
Decision fusion
,
Egocentric activity recognition
,
First-person vision
,
Multi-modality
URI
https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85205268480&origin=inward
https://hdl.handle.net/11511/111706
Journal
Neural Computing and Applications
DOI
https://doi.org/10.1007/s00521-024-10463-0
Collections
Graduate School of Informatics, Article
Citation Formats
IEEE
ACM
APA
CHICAGO
MLA
BibTeX
M. A. Arabacı, E. Sürer, and A. Temizel, “Egocentric activity recognition using two-stage decision fusion,”
Neural Computing and Applications
, pp. 0–0, 2024, Accessed: 00, 2024. [Online]. Available: https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85205268480&origin=inward.