Multi-modal egocentric activity recognition using multi-kernel learning

Creative Commons License

Arabaci M. A., Ozkan F., SÜRER E., Jancovic P., TEMİZEL A.

Multimedia Tools and Applications, vol.80, no.11, pp.16299-16328, 2021 (SCI-Expanded) identifier identifier

  • Publication Type: Article / Article
  • Volume: 80 Issue: 11
  • Publication Date: 2021
  • Doi Number: 10.1007/s11042-020-08789-7
  • Journal Name: Multimedia Tools and Applications
  • Journal Indexes: Science Citation Index Expanded (SCI-EXPANDED), Scopus, FRANCIS, ABI/INFORM, Applied Science & Technology Source, Compendex, Computer & Applied Sciences, INSPEC, zbMATH
  • Page Numbers: pp.16299-16328
  • Keywords: Egocentric, First-person vision, Activity recognition, Multi-kernel learning, Multi-modality, SPEAKER, FUSION, KERNEL, DEPTH
  • Middle East Technical University Affiliated: Yes


© 2020, Springer Science+Business Media, LLC, part of Springer Nature.Existing methods for egocentric activity recognition are mostly based on extracting motion characteristics from videos. On the other hand, ubiquity of wearable sensors allow acquisition of information from different sources. Although the increase in sensor diversity brings out the need for adaptive fusion, most of the studies use pre-determined weights for each source. In addition, there are a limited number of studies making use of optical, audio and wearable sensors. In this work, we propose a new framework that adaptively weighs the visual, audio and sensor features in relation to their discriminative abilities. For that purpose, multi-kernel learning (MKL) is used to fuse multi-modal features where the feature and kernel selection/weighing and recognition tasks are performed concurrently. Audio-visual information is used in association with the data acquired from wearable sensors since they hold information on different aspects of activities and help building better models. The proposed framework can be used with different modalities to improve the recognition accuracy and easily be extended with additional sensors. The results show that using multi-modal features with MKL outperforms the existing methods.