Automatic annotation of daily activity from smartphone-based multisensory streams

25Citations
Citations of this article
56Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a system for automatic annotation of daily experience from multisensory streams on smartphones. Using smartphones as platform facilitates collection of naturalistic daily activity, which is difficult to collect with multiple on-body sensors or array of sensors affixed to indoor locations. However, recognizing daily activities in unconstrained settings is more challenging than in controlled environments: 1) multiples heterogeneous sensors equipped in smartphones are noisier, asynchronous, vary in sampling rates and can have missing data; 2) unconstrained daily activities are continuous, can occur concurrently, and have fuzzy onset and offset boundaries; 3) ground-truth labels obtained from the user's self-report can be erroneous and accurate only in a coarse time scale. To handle these problems, we present in this paper a flexible framework for incorporating heterogeneous sensory modalities combined with state-of-the-art classifiers for sequence labeling. We evaluate the system with real-life data containing 11721 minutes of multisensory recordings, and demonstrate the accuracy and efficiency of the proposed system for practical lifelogging applications. © 2013 ICST Institute for Computer Science, Social Informatics and Telecommunications Engineering.

Cite

CITATION STYLE

APA

Hamm, J., Stone, B., Belkin, M., & Dennis, S. (2013). Automatic annotation of daily activity from smartphone-based multisensory streams. In Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering (Vol. 110 LNICST, pp. 328–342). https://doi.org/10.1007/978-3-642-36632-1_19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free