A comparative home activity monitoring study using visual and inertial sensors
Monitoring actions at home can provide essential information for rehabilitation management. This paper presents a comparative study and a dataset for the fully automated, sample-accurate recognition of common home actions in the living room environment using commercial-grade, inexpensive inertial an...
Saved in:
| Published in | 2015 17th International Conference on E-health Networking, Application & Services (HealthCom) pp. 644 - 647 |
|---|---|
| Main Authors | , , , , , , , |
| Format | Conference Proceeding |
| Language | English |
| Published |
IEEE
01.10.2015
|
| Subjects | |
| Online Access | Get full text |
| DOI | 10.1109/HealthCom.2015.7454583 |
Cover
| Summary: | Monitoring actions at home can provide essential information for rehabilitation management. This paper presents a comparative study and a dataset for the fully automated, sample-accurate recognition of common home actions in the living room environment using commercial-grade, inexpensive inertial and visual sensors. We investigate the practical home-use of body-worn mobile phone inertial sensors together with an Asus Xmotion RGB-Depth camera to achieve monitoring of daily living scenarios. To test this setup against realistic data, we introduce the challenging SPHERE-H130 action dataset containing 130 sequences of 13 household actions recorded in a home environment. We report automatic recognition results at maximal temporal resolution, which indicate that a vision-based approach outperforms accelerometer provided by two phone-based inertial sensors by an average of 14.85% accuracy for home actions. Further, we report improved accuracy of a vision-based approach over accelerometry on particularly challenging actions as well as when generalising across subjects. |
|---|---|
| DOI: | 10.1109/HealthCom.2015.7454583 |