Automatically detecting pain using facial actions

Pain is generally measured by patient self-report, normally via verbal communication. However, if the patient is a child or has limited ability to communicate (i.e. the mute, mentally impaired, or patients having assisted breathing) self-report may not be a viable measurement. In addition, these sel...

Full description

Saved in:
Bibliographic Details
Published in2009 3rd International Conference on Affective Computing and Intelligent Interaction and Workshops Vol. 2009; pp. 1 - 8
Main Authors Lucey, P., Cohn, J., Lucey, S., Matthews, I., Sridharan, S., Prkachin, K.M.
Format Conference Proceeding Journal Article
LanguageEnglish
Published United States IEEE 08.12.2009
Subjects
Online AccessGet full text
ISBN9781424448005
142444800X
ISSN2156-8103
2156-8111
DOI10.1109/ACII.2009.5349321

Cover

More Information
Summary:Pain is generally measured by patient self-report, normally via verbal communication. However, if the patient is a child or has limited ability to communicate (i.e. the mute, mentally impaired, or patients having assisted breathing) self-report may not be a viable measurement. In addition, these self-report measures only relate to the maximum pain level experienced during a sequence so a frame-by-frame measure is currently not obtainable. Using image data from patients with rotator-cuff injuries, in this paper we describe an AAM-based automatic system which can detect pain on a frame-by-frame level. We do this two ways: directly (straight from the facial features); and indirectly (through the fusion of individual AU detectors). From our results, we show that the latter method achieves the optimal results as most discriminant features from each AU detector (i.e. shape or appearance) are used.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISBN:9781424448005
142444800X
ISSN:2156-8103
2156-8111
DOI:10.1109/ACII.2009.5349321