A novel human posture estimation using single depth image from Kinect v2 sensor

In this paper, we propose an approach to estimate general posture of the human-body. We perform various 2D scans of the body to carry out their Fast-Fourier-Transform(FFT) to extract features which can be fed to a 2-layer feed-forward neural-network. This approach can offer an effective procedure gi...

Full description

Saved in:
Bibliographic Details
Published inAnnual IEEE Systems Conference pp. 1 - 7
Main Authors Maryam, S. Rasouli D., Payandeh, Shahram
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.04.2018
Subjects
Online AccessGet full text
ISSN2472-9647
DOI10.1109/SYSCON.2018.8369566

Cover

Abstract In this paper, we propose an approach to estimate general posture of the human-body. We perform various 2D scans of the body to carry out their Fast-Fourier-Transform(FFT) to extract features which can be fed to a 2-layer feed-forward neural-network. This approach can offer an effective procedure given limited resources which are usually available for deep learning approaches. In comparison with the literature, our proposed method doesn't require any specific skeleton points of the human body, results in a reduced level of computational complexities. We compared our method with the state-of-the-art and the results show an increased level of classification accuracy. The training dataset is captured from a single subject moving with various postures. If the test-data is also captured from the same subject, then the overall level of accuracy is 98.2% while if the test images are captured from different subjects then the accuracy of the estimation would be near 91.5%. We also test our method in the restricted field of view where parts of the subject's body are not in the field of sensor view, the result shows that in this case, the accuracy is as high as 79.1%.
AbstractList In this paper, we propose an approach to estimate general posture of the human-body. We perform various 2D scans of the body to carry out their Fast-Fourier-Transform(FFT) to extract features which can be fed to a 2-layer feed-forward neural-network. This approach can offer an effective procedure given limited resources which are usually available for deep learning approaches. In comparison with the literature, our proposed method doesn't require any specific skeleton points of the human body, results in a reduced level of computational complexities. We compared our method with the state-of-the-art and the results show an increased level of classification accuracy. The training dataset is captured from a single subject moving with various postures. If the test-data is also captured from the same subject, then the overall level of accuracy is 98.2% while if the test images are captured from different subjects then the accuracy of the estimation would be near 91.5%. We also test our method in the restricted field of view where parts of the subject's body are not in the field of sensor view, the result shows that in this case, the accuracy is as high as 79.1%.
Author Payandeh, Shahram
Maryam, S. Rasouli D.
Author_xml – sequence: 1
  givenname: S. Rasouli D.
  surname: Maryam
  fullname: Maryam, S. Rasouli D.
  organization: Network Robotics and Sensing Laboratory, School of Engineering Science, Simon Fraser University(SFU), Burnaby, Canada
– sequence: 2
  givenname: Shahram
  surname: Payandeh
  fullname: Payandeh, Shahram
  organization: Network Robotics and Sensing Laboratory, School of Engineering Science, Simon Fraser University(SFU), Burnaby, Canada
BookMark eNotj8tOwzAURA0Cibb0C7rxDyT4lZt4WUVAERVZtBtWVbCv26DEiWKnEn9PEd3MLI40OjMnd773SMiKs5Rzpp92n7uy-kgF40VaSNAZwA2Z80wWIAEU3JKZULlINKj8gSxD-GaMceAgNZ-Rak19f8aWnqau9nToQ5xGpBhi09Wx6T2dQuOP9C9apBaHeKIXdETqxr6j741HE-lZ0IA-9OMjuXd1G3B57QXZvzzvy02yrV7fyvU2aTSLidJQ5IDGskxoU1hm0FnFEJkDlSFaa6xjmTO5cdYJrqWudVYA_5JcKKHkgqz-ZxtEPAzjxWj8OVzvy1_JuVHi
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DOI 10.1109/SYSCON.2018.8369566
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan All Online (POP All Online) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Engineering
EISBN 1538636646
9781538636640
EISSN 2472-9647
EndPage 7
ExternalDocumentID 8369566
Genre orig-research
GroupedDBID 6IE
6IF
6IK
6IL
6IN
AAJGR
AAWTH
ABLEC
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
IPLJI
OCL
RIE
RIL
RNS
ID FETCH-LOGICAL-i90t-496876ecd0529c8d0cefd40ee0f645eeddcdf05fc7cfdf21939a95861b3124243
IEDL.DBID RIE
IngestDate Wed Aug 27 02:50:09 EDT 2025
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i90t-496876ecd0529c8d0cefd40ee0f645eeddcdf05fc7cfdf21939a95861b3124243
PageCount 7
ParticipantIDs ieee_primary_8369566
PublicationCentury 2000
PublicationDate 2018-April
PublicationDateYYYYMMDD 2018-04-01
PublicationDate_xml – month: 04
  year: 2018
  text: 2018-April
PublicationDecade 2010
PublicationTitle Annual IEEE Systems Conference
PublicationTitleAbbrev SYSCON
PublicationYear 2018
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0001616391
Score 1.6858315
Snippet In this paper, we propose an approach to estimate general posture of the human-body. We perform various 2D scans of the body to carry out their...
SourceID ieee
SourceType Publisher
StartPage 1
SubjectTerms Depth data
Estimation
Feature extraction
Hidden Markov models
human posture estimation
Kinect sensor
Neural networks
Robot sensing systems
Skeleton
Two dimensional displays
Title A novel human posture estimation using single depth image from Kinect v2 sensor
URI https://ieeexplore.ieee.org/document/8369566
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3NT8IwFG-Ak178AON3evDoxrZ2Hz0aIiEawARM8ES29hWNZCMwOPjX-7pNUOPBy9Js6da0Wd_79b3f7xFyw7QfJyDwR-KAAIWHjiV0IqzY5YjjpAhVwa3qD4LeM3-Y-JMaud1yYQCgSD4D2zSLWL7K5NoclbUjFqA7H9RJPQxFydXanacE6FkItxIWch3RHr2MOsOByd6K7KrnjxIqhQXpHpD-17fLxJF3e50ntvz4Jcv438EdktaOq0eftlboiNQgPSb732QGm2R4R9NsA3NaFOSji2xlwgbU6GuUxEVqst9n1FzmQBUs8leKj2ZADfuEPuKrZE43Hl0h6M2WLTLu3o87PasqpGC9CSe3uAhwzwOpTFRPRsqRoBV3ABwdcB-Hp6TSjq9lKLXSuIUxEQs_CtyEuYY9wk5II81SOCU01AzxjTLynujKILRjsecDD4XnAYKj-Iw0zcxMF6VUxrSalPO_b1-QPbM6ZSLMJWnkyzVcoY3Pk-ticT8BhMmnQQ
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LT8JAEN4gHtSLDzC-3YNHC31sH3s0RILyMgETPJF2dxaNpCVQOPjrnW0rqPHgpWnatN3spDvz7cz3DSE3jnLDCDj-SAwQoDDfNLiKuBFaDHGc4L7MuFXdntd6Zo8jd1Qit2suDABkxWdQ06dZLl8mYqm3yuqB42E4722RbRdRhZ-ztTY7Kh7GFtwqpIUsk9cHL4NGv6frt4Ja8eyPJiqZD2nuk-7X1_PSkffaMo1q4uOXMON_h3dAqhu2Hn1a-6FDUoL4iOx9ExqskP4djZMVTGnWko_OkoVOHFCtsJFTF6muf59QfZgClTBLXynemgDV_BPaxleJlK5sukDYm8yrZNi8HzZaRtFKwXjjZmow7uGqB0LqvJ4IpClASWYCmMpjLg5PCqlMVwlfKKlwEXN4yN3AsyLH0vwR55iU4ySGE0J95SDCkVrgE4MZNIMT2i4wn9s2IDwKT0lFz8x4lotljItJOfv78jXZaQ27nXHnodc-J7vaUnlZzAUpp_MlXKLHT6OrzNCfflmqkg
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=proceeding&rft.title=Annual+IEEE+Systems+Conference&rft.atitle=A+novel+human+posture+estimation+using+single+depth+image+from+Kinect+v2+sensor&rft.au=Maryam%2C+S.+Rasouli+D.&rft.au=Payandeh%2C+Shahram&rft.date=2018-04-01&rft.pub=IEEE&rft.eissn=2472-9647&rft.spage=1&rft.epage=7&rft_id=info:doi/10.1109%2FSYSCON.2018.8369566&rft.externalDocID=8369566