Localising speech, footsteps and other sounds using resource-constrained devices

While a number of acoustic localisation systems have been proposed over the last few decades, these have typically either relied on expensive dedicated microphone arrays and workstation-class processing, or have been developed to detect a very specific type of sound in a particular scenario. However...

Full description

Saved in:
Bibliographic Details
Published in2011 10th International Conference on Information Processing in Sensor Networks pp. 330 - 341
Main Authors Yukang Guo, Hazas, M
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.04.2011
Subjects
Online AccessGet full text
ISBN9781612848549
1612848540

Cover

Abstract While a number of acoustic localisation systems have been proposed over the last few decades, these have typically either relied on expensive dedicated microphone arrays and workstation-class processing, or have been developed to detect a very specific type of sound in a particular scenario. However, as people live and work indoors, they generate a wide variety of sounds as they interact and move about. These human-generated sounds can be used to infer the positions of people, without requiring them to wear trackable tags. In this paper, we take a practical yet general approach to localising a number of human-generated sounds. Drawing from signal processing literature, we identify methods for resource-constrained devices in a sensor network to detect, classify and locate acoustic events such as speech, footsteps and objects being placed onto tables. We evaluate the classification and time-of-arrival estimation algorithms using a data set of human-generated sounds we captured with sensor nodes in a controlled setting. We show that despite the variety and complexity of the sounds, their localisation is feasible for sensor networks, with typical accuracies of a half metre or better. We specifically discuss the processing and networking considerations, and explore the performance trade-offs which can be made to further conserve resources.
AbstractList While a number of acoustic localisation systems have been proposed over the last few decades, these have typically either relied on expensive dedicated microphone arrays and workstation-class processing, or have been developed to detect a very specific type of sound in a particular scenario. However, as people live and work indoors, they generate a wide variety of sounds as they interact and move about. These human-generated sounds can be used to infer the positions of people, without requiring them to wear trackable tags. In this paper, we take a practical yet general approach to localising a number of human-generated sounds. Drawing from signal processing literature, we identify methods for resource-constrained devices in a sensor network to detect, classify and locate acoustic events such as speech, footsteps and objects being placed onto tables. We evaluate the classification and time-of-arrival estimation algorithms using a data set of human-generated sounds we captured with sensor nodes in a controlled setting. We show that despite the variety and complexity of the sounds, their localisation is feasible for sensor networks, with typical accuracies of a half metre or better. We specifically discuss the processing and networking considerations, and explore the performance trade-offs which can be made to further conserve resources.
Author Yukang Guo
Hazas, M
Author_xml – sequence: 1
  surname: Yukang Guo
  fullname: Yukang Guo
  organization: Sch. of Comput. & Commun., Lancaster Univ., Lancaster, UK
– sequence: 2
  givenname: M
  surname: Hazas
  fullname: Hazas, M
  organization: Sch. of Comput. & Commun., Lancaster Univ., Lancaster, UK
BookMark eNotjM1KxDAURiMq6Ix9Ajd5AAtJk9s0Sxn8g4IuZj-kNzdOZExK0hF8ewf12xwOHL4Vu0g50RlbSQ1CCZCdPGeNNYPsZTfoAbS9Yk2tH-K0vrcdqGv2NmZ0h1hjeud1JsL9HQ85L3WhuXKXPM_Lngqv-Zh85cffsNBJC1KLOdWluJjIc09fEanesMvgDpWaf67Z9vFhu3lux9enl8392EYrlnaSQQRpbaDOTA4ECNn7KQSBQDqgMcpIAufRaRWERA9eBGMn6lABklZrdvt3G4loN5f46cr3DoyxQlv1A-XGTlQ
ContentType Conference Proceeding
DBID 6IE
6IL
CBEJK
RIE
RIL
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Xplore POP ALL
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP All) 1998-Present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 1450305121
9781450305129
EndPage 341
ExternalDocumentID 5779049
Genre orig-research
GroupedDBID 6IE
6IF
6IK
6IL
6IN
AAJGR
AAWTH
ADFMO
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
IEGSK
IERZE
OCL
RIE
RIL
ID FETCH-LOGICAL-i90t-b1f0f199fe27ba505016dbff0c5e4fc77371e5adca43f01cd5d0f79be2c35ce43
IEDL.DBID RIE
ISBN 9781612848549
1612848540
IngestDate Wed Aug 27 02:53:16 EDT 2025
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i90t-b1f0f199fe27ba505016dbff0c5e4fc77371e5adca43f01cd5d0f79be2c35ce43
PageCount 12
ParticipantIDs ieee_primary_5779049
PublicationCentury 2000
PublicationDate 2011-April
PublicationDateYYYYMMDD 2011-04-01
PublicationDate_xml – month: 04
  year: 2011
  text: 2011-April
PublicationDecade 2010
PublicationTitle 2011 10th International Conference on Information Processing in Sensor Networks
PublicationTitleAbbrev IPSN
PublicationYear 2011
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0000669253
Score 1.5445609
Snippet While a number of acoustic localisation systems have been proposed over the last few decades, these have typically either relied on expensive dedicated...
SourceID ieee
SourceType Publisher
StartPage 330
SubjectTerms Acoustics
Algorithm design and analysis
Design
Estimation
Experimentation
Feature extraction
General Terms-Algorithms
Measurement
Microphones
Noise
Speech
Title Localising speech, footsteps and other sounds using resource-constrained devices
URI https://ieeexplore.ieee.org/document/5779049
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PS8MwFA7bTp5UNvE3OXhcR9skS3MWxxAnO0zYbTTJCw6hHWt78a_3Je0migdvbXnQkEfyffnyfhDyoOU0Z1zLSCTaRVxoiDIGgCvecZ_aiajopYHF63T-xp_XYt0j42MuDACE4DOY-Mdwl29L03ipDA_vUiGj7ZO-zKZtrtZRT0HoVKnwbRuQw-CemyEV6Uo6Hd7Vjw4qAUBmp2Rx-HUbN_IxaWo9MZ-_qjL-d2xnZPSdqkeXRxA6Jz0ohmT54hFq62UAWu0AzPuYurKs0aO7iuaFpSHvila-p1JFm2C474T8yHjK6DtHgKUWwk4yIqvZ0-pxHnWtE6KtiutIJy52iVIOUqlzJDlI7Kx2LjYCuDNSMpmAyK3JOXNxYqywsZNKQ2qYMMDZBRkUZQGXhHKFJswoG6NP0zjTaMGQN-Gpmunc5Fdk6Cdjs2uLY2y6ebj--_MNOWlFWR_6cksG9b6BO0T1Wt8Hd34Bn8mjUA
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3PT8IwFG4QD3pSA8bf9uCRkW1tKT0bCSoQDphwI2v7GonJRth28a_3dRsYjQdv2_KSNX1pv69f3w9CHrQcJIxrGYhIu4ALDcGQAeCKd9yndiIqemlgOhuM3_jLUixbpLfPhQGAKvgM-v6xusu3mSm9VIaHd6mQ0R6QQ8E5F3W21l5RQfBUsfCNG5DF4K47RDLSFHXavasfPVQqCBmdkOnu53XkyEe_LHTffP6qy_jf0Z2S7neyHp3vYeiMtCDtkPnEY9TaCwE03wCY9x51WVagTzc5TVJLq8wrmvuuSjktK8NtI-UHxpNG3zsCLLVQ7SVdshg9LR7HQdM8IVirsAh05EIXKeUgljpBmoPUzmrnQiOAOyMlkxGIxJqEMxdGxgobOqk0xIYJA5ydk3aapXBBKFdowoyyIXo1DocaLRgyJzxXM52Y5JJ0_GSsNnV5jFUzD1d_f74nR-PFdLKaPM9er8lxLdH6QJgb0i62Jdwixhf6rnLtF1w0pp0
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2011+10th+International+Conference+on+Information+Processing+in+Sensor+Networks&rft.atitle=Localising+speech%2C+footsteps+and+other+sounds+using+resource-constrained+devices&rft.au=Yukang+Guo&rft.au=Hazas%2C+M&rft.date=2011-04-01&rft.pub=IEEE&rft.isbn=9781612848549&rft.spage=330&rft.epage=341&rft.externalDocID=5779049
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9781612848549/lc.gif&client=summon&freeimage=true
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9781612848549/mc.gif&client=summon&freeimage=true
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9781612848549/sc.gif&client=summon&freeimage=true