DOI QR코드

DOI QR Code

Acoustic Monitoring and Localization for Social Care

  • Goetze, Stefan (Fraunhofer Institute for Digital Media Technology, Project group Hearing, Speech and Audio Technology) ;
  • Schroder, Jens (Fraunhofer Institute for Digital Media Technology, Project group Hearing, Speech and Audio Technology) ;
  • Gerlach, Stephan (Fraunhofer Institute for Digital Media Technology, Project group Hearing, Speech and Audio Technology) ;
  • Hollosi, Danilo (Fraunhofer Institute for Digital Media Technology, Project group Hearing, Speech and Audio Technology) ;
  • Appell, Jens-E. (Fraunhofer Institute for Digital Media Technology, Project group Hearing, Speech and Audio Technology) ;
  • Wallhoff, Frank (Fraunhofer Institute for Digital Media Technology, Project group Hearing, Speech and Audio Technology, and Jade-Hochschule)
  • Received : 2011.07.18
  • Accepted : 2012.02.06
  • Published : 2012.03.30

Abstract

Increase in the number of older people due to demographic changes poses great challenges to the social healthcare systems both in the Western and as well as in the Eastern countries. Support for older people by formal care givers leads to enormous temporal and personal efforts. Therefore, one of the most important goals is to increase the efficiency and effectiveness of today's care. This can be achieved by the use of assistive technologies. These technologies are able to increase the safety of patients or to reduce the time needed for tasks that do not relate to direct interaction between the care giver and the patient. Motivated by this goal, this contribution focuses on applications of acoustic technologies to support users and care givers in ambient assisted living (AAL) scenarios. Acoustic sensors are small, unobtrusive and can be added to already existing care or living environments easily. The information gathered by the acoustic sensors can be analyzed to calculate the position of the user by localization and the context by detection and classification of acoustic events in the captured acoustic signal. By doing this, possibly dangerous situations like falls, screams or an increased amount of coughs can be detected and appropriate actions can be initialized by an intelligent autonomous system for the acoustic monitoring of older persons. The proposed system is able to reduce the false alarm rate compared to other existing and commercially available approaches that basically rely only on the acoustic level. This is due to the fact that it explicitly distinguishes between the various acoustic events and provides information on the type of emergency that has taken place. Furthermore, the position of the acoustic event can be determined as contextual information by the system that uses only the acoustic signal. By this, the position of the user is known even if she or he does not wear a localization device such as a radio-frequency identification (RFID) tag.

Keywords

References

  1. Commission of the European Communities, "Europe's demographic future: facts and figures," Staff Working Document SEC (2007) 638, Brussels: Commission of the European Communities, 2007.
  2. Statistisches Bundesamt [Federal Statistical Office], "Demografischer wandel in Deutschland: auswirkungen auf krankenhausbehandlungen und pflegebedürftige im bund und in den landern [Demographic change in Germany: impacts on hospital treatments and people in need of care]," 2008, Available: https://www.statistik.bayern.de/veroeffentlichungen/ download/A1832E%20200851/A1832E%20200851.pdf. German.
  3. R. C. Petersen, "Mild cognitive impairment as a diagnostic entity," Journal of Internal Medicine, vol. 256, no. 3, pp. 183-194, 2004. https://doi.org/10.1111/j.1365-2796.2004.01388.x
  4. M. A. Rudberg, S. E. Furner, J. E. Dunn, and C. K. Cassel, "The relationship of visual and hearing impairments to disability: an analysis using the longitudinal study of aging," Journal of Gerontology, vol. 48, no. 6, pp. M261-M265, 1993. https://doi.org/10.1093/geronj/48.6.M261
  5. S. Uimonen, K. Huttunen., K. Jounio-Ervasti, and M. Sorri, "Do we know the real need for hearing rehabilitation at the population level? Hearing impairments in the 5- to 75-yearold cross-sectional Finnish population," British Journal of Audiology, vol. 33, no. 1, pp. 53-59, 1999. https://doi.org/10.3109/03005364000000099
  6. S. Goetze, F. Xiong, J. Rennies, T. Rohdenburg, and J. E. Appell, "Hands-free telecommunication for elderly persons suffering from hearing deficiencies," 12th IEEE International Conference on E-Health Networking, Application and Services, Lyon, France, 2010, pp. 209-224.
  7. G. van den Broek, F. Cavallo, and C. Wehrmann, AALIANCE Ambient Assisted Living Roadmap, Amsterdam, Netherlands: IOS Press, 2010.
  8. S. Boll, W. Heuten, E. M. Meyer, and M. Meis, "Development of a multimodal reminder system for older persons in their residential home," Informatics for Health and Social Care, vol. 35, no. 3-4, pp. 104-124, 2010. https://doi.org/10.3109/17538157.2010.528651
  9. C. Lisetti, F. Nasoz, C. LeRouge, O. Ozyer, and K. Alvarez, "Developing multimodal intelligent affective interfaces for tele-home health care," International Journal of Human- Computer Studies, vol. 59, no. 1-2, pp. 245-255, 2003. https://doi.org/10.1016/S1071-5819(03)00051-X
  10. S. Chew, W. Tay, D. Smit, and C. Bartneck, "Do social robots walk or roll?," Proceedings of the Second International Conference on Social Robotics, Singapore, 2010, pp. 355-361.
  11. J. Rennies, S. Goetze, and J. E. Appell, "Personalized acoustic interfaces for human-computer interaction," Human-Centered Design of E-Health Technologies: Concepts, Methods and Applications, Hershey, PA: IGI Global, 2011. pp. 180-207.
  12. S. Goetze, N. Moritz, J. E. Appell, M. Meis, C. Bartsch, and J. Bitzer, "Acoustic user interfaces for ambient-assisted living technologies," Informatics for Health and Social Care, vol. 35, no. 3-4, pp. 125-143, 2010. https://doi.org/10.3109/17538157.2010.528655
  13. E. Hansler and G. Schmidt, Speech and Audio Processing in Adverse Environments, Heidelberg, Germany: Springer, 2008.
  14. P. W. J. van Hengel and J. Anemuller, "Audio event detection for in-home care," NAG/DAGA International Conference on Acoustics, Rotterdam, Amsterdam, 2009, pp. 618-620.
  15. P. W. J. van Hengel, M. Huisman, and J. E. Appell, "Sounds like trouble," Human Factors Security and Safety, Maastricht, Netherlands: Shaker Publishing, 2009, pp. 369-375.
  16. D. Hollosi, J. Schroder, S. Goetze, and J. E. Appell, "Voice activity detection driven acoustic event classification for monitoring in smart homes," 3rd International Symposium on Applied Sciences in Biomedical and Communication Technologies, Rome, Italy, 2010, pp. 1-5.
  17. J. Schroder, S. Wabnik, P. W. J. van Hengel, and S. Goetze, "Detection and classification of acoustic events for in-home care," Ambient Assisted Living, Heidelberg, Germany: Springer, 2011, pp. 181-196.
  18. B. Pfister and T. Kaufmann, "Sprachverarbeitung: grundlagen und methoden der sprachsynthese und spracherkennung," Heidelberg, Germany: Springer, 2008.
  19. C. H. Knapp and G. C. Carter, "The generalized correlation method for estimation of time delay," IEEE Transactions on Acoustics, Speech and Signal Processing, vol. 24, no. 4, pp. 320-327, 1976. https://doi.org/10.1109/TASSP.1976.1162830
  20. G. Doblinger, "Localization and tracking of acoustical sources," Topics in Acoustic Echo and Noise Control, Heidelberg, Germany: Springer, 2006, pp. 91-122.
  21. J. Bitzer and K. U. Simmer, "Superdirective microphone arrays," Microphone Arrays: Signal Processing Techniques and Applications, Heidelberg, Germany: Springer, 2001, pp. 19-38.
  22. J. Ramirez, J. C. Segura, C. Benitez, A. de la Torre, and A. Rubio, "Efficient voice activity detection algorithms using long-term speech information," Speech Communication, vol. 42, pp. 3-4, pp. 271-287, 2004. https://doi.org/10.1016/j.specom.2003.10.002
  23. N. Cho and E Kim, "Enhanced voice activity detection using acoustic event detection and classification," IEEE Transactions on Consumer Electronics, vol. 57, no. 1, pp. 196-202, 2011. https://doi.org/10.1109/TCE.2011.5735502
  24. B. T. Meyer, T. Brand, and B. Kollmeier, "Effect of speechintrinsic variations on human and automatic recognition of spoken phonemes," Journal of the Acoustical Society of America, vol. 129, no. 1, pp. 388-403, 2011. https://doi.org/10.1121/1.3514525
  25. D. Macho, L. Mauuary, B. Noe, Y. M. Cheng, D. Ealey, D. Jouvet, H. Kelleher, D. Pearce, and F. Saadoun, "Evaluation of a noise-robust DSR front-end on aurora databases," 7th International Conference on Spoken Language Processing, Denver, CA, 2002, pp. 17-22.
  26. V. Hohmann, "Frequency analysis and synthesis using a Gammatone filterbank," Acta Acustica united with Acoustica, vol. 88, no. 3, pp. 433-442, 2002.
  27. Y. Amit, A. Koloydenko, and P. Niyogi, "Robust acoustic object detection," Journal of the Acoustical Society of America, vol. 118, no. 4, pp. 2634-2648, 2005. https://doi.org/10.1121/1.2011411
  28. R.O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, 2nd ed. New York, NY: Wiley, 2001.
  29. C. M. Bishop, Pattern Recognition and Machine Learning, Heidelberg, Germany: Springer, 2006.
  30. D. Arthur and S. Vassilvitskii, "k-means++: the advantages of careful seeding," Stanford InfoLab, Stanford, CA, Technical Report 2006-13, 2006.
  31. J. Chen, J. Benesty, and Y. A. Huang, "Time delay estimation in room acoustic environments: an overview," EURASIP Journal on Applied Signal Processing, vol. 2006, pp. 1-19, 2006. https://doi.org/10.1155/ASP/2006/96421
  32. A. Savard, "Overview of homophonic pitch detection algorithms," Schulich School of Music, McGill University, Montreal, Technical Report MUMT-612, 2006.
  33. A. M. Noll, "Cepstrum pitch determination," Journal of the Acoustical Society of America, vol. 41, no. 2, pp. 293-309, 1967. https://doi.org/10.1121/1.1910339
  34. A. V. Oppenheim and R. W. Schafer, "From frequency to quefrency: a history of the cepstrum," IEEE Signal Processing Magazine, vol. 21, no. 5, pp. 95-106, 2004. https://doi.org/10.1109/MSP.2004.1328092
  35. P. Vary and R. Martin, Digital Speech Transmission: Enhancement, Coding and Error Concealment, Chichester, UK: John Wiley & Sons Ltd, 2006.
  36. M. Kepesi, F. Pernkopf, and M. Wohlmayr, "Joint positionpitch tracking for 2-channel audio," 5th International Workshop on Content-Based Multimedia Indexing, Bordeaux, France, 2007, pp. 303-306.

Cited by

  1. Metrics for Polyphonic Sound Event Detection vol.6, pp.6, 2016, https://doi.org/10.3390/app6060162
  2. Self-improvement of voice interface with user-input spoken query at early stage of commercialization vol.59, pp.4, 2013, https://doi.org/10.1109/TCE.2013.6689699
  3. Convolutional Recurrent Neural Networks for Polyphonic Sound Event Detection vol.25, pp.6, 2017, https://doi.org/10.1109/TASLP.2017.2690575
  4. An integrated system for voice command recognition and emergency detection based on audio signals vol.42, pp.13, 2015, https://doi.org/10.1016/j.eswa.2015.02.036