Fig. 1. Architecture of the proposed system.
Fig. 2. Framework of the training and testing procedure for the proposed SED system.
Fig. 3. Diagram of a process for generating haptic vibration.
Fig. 4. Results of sound-to-haptic conversion using harmonic-percussive source separation.
Table 1. Comparison of the segment-based error rate and F-score for different combinations of classifiers and features.
References
- M. Zöhrer and F. Pernkopf, "Gated recurrent networks applied to acoustic scene classification and acoustic event detection," Proc. Detection and Classification of Acoustic Scenes and Events 2016, 1-5 (2016).
- B. H. Kim, H.-G. Kim, J. Jeong, and J. Y. Kim, "VoIP receiver-based adaptive playout scheduling and packet loss concealment technique," IEEE Trans. Consum. Electron., 59, 250-258 (2013). https://doi.org/10.1109/TCE.2013.6490267
- K. Kumatani, J. McDonough, J. F. Lehman, and B. Raj, "Channel selection based on multichannel crosscorrelation coefficients for distant speech recognition," Proc. Joint Workshop Hands-free Speech Commun. Microphone Arrays, 1-6 (2011).
- D. Pavlidi, A. Griffin, M. Puigt, and A. Mouchtaris, "Real-time multiple sound source localization and counting using a circular microphone array," IEEE Trans. Audio, Speech, Lang. Process., 21, 2193-2206 (2013). https://doi.org/10.1109/TASL.2013.2272524
- R. Lu and Z. Duan, "Bidirectional GRU for sound event detection," Proc. Detection and Classification of Acoustic Scenes and Events 2017, 1-4 (2017).
- A. Mesaros, T. Heittola, and T. Virtanen, "TUT database for acoustic scene classification and sound event detection," Proc. 24th Eur. Signal Process. Conf., 1128-1132 (2016).
- A. Liutkus, D. Fitzgerald, Z. Rafii, B. Pardo, and L. Daudet, "Kernel additive models for source separation," IEEE Trans. Signal Process., 62, 4298-4310 (2014). https://doi.org/10.1109/TSP.2014.2332434