• Title/Summary/Keyword: 오디오 핑거프린트

Search Result 17, Processing Time 0.022 seconds

Audio Fingerprinting Based Spatial Audio Reproduction System (오디오 핑거프린팅기반 입체음향 재현 시스템)

  • Ryu, Sang Hyeon;Kim, Hyoung-Gook
    • Journal of the Institute of Electronics and Information Engineers
    • /
    • v.50 no.12
    • /
    • pp.217-223
    • /
    • 2013
  • This paper proposes a spatial audio reproduction system based on audio fingerprinting that combines the audio fingerprinting and the spatial audio processing. In the proposed system, a salient audio peak pair fingerprint based on modulation spectrum improves the accuracy of the audio fingerprinting system in real noisy environments and spatial audio information as metadata gives a listener a sensation of being listening to the sound in the space, where the sound is actually recorded.

Audio Fingerprinting Based on Constant Q Transform for TV Commercial Advertisement Identification (TV 광고 식별을 위한 Constant-Q 변환 기반의 오디오 핑거프린팅 방식)

  • Ryu, Sang Hyeon;Kim, Hyoung-Gook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.33 no.3
    • /
    • pp.210-215
    • /
    • 2014
  • In spite of distortion caused by noise and echo, the audio fingerprinting technique must identify successfully an audio source. This audio fingerprinting technique is applying for TV commercial advertisement identification. In this paper, we propose a robust audio fingerprinting method for TV commercial advertisement identification. In the proposed method, a prominent audio peak pair fingerprint based on constant Q transform improves the accuracy of the audio fingerprinting system in real noisy environments. Experimental results confirm that the proposed method is quite robust than previous audio fingerprinting method in different noise conditions and achieves promising accurate results.

Audio Fingerprinting Using a Robust Hash Function Based on the MCLT Peak-Pair (MCLT 피크쌍 기반의 강인한 해시 함수를 이용한 오디오 핑거프린팅)

  • Lee, Jun-Yong;Kim, Hyoung-Gook
    • The Journal of the Acoustical Society of Korea
    • /
    • v.34 no.2
    • /
    • pp.157-162
    • /
    • 2015
  • In this paper, we propose an audio fingerprinting using robust hash based on the MCLT (Modulated Complex Lapped Transform) peak-pair. In existing methods, the robust audio fingerprinting is not generated if various distortions occurred; time-scaling, pith-shifting and equalization. To solve this problem, we used the spectrum of the MCLT, an adaptive thresholding method for detection of prominent peaks and the novel hash function in the audio fingerprinting. Experimental results show that the proposed method is highly robust in various distorted environments and achieves better identification rates compared to other methods.

Energy and Statistical Filtering for a Robust Audio Fingerprinting System (강인한 오디오 핑거프린팅 시스템을 위한 에너지와 통계적 필터링)

  • Jeong, Byeong-Jun;Kim, Dae-Jin
    • The Journal of the Korea Contents Association
    • /
    • v.12 no.5
    • /
    • pp.1-9
    • /
    • 2012
  • The popularity of digital music and smart phones led to develope noise-robust real-time audio fingerprinting system in various ways. In particular, The Multiple Hashing(MLH) of fingerprint algorithms is robust to noise and has an elaborate structure. In this paper, we propose a filter engine based on MLH to achieve better performance. In this approach, we compose a energy-intensive filter to improve the accuracy of Q/R from music database and a statistic filter to remove continuity and redundancy. The energy-intensive filter uses the Discrite Cosine Transform(DCT)'s feature gathering energy to low-order bits and the statistic filters use the correlation between searched fingerprint's information. Experimental results show that the superiority of proposed algorithm consists of the energy and statistical filtering in noise environment. It is found that the proposed filter engine achieves more robust to noise than Philips Robust Hash(PRH), and a more compact way than MLH.

Trends in Acoustic Data Hiding and Tagging Technologies for Enhancement of Media Accessibility (미디어 접근편의성 향상을 위한 음향 데이터 삽입 및 색인 기술 동향)

  • Sung, J.;Beack, S.;Lee, M.;Lee, T.
    • Electronics and Telecommunications Trends
    • /
    • v.32 no.3
    • /
    • pp.36-45
    • /
    • 2017
  • 오디오 워터마크, 음향 데이터 전송 및 오디오 핑거프린트 등으로 대표되는 음향 데이터 삽입 및 색인 기술은 최근 다양한 미디어 활용 인프라의 보급과 새로운 형태의 미디어 생태계가 등장함에 따라 중요성이 더욱 커지고 있으며, 콘텐츠 제어 및 식별을 비롯한 다양한 응용 서비스의 기반 기술로 활용될 수 있다. 본고에서는 음향 신호 기반 데이터 삽입 및 색인 기술 개발 현황과 관련 서비스 동향에 대해서 소개한다.

Music Search Algorithm for Automotive Infotainment System (자동차 환경의 인포테인먼트 시스템을 위한 음악 검색 알고리즘)

  • Kim, Hyoung-Gook;Kim, Jae-Man
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.12 no.1
    • /
    • pp.81-87
    • /
    • 2013
  • In this paper, we propose a music search algorithm for automotive infotainment system. The proposed method extracts fingerprints using the high peaks based on log-spectrum of the music signal, and the extracted music fingerprints store in cloud server applying a hash value. In the cloud server, the most similar music is retrieved by comparing the user's query music with the fingerprints stored in hash table of cloud server. To evaluate the performance of the proposed music search algorithm, we measure an accuracy of the retrieved results according to various length of the query music and measure a retrieval time according to the number of stored music database in hash table.

Development of Music Recommendation System based on Customer Sentiment Analysis (소비자 감성 분석 기반의 음악 추천 알고리즘 개발)

  • Lee, Seung Jun;Seo, Bong-Goon;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.197-217
    • /
    • 2018
  • Music is one of the most creative act that can express human sentiment with sound. Also, since music invoke people's sentiment to get empathized with it easily, it can either encourage or discourage people's sentiment with music what they are listening. Thus, sentiment is the primary factor when it comes to searching or recommending music to people. Regard to the music recommendation system, there are still lack of recommendation systems that are based on customer sentiment. An algorithm's that were used in previous music recommendation systems are mostly user based, for example, user's play history and playlists etc. Based on play history or playlists between multiple users, distance between music were calculated refer to basic information such as genre, singer, beat etc. It can filter out similar music to the users as a recommendation system. However those methodology have limitations like filter bubble. For example, if user listen to rock music only, it would be hard to get hip-hop or R&B music which have similar sentiment as a recommendation. In this study, we have focused on sentiment of music itself, and finally developed methodology of defining new index for music recommendation system. Concretely, we are proposing "SWEMS" index and using this index, we also extracted "Sentiment Pattern" for each music which was used for this research. Using this "SWEMS" index and "Sentiment Pattern", we expect that it can be used for a variety of purposes not only the music recommendation system but also as an algorithm which used for buildup predicting model etc. In this study, we had to develop the music recommendation system based on emotional adjectives which people generally feel when they listening to music. For that reason, it was necessary to collect a large amount of emotional adjectives as we can. Emotional adjectives were collected via previous study which is related to them. Also more emotional adjectives has collected via social metrics and qualitative interview. Finally, we could collect 134 individual adjectives. Through several steps, the collected adjectives were selected as the final 60 adjectives. Based on the final adjectives, music survey has taken as each item to evaluated the sentiment of a song. Surveys were taken by expert panels who like to listen to music. During the survey, all survey questions were based on emotional adjectives, no other information were collected. The music which evaluated from the previous step is divided into popular and unpopular songs, and the most relevant variables were derived from the popularity of music. The derived variables were reclassified through factor analysis and assigned a weight to the adjectives which belongs to the factor. We define the extracted factors as "SWEMS" index, which describes sentiment score of music in numeric value. In this study, we attempted to apply Case Based Reasoning method to implement an algorithm. Compare to other methodology, we used Case Based Reasoning because it shows similar problem solving method as what human do. Using "SWEMS" index of each music, an algorithm will be implemented based on the Euclidean distance to recommend a song similar to the emotion value which given by the factor for each music. Also, using "SWEMS" index, we can also draw "Sentiment Pattern" for each song. In this study, we found that the song which gives a similar emotion shows similar "Sentiment Pattern" each other. Through "Sentiment Pattern", we could also suggest a new group of music, which is different from the previous format of genre. This research would help people to quantify qualitative data. Also the algorithms can be used to quantify the content itself, which would help users to search the similar content more quickly.