• Title/Summary/Keyword: 선별 알고리즘

Search Result 292, Processing Time 0.026 seconds

Study on Target Advertisement Technology using a method of reasoning TV Viewer's Profiles (시청자 프로파일 추론 기법을 이용한 표적 광고 기술 연구)

  • Kim M;Lim J;Kang S;Kim M;Kang K
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2004.11a
    • /
    • pp.147-150
    • /
    • 2004
  • 기존의 방송환경은 시청자에게 단 방향의 서비스를 제공하며, 시청자 각자의 취향에 상관없이 소극적인 자세를 요구한다. 이러한 일방적인 사용자 시청환경에서는 대개의 프로그램들은 방송사에 의해 다수 사용자의 대표적인 시간대에 적절하게 배치되고, 적절하게 배분된 프로그램의 인기도 혹은 시청률, 시청 연령, 시간대에 따른 광고가 그 프로그램을 지원하게 된다. 시청자의 적극적 선별 노력에 의해 선택된 프로그램과 달리 방송 프로그램 콘텐츠와 같이 전송되는 광고는 사용자가 원하는 대개의 프로그램을 지원하는 중요한 수단이지만, 정작 시청자의 관심과는 상관없이 제공된다. 또한, 디지털 방송의 시작과 함께 양방향 서비스 환경에서도 사용자를 고려한 광고의 적절한 분배가 이루어지지 않고 있다. 이러한 일률적인 광고 콘텐츠의 제공은 시청자의 무관심을 초래할 수 있으며, 이러한 경향이 높을수록 효율적인 정보 제공에 많은 제한을 받을 수 있다. 본 논문에서는, 제한적 정보 제공의 해소와 새로운 시청 환경 제안을 위한, 시청자의 성별 및 연령 대를 추론하는 알고리즘을 제안하고 이를 맞춤형 광고, 즉 표적 광고 서비스(Target Advertisement Service) 기법을 제안한다.

  • PDF

Web Document Prediction System by using Web Log Mining (웹 로그 마이닝을 이용한 웹 문서 예측 시스템)

  • Lee Bum-suk;Hwang Byung-yeon
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2005.11b
    • /
    • pp.97-99
    • /
    • 2005
  • 웹 문서 수의 급격한 증가는 사용자로 하여금 방대한 양의 웹 문서들로부터 필요한 정보를 선별하기 위한 시간과 비용을 낭비하게 만들었다. 따라서 이러한 문제를 해결하기 위한 연구의 필요성이 점차 증가하였는데, 그 중 웹 서버 로그 데이터에 마이닝 기법을 적용하여 사용자들의 사이트 내 문서의 접근 패턴을 분석하고, 그 데이터를 이용하여 동적으로 변화하는 적응형 웹 사이트를 제공하려는 것이 대표적인 연구 사례이다. 본 논문에서는 웹 서버 로그 마이닝을 이용하여 사용자가 필요로 하거나, 관심을 가지고 있는 페이지를 예측하여 추천해 주는 시스템에 대해 소개한다. 이러한 시스템을 구현하기 위해 순차 패턴 마이닝이나 빈발 에피소드 발견 기법 등의 알고리즘을 사용할 수 있다. 제안하는 시스템에서는 사용자 접근 패턴을 분석할 때 순차 패턴 마이닝 기법을 사용하고, 사용자의 이동 패턴을 근거로 웹 문서를 예측하여 추천해줄 때에는 에피소드 발견 기법에서의 window 개념을 이용한다. 본 논문에서 제안한 시스템은 웹 문서를 사용자가 머물었던 시간에 따라 관심 있는 문서와 지나간 문서로 구분하여 관심 있는 문서에 대해서안 마이닝을 수행한다. 또한 일정한 크기를 갖는 History window에 의해 다음 문서를 추천해주기 때문에 사용자의 모든 로그를 저장하지 않으므로 보다 효율적이다.

  • PDF

A Study on improvement of performance of collaborative filtering recommendation system using social data (소셜 데이터를 이용한 협업필터링 추천 시스템 성능 개선 연구)

  • Joo, Jong-Min;Yang, Hyung-Jeong;Kim, Nam-Hun;Park, Sung-Hyun;Lee, Gun-Woo
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2017.11a
    • /
    • pp.660-663
    • /
    • 2017
  • 다양한 소셜 네트워크 서비스가 발달되고 많은 사람들이 소셜 미디어에 참여하면서 방대한 양의 정보가 발생하고 있다. 따라서 원하는 정보를 선별하고 가공하는 연구도 활발히 진행되고 있다. 협업필터링은 이러한 정보를 토대로 사용자에게 맞춤형 아이템을 추천해주는 알고리즘이다. 하지만 정확한 추천을 위해서는 매우 방대한 양의 정보가 필요하다. 또한 협업필터링에는 초기에는 제대로 추천이 이루어지지 않는 콜드스타터 문제가 있다. 이러한 문제를 해결하기 위해 본 논문에서는 소셜 네트워크 서비스 중의 하나인 트위터 데이터를 활용하여 협업필터링 추천 시스템의 성능을 높이고자 한다. 협업필터링의 평점에 특정 아이템 관련 트윗을 수집해서 긍정/부정을 측정하여 가중치를 부여한다. RMSE 평가 방법을 통한 실험 결과, 소셜 미디어의 긍부정 영향력을 측정하여 적용했을 때가 기존의 협업필터링 방식에 비해 약 5.5%의 성능 향상을 확인하였다.

A Study on Keywords Extraction based on Semantic Analysis of Document (문서의 의미론적 분석에 기반한 키워드 추출에 관한 연구)

  • Song, Min-Kyu;Bae, Il-Ju;Lee, Soo-Hong;Park, Ji-Hyung
    • Proceedings of the Korea Inteligent Information System Society Conference
    • /
    • 2007.11a
    • /
    • pp.586-591
    • /
    • 2007
  • 지식 관리 시스템, 정보 검색 시스템, 그리고 전자 도서관 시스템 등의 문서를 다루는 시스템에서는 문서의 구조화 및 문서의 저장이 필요하다. 문서에 담겨있는 정보를 추출하기 위해 가장 우선시되어야 하는 것은 키워드의 선별이다. 기존 연구에서 가장 널리 사용된 알고리즘은 단어의 사용 빈도를 체크하는 TF(Term Frequency)와 IDF(Inverted Document Frequency)를 활용하는 TF-IDF 방법이다. 그러나 TF-IDF 방법은 문서의 의미를 반영하지 못하는 한계가 존재한다. 이를 보완하기 위하여 본 연구에서는 세 가지 방법을 활용한다. 첫 번째는 문헌 속에서의 단어의 위치 및 서론, 결론 등의 특정 부분에 사용된 단어의 활용도를 체크하는 문헌구조적 기법이고, 두 번째는 강조 표현, 비교 표현 등의 특정 사용 문구를 통제 어휘로 지정하여 활용하는 방법이다. 마지막으로 어휘의 사전적 의미를 분석하여 이를 메타데이터로 활용하는 방법인 언어학적 기법이 해당된다. 이를 통하여 키워드 추출 과정에서 문서의 의미 분석도 수행하여 키워드 추출의 효율을 높일 수 있다.

  • PDF

A Study of Machine Learning Model for Prediction of Swelling Waves Occurrence on East Sea (동해안 너울성 파도 예측을 위한 머신러닝 모델 연구)

  • Kang, Donghoon;Oh, Sejong
    • The Journal of Korean Institute of Information Technology
    • /
    • v.17 no.9
    • /
    • pp.11-17
    • /
    • 2019
  • In recent years, damage and loss of life and property have been occurred frequently due to swelling waves in the East Sea. Swelling waves are not easy to predict because they are caused by various factors. In this research, we build a model for predicting the swelling waves occurrence in the East Coast of Korea using machine learning technique. We collect historical data of unloading interruption in the Pohang Port, and collect air pressure, wind speed, direction, water temperature data of the offshore Pohang Port. We select important variables for prediction, and test various machine learning prediction algorithms. As a result, tide level, water temperature, and air pressure were selected, and Random Forest model produced best performance. We confirm that Random Forest model shows best performance and it produces 88.86% of accuracy

Corporate Default Prediction Model Using Deep Learning Time Series Algorithm, RNN and LSTM (딥러닝 시계열 알고리즘 적용한 기업부도예측모형 유용성 검증)

  • Cha, Sungjae;Kang, Jungseok
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.1-32
    • /
    • 2018
  • In addition to stakeholders including managers, employees, creditors, and investors of bankrupt companies, corporate defaults have a ripple effect on the local and national economy. Before the Asian financial crisis, the Korean government only analyzed SMEs and tried to improve the forecasting power of a default prediction model, rather than developing various corporate default models. As a result, even large corporations called 'chaebol enterprises' become bankrupt. Even after that, the analysis of past corporate defaults has been focused on specific variables, and when the government restructured immediately after the global financial crisis, they only focused on certain main variables such as 'debt ratio'. A multifaceted study of corporate default prediction models is essential to ensure diverse interests, to avoid situations like the 'Lehman Brothers Case' of the global financial crisis, to avoid total collapse in a single moment. The key variables used in corporate defaults vary over time. This is confirmed by Beaver (1967, 1968) and Altman's (1968) analysis that Deakins'(1972) study shows that the major factors affecting corporate failure have changed. In Grice's (2001) study, the importance of predictive variables was also found through Zmijewski's (1984) and Ohlson's (1980) models. However, the studies that have been carried out in the past use static models. Most of them do not consider the changes that occur in the course of time. Therefore, in order to construct consistent prediction models, it is necessary to compensate the time-dependent bias by means of a time series analysis algorithm reflecting dynamic change. Based on the global financial crisis, which has had a significant impact on Korea, this study is conducted using 10 years of annual corporate data from 2000 to 2009. Data are divided into training data, validation data, and test data respectively, and are divided into 7, 2, and 1 years respectively. In order to construct a consistent bankruptcy model in the flow of time change, we first train a time series deep learning algorithm model using the data before the financial crisis (2000~2006). The parameter tuning of the existing model and the deep learning time series algorithm is conducted with validation data including the financial crisis period (2007~2008). As a result, we construct a model that shows similar pattern to the results of the learning data and shows excellent prediction power. After that, each bankruptcy prediction model is restructured by integrating the learning data and validation data again (2000 ~ 2008), applying the optimal parameters as in the previous validation. Finally, each corporate default prediction model is evaluated and compared using test data (2009) based on the trained models over nine years. Then, the usefulness of the corporate default prediction model based on the deep learning time series algorithm is proved. In addition, by adding the Lasso regression analysis to the existing methods (multiple discriminant analysis, logit model) which select the variables, it is proved that the deep learning time series algorithm model based on the three bundles of variables is useful for robust corporate default prediction. The definition of bankruptcy used is the same as that of Lee (2015). Independent variables include financial information such as financial ratios used in previous studies. Multivariate discriminant analysis, logit model, and Lasso regression model are used to select the optimal variable group. The influence of the Multivariate discriminant analysis model proposed by Altman (1968), the Logit model proposed by Ohlson (1980), the non-time series machine learning algorithms, and the deep learning time series algorithms are compared. In the case of corporate data, there are limitations of 'nonlinear variables', 'multi-collinearity' of variables, and 'lack of data'. While the logit model is nonlinear, the Lasso regression model solves the multi-collinearity problem, and the deep learning time series algorithm using the variable data generation method complements the lack of data. Big Data Technology, a leading technology in the future, is moving from simple human analysis, to automated AI analysis, and finally towards future intertwined AI applications. Although the study of the corporate default prediction model using the time series algorithm is still in its early stages, deep learning algorithm is much faster than regression analysis at corporate default prediction modeling. Also, it is more effective on prediction power. Through the Fourth Industrial Revolution, the current government and other overseas governments are working hard to integrate the system in everyday life of their nation and society. Yet the field of deep learning time series research for the financial industry is still insufficient. This is an initial study on deep learning time series algorithm analysis of corporate defaults. Therefore it is hoped that it will be used as a comparative analysis data for non-specialists who start a study combining financial data and deep learning time series algorithm.

A Study on the Effect of On-Dock System in Container Terminals - Focusing on GwangYang Port - (컨테이너터미널에서 On-Dock 시스템 효과분석에 관한 연구 - 광양항을 중심으로 -)

  • Cha, Sang-Hyun;Noh, Chang-Kyun
    • Journal of Navigation and Port Research
    • /
    • v.39 no.1
    • /
    • pp.45-53
    • /
    • 2015
  • These days Container Terminals are focusing on increasing the quantity of containers and shipping lines choose Terminals by referring to the key elements of a terminal to perform the overall operation the fastest such as the location of the terminal, discharging ability, keeping environment, and other elements related to shipping in general. Container terminal is able to offer On-Dock service has become an important factor for shipping lines to choose that terminal. In this paper, we propose an algorithm for On-Dock system work algorithm, the algorithm Empty container exports, Full Container algorithm and The aim of our study focus on both container's gate out time and search for the effective terminal operation which is using the general On-Dock system through several algorithm like container batch priority, gate in and out job priority and empty container yard equipment allocation rule based on the automatic allocation method and manual allocation scheme for container. Gathering these information, it gives the priority and yard location of gate-out containers to control. That is, by selecting an optimum algorithm container, container terminals Empty reduces the container taken out time, it is possible to minimize unnecessary re-handling of the yard container can be enhanced with respect to the efficiency of the equipment. Operations and operating results of the Non On-Dock and On-Dock system is operated by the out work operations (scenarios) forms that are operating in the real Gwangyang Container Terminal derived results. Gwangyang Container terminal and apply the On-Dock system, Non On-Dock can be taken out this time, about 5 minutes more quickly when applying the system. when managing export orders for berths where On-Dock service is needed, ball containers are allocated and for import cargoes, D/O is managed and after carryout, return management, container damage, cleaning, fixing and controlling services are supported hence the berth service can be strengthened and container terminal business can grow.

A Real-time Dual-mode Temporal Synchronization and Compensation based on Reliability Measure in Stereoscopic Video (3D 입체 영상 시스템에서 신뢰도를 활용한 듀얼 모드 실시간 동기 에러 검출 및 보상 방법)

  • Kim, Giseok;Cho, Jae-Soo;Lee, Gwangsoon;Lee, Eung-Don
    • Journal of Broadcast Engineering
    • /
    • v.19 no.6
    • /
    • pp.896-906
    • /
    • 2014
  • In this paper, a real-time dual-mode temporal synchronization and compensation method based on a new reliability measure in stereoscopic video is proposed. The goal of temporal alignment is to detect the temporal asynchrony and recover synchronization of the two video streams. The accuracy of the temporal synchronization algorithm depends on the 3DTV contents. In order to compensate the temporal synchronization error, it is necessary to judge whether the result of the temporal synchronization is reliable or not. Based on our recently developed temporal synchronization method[1], we define a new reliability measure for the result of the temporal synchronization method. Furthermore, we developed a dual-mode temporal synchronization method, which uses a usual texture matching method and the temporal spatiogram method[1]. The new reliability measure is based on two distinctive features, a dynamic feature for scene change and a matching distinction feature. Various experimental results show the effectiveness of the proposed method. The proposed algorithms are evaluated and verified through an experimental system implemented for 3DTV.

Sensor Fault Detection Scheme based on Deep Learning and Support Vector Machine (딥 러닝 및 서포트 벡터 머신기반 센서 고장 검출 기법)

  • Yang, Jae-Wan;Lee, Young-Doo;Koo, In-Soo
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.18 no.2
    • /
    • pp.185-195
    • /
    • 2018
  • As machines have been automated in the field of industries in recent years, it is a paramount importance to manage and maintain the automation machines. When a fault occurs in sensors attached to the machine, the machine may malfunction and further, a huge damage will be caused in the process line. To prevent the situation, the fault of sensors should be monitored, diagnosed and classified in a proper way. In the paper, we propose a sensor fault detection scheme based on SVM and CNN to detect and classify typical sensor errors such as erratic, drift, hard-over, spike, and stuck faults. Time-domain statistical features are utilized for the learning and testing in the proposed scheme, and the genetic algorithm is utilized to select the subset of optimal features. To classify multiple sensor faults, a multi-layer SVM is utilized, and ensemble technique is used for CNN. As a result, the SVM that utilizes a subset of features selected by the genetic algorithm provides better performance than the SVM that utilizes all the features. However, the performance of CNN is superior to that of the SVM.

Automated algorithm of automated auditory brainstem response for neonates (신생아 청성뇌간 반응의 자동 판독 알고리즘)

  • Jung, Won-Hyuk;Hong, Hyun-Ki;Nam, Ki-Chang;Cha, Eun-Jong;Kim, Deok-Won
    • Journal of the Institute of Electronics Engineers of Korea SC
    • /
    • v.44 no.1
    • /
    • pp.100-107
    • /
    • 2007
  • AABR(automated auditory brainstem response) test is used for the screening purpose of hearing ability of neonates. In this paper, algorithm using Rolle's theorem is suggested for automatic detection of the ensemble averaged ABR waveform. The ABR waveforms were recorded from 55 normal-hearing ears of neonates at screening levels varying from 30 to 60 dBnHL. Recorded signals were analyzed by expert audiologist and by the proposed algorithm. The results showed that the proposed algorithm correctly identified latencies of the major ABR waves (III, V) with latent difference below 0.2 ms. No significant differences were found between the two methods. We also analyzed the ABR signals using derivative algorithm and compared the results with proposed algorithm. The number of detected candidate waves using the proposed algorithm was 47 % less than that of the existing one. The proposed method had lower relative errors (0.01 % error at 60dBnHL) compared to the existing one. By using proposed algorithm, clinicians can detect and label waves III and V more objectively and quantitatively than the manual detection method.