• Title/Summary/Keyword: Content Based Filtering

Search Result 227, Processing Time 0.022 seconds

An Expert System for the Estimation of the Growth Curve Parameters of New Markets (신규시장 성장모형의 모수 추정을 위한 전문가 시스템)

  • Lee, Dongwon;Jung, Yeojin;Jung, Jaekwon;Park, Dohyung
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.17-35
    • /
    • 2015
  • Demand forecasting is the activity of estimating the quantity of a product or service that consumers will purchase for a certain period of time. Developing precise forecasting models are considered important since corporates can make strategic decisions on new markets based on future demand estimated by the models. Many studies have developed market growth curve models, such as Bass, Logistic, Gompertz models, which estimate future demand when a market is in its early stage. Among the models, Bass model, which explains the demand from two types of adopters, innovators and imitators, has been widely used in forecasting. Such models require sufficient demand observations to ensure qualified results. In the beginning of a new market, however, observations are not sufficient for the models to precisely estimate the market's future demand. For this reason, as an alternative, demands guessed from those of most adjacent markets are often used as references in such cases. Reference markets can be those whose products are developed with the same categorical technologies. A market's demand may be expected to have the similar pattern with that of a reference market in case the adoption pattern of a product in the market is determined mainly by the technology related to the product. However, such processes may not always ensure pleasing results because the similarity between markets depends on intuition and/or experience. There are two major drawbacks that human experts cannot effectively handle in this approach. One is the abundance of candidate reference markets to consider, and the other is the difficulty in calculating the similarity between markets. First, there can be too many markets to consider in selecting reference markets. Mostly, markets in the same category in an industrial hierarchy can be reference markets because they are usually based on the similar technologies. However, markets can be classified into different categories even if they are based on the same generic technologies. Therefore, markets in other categories also need to be considered as potential candidates. Next, even domain experts cannot consistently calculate the similarity between markets with their own qualitative standards. The inconsistency implies missing adjacent reference markets, which may lead to the imprecise estimation of future demand. Even though there are no missing reference markets, the new market's parameters can be hardly estimated from the reference markets without quantitative standards. For this reason, this study proposes a case-based expert system that helps experts overcome the drawbacks in discovering referential markets. First, this study proposes the use of Euclidean distance measure to calculate the similarity between markets. Based on their similarities, markets are grouped into clusters. Then, missing markets with the characteristics of the cluster are searched for. Potential candidate reference markets are extracted and recommended to users. After the iteration of these steps, definite reference markets are determined according to the user's selection among those candidates. Then, finally, the new market's parameters are estimated from the reference markets. For this procedure, two techniques are used in the model. One is clustering data mining technique, and the other content-based filtering of recommender systems. The proposed system implemented with those techniques can determine the most adjacent markets based on whether a user accepts candidate markets. Experiments were conducted to validate the usefulness of the system with five ICT experts involved. In the experiments, the experts were given the list of 16 ICT markets whose parameters to be estimated. For each of the markets, the experts estimated its parameters of growth curve models with intuition at first, and then with the system. The comparison of the experiments results show that the estimated parameters are closer when they use the system in comparison with the results when they guessed them without the system.

Information Mediating in Social Network Sites : A Simulation Study (소셜 네트워크 사이트의 정보 매개하기 : 시뮬레이션 연구)

  • Rho, Sangkyu;Kim, Taekyung;Park, Jinsoo
    • The Journal of Society for e-Business Studies
    • /
    • v.18 no.1
    • /
    • pp.33-55
    • /
    • 2013
  • Information sharing behavior in the Internet has raised much interest. Recently, social network sites provide a new information sharing channel for the users who want to connect with others based on common social background or tastes. Especially, we focus that a social network site becomes one of major routes for information sharing about socially influential issues. Therefore, studying how information is diffused via a social network site may give theoretically, practically significant implication. Based on the assertion, we investigated user's behavior to mediate other user's information messages. We define information mediating behavior as concurrent actions of filtering and distributing behavior of the digital content that is originated from one of the connected users. In this study, we intended to understand the effects of information mediating behavior, and tried to understand characteristics of re-mediating of previously mediated information. Using an agent-based simulation model, we found that information mediating behavior increased the extent of information diffusion significantly. In addition, even a small degree of mediating probability could boost up the level of information diffusion in the case of a re-mediating condition. We believe that those findings provide remarkable insight of research and business application on both of information sharing and diffusion in a social network site.

Contents Conversion System for Mobile Devices using Light-Weight Web Document (웹 문서 경량화에 의한 모바일용 콘텐츠 변환 시스템)

  • Kim Jeong-Hee;Kwon Hoon;Kwak Ho-Young
    • Journal of Internet Computing and Services
    • /
    • v.6 no.6
    • /
    • pp.13-22
    • /
    • 2005
  • This paper aims to develop a system for converting web contents to mobile contents that can be used on mobile devices. Since web contents generally consist of pop-up ad windows, a bunch of unnecessary images and useless links, it is difficult to efficiently display them on common mobile devices that have lower bandwidth and memory, as well as much smaller screen, than the online environment. It is also troublesome for mobile device users to directly access contents. Thus, there has been a great demand for a new method for extracting useful and adequate contents from web documents, and optimizing them for use on mobile phones, In the paper, a system based on WAP 2,0 and XHTML Basic, which is a content creation language adopted for WAP 2,0, has been suggested. The system is designed to convert web contents by using the conversion rules of the existing filtering method after making the size of web documents smaller. The adopted conversion rules use the XHTML Basic's module units so that modification and deletion can be carried out with ease. In addition, it has been defined in a XSL document written in XSLT to maintain the extensibility of conversion and the validity of documents, In order to allow it to efficiently work together with WAP l.X's legacy services, the system has been built in a way that can have modules, which analyze information about CC/PP profiles and mobile device headers.

  • PDF

Preparation of Natural Seasoning using Enzymatic Hydrolysates from Byproducts of Alaska Pollock Theragra chalcogramma and Sea Tangle Laminaria japonica (명태(Theragra chalcogramma) 및 다시마(Laminaria japonica) 부산물 유래 효소 가수분해물을 이용한 천연 풍미 소재의 제조)

  • Kim, Jeong Gyun;Noh, Yuni;Park, Kwon Hyun;Lee, Ji Sun;Kim, Hyeon Jeong;Kim, Min Ji;Yoon, Moo Ho;Kim, Jin-Soo;Heu, Min Soo
    • Korean Journal of Fisheries and Aquatic Sciences
    • /
    • v.45 no.6
    • /
    • pp.545-552
    • /
    • 2012
  • This study developed a natural seasoning (NS) and characterized its food components. Hydrolysate from Alaska Pollock Theragra chalcogramma heads and sea tangle Laminaria japonica byproduct were obtained by incubating them with Neutrase for 4 h. NS was prepared by mixing sorbitol 2%, salt 2%, ginger powder 0.04%, garlic powder 0.2%, onion powder 0.2% and inosine monophosphate (IMP) 0.1% based on concentrated hydrolysates from Alaska pollock head and sea tangle byproduct before vaccum filtering. The proximate composition of NS was 82.7% moisture, 9.0% crude protein, and 5.1% ash. It had a higher crude protein content than commercial anchovy sauce (CS), it was lower in moisture and ash. The 1,1-diphenyl-2-picrylhydrazyl (DPPH) radical scavenging activity and angiotensin-I converting enzyme (ACE) inhibiting activity of NS were 90.1% and 88.9%, respectively, which were superior to those of CS. The free amino acid content and total taste value of NS were 1,626.0 mg/100 mL and 165.86, respectively, which were higher than those of CS. According to the results of taste value, the major free amino acids were glutamic acid and aspartic acid. In the sensory evaluation, the color and taste of NS were superior to those of CS. No difference in fish odor between NS and CS was found.

Soil Improvement Effect of Waste Lime Sludge Using Prefabricated Vertical Drains (연직배수재를 이용한 폐석회 슬러지의 지반개량 효과)

  • Shin, Eun-Chul;Park, Jeong-Jun;Kim, Jong-In
    • Journal of the Korean GEO-environmental Society
    • /
    • v.6 no.2
    • /
    • pp.51-60
    • /
    • 2005
  • The disposal problem of waste lime which is a residual product of lime industry have caused a lots of arguments in the past few years. Further more, waste lime contains a high moisture content which causes the disposal of waste lime is a great difficulty. The purpose of this study is to investigate for the effective dewatering solutions by placing various prefabricated vertical drains. The moisture content and degree of consolidation, pore water pressure, changes of settlement, bearing capacity with various vertical drains in waste lime were analyzed. The laboratory test results indicate that PBD is 2 times higher than circular drain in coefficient of consolidation. Based on the laboratory test results, settlement, pore water pressure, and dewatering measurements are shown in similar tendency. It is considered that PBD can drain primitive pore water much efficiently. The picture of SEM shows that circular drain filter has a serious clogging problem in comparison with PBD. In conclusion, PBD holds a superiority in waste lime's ground improvement and dewatering pore water pressure from the waste lime sludge. Also, circular drain is desired for some modification in its filtering system.

  • PDF

A Study on the Effect of Network Centralities on Recommendation Performance (네트워크 중심성 척도가 추천 성능에 미치는 영향에 대한 연구)

  • Lee, Dongwon
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.23-46
    • /
    • 2021
  • Collaborative filtering, which is often used in personalization recommendations, is recognized as a very useful technique to find similar customers and recommend products to them based on their purchase history. However, the traditional collaborative filtering technique has raised the question of having difficulty calculating the similarity for new customers or products due to the method of calculating similaritiesbased on direct connections and common features among customers. For this reason, a hybrid technique was designed to use content-based filtering techniques together. On the one hand, efforts have been made to solve these problems by applying the structural characteristics of social networks. This applies a method of indirectly calculating similarities through their similar customers placed between them. This means creating a customer's network based on purchasing data and calculating the similarity between the two based on the features of the network that indirectly connects the two customers within this network. Such similarity can be used as a measure to predict whether the target customer accepts recommendations. The centrality metrics of networks can be utilized for the calculation of these similarities. Different centrality metrics have important implications in that they may have different effects on recommended performance. In this study, furthermore, the effect of these centrality metrics on the performance of recommendation may vary depending on recommender algorithms. In addition, recommendation techniques using network analysis can be expected to contribute to increasing recommendation performance even if they apply not only to new customers or products but also to entire customers or products. By considering a customer's purchase of an item as a link generated between the customer and the item on the network, the prediction of user acceptance of recommendation is solved as a prediction of whether a new link will be created between them. As the classification models fit the purpose of solving the binary problem of whether the link is engaged or not, decision tree, k-nearest neighbors (KNN), logistic regression, artificial neural network, and support vector machine (SVM) are selected in the research. The data for performance evaluation used order data collected from an online shopping mall over four years and two months. Among them, the previous three years and eight months constitute social networks composed of and the experiment was conducted by organizing the data collected into the social network. The next four months' records were used to train and evaluate recommender models. Experiments with the centrality metrics applied to each model show that the recommendation acceptance rates of the centrality metrics are different for each algorithm at a meaningful level. In this work, we analyzed only four commonly used centrality metrics: degree centrality, betweenness centrality, closeness centrality, and eigenvector centrality. Eigenvector centrality records the lowest performance in all models except support vector machines. Closeness centrality and betweenness centrality show similar performance across all models. Degree centrality ranking moderate across overall models while betweenness centrality always ranking higher than degree centrality. Finally, closeness centrality is characterized by distinct differences in performance according to the model. It ranks first in logistic regression, artificial neural network, and decision tree withnumerically high performance. However, it only records very low rankings in support vector machine and K-neighborhood with low-performance levels. As the experiment results reveal, in a classification model, network centrality metrics over a subnetwork that connects the two nodes can effectively predict the connectivity between two nodes in a social network. Furthermore, each metric has a different performance depending on the classification model type. This result implies that choosing appropriate metrics for each algorithm can lead to achieving higher recommendation performance. In general, betweenness centrality can guarantee a high level of performance in any model. It would be possible to consider the introduction of proximity centrality to obtain higher performance for certain models.

Clustering-based Hierarchical Scene Structure Construction for Movie Videos (영화 비디오를 위한 클러스터링 기반의 계층적 장면 구조 구축)

  • Choi, Ick-Won;Byun, Hye-Ran
    • Journal of KIISE:Software and Applications
    • /
    • v.27 no.5
    • /
    • pp.529-542
    • /
    • 2000
  • Recent years, the use of multimedia information is rapidly increasing, and the video media is the most rising one than any others, and this field Integrates all the media into a single data stream. Though the availability of digital video is raised largely, it is very difficult for users to make the effective video access, due to its length and unstructured video format. Thus, the minimal interaction of users and the explicit definition of video structure is a key requirement in the lately developing image and video management systems. This paper defines the terms and hierarchical video structure, and presents the system, which construct the clustering-based video hierarchy, which facilitate users by browsing the summary and do a random access to the video content. Instead of using a single feature and domain-specific thresholds, we use multiple features that have complementary relationship for each other and clustering-based methods that use normalization so as to interact with users minimally. The stage of shot boundary detection extracts multiple features, performs the adaptive filtering process for each features to enhance the performance by eliminating the false factors, and does k-means clustering with two classes. The shot list of a result after the proposed procedure is represented as the video hierarchy by the intelligent unsupervised clustering technique. We experimented the static and the dynamic movie videos that represent characteristics of various video types. In the result of shot boundary detection, we had almost more than 95% good performance, and had also rood result in the video hierarchy.

  • PDF

Effective Morphological Layer Segmentation Based on Edge Information for Screen Image Coding (스크린 이미지 부호화를 위한 에지 정보 기반의 효과적인 형태학적 레이어 분할)

  • Park, Sang-Hyo;Lee, Si-Woong
    • The Journal of the Korea Contents Association
    • /
    • v.13 no.12
    • /
    • pp.38-47
    • /
    • 2013
  • An image coding based on MRC model, a kind of multi-layer image model, first segments a screen image into foreground, mask, and background layers, and then compresses each layer using a codec that is suitable to the layer. The mask layer defines the position of foreground regions such as textual and graphical contents. The colour signal of the foreground (background) region is saved in the foreground (background) layer. The mask layer which contains the segmentation result of foreground and background regions is of importance since its accuracy directly affects the overall coding performance of the codec. This paper proposes a new layer segmentation algorithm for the MRC based image coding. The proposed method extracts text pixels from the background using morphological top hat filtering. The application of white or black top hat transformation to local blocks is controlled by the information of relative brightness of text compared to the background. In the proposed method, the boundary information of text that is extracted from the edge map of the block is used for the robust decision on the relative brightness of text. Simulation results show that the proposed method is superior to the conventional methods.

Design and Implementation of High-dimensional Index Structure for the support of Concurrency Control (필터링에 기반한 고차원 색인구조의 동시성 제어기법의 설계 및 구현)

  • Lee, Yong-Ju;Chang, Jae-Woo;Kim, Hang-Young;Kim, Myung-Joon
    • The KIPS Transactions:PartD
    • /
    • v.10D no.1
    • /
    • pp.1-12
    • /
    • 2003
  • Recently, there have been many indexing schemes for multimedia data such as image, video data. But recent database applications, for example data mining and multimedia database, are required to support multi-user environment. In order for indexing schemes to be useful in multi-user environment, a concurrency control algorithm is required to handle it. So we propose a concurrency control algorithm that can be applied to CBF (cell-based filtering method), which uses the signature of the cell for alleviating the dimensional curse problem. In addition, we extend the SHORE storage system of Wisconsin university in order to handle high-dimensional data. This extended SHORE storage system provides conventional storage manager functions, guarantees the integrity of high-dimensional data and is flexible to the large scale of feature vectors for preventing the usage of large main memory. Finally, we implement the web-based image retrieval system by using the extended SHORE storage system. The key feature of this system is platform-independent access to the high-dimensional data as well as functionality of efficient content-based queries. Lastly. We evaluate an average response time of point query, range query and k-nearest query in terms of the number of threads.

Predicting the Performance of Recommender Systems through Social Network Analysis and Artificial Neural Network (사회연결망분석과 인공신경망을 이용한 추천시스템 성능 예측)

  • Cho, Yoon-Ho;Kim, In-Hwan
    • Journal of Intelligence and Information Systems
    • /
    • v.16 no.4
    • /
    • pp.159-172
    • /
    • 2010
  • The recommender system is one of the possible solutions to assist customers in finding the items they would like to purchase. To date, a variety of recommendation techniques have been developed. One of the most successful recommendation techniques is Collaborative Filtering (CF) that has been used in a number of different applications such as recommending Web pages, movies, music, articles and products. CF identifies customers whose tastes are similar to those of a given customer, and recommends items those customers have liked in the past. Numerous CF algorithms have been developed to increase the performance of recommender systems. Broadly, there are memory-based CF algorithms, model-based CF algorithms, and hybrid CF algorithms which combine CF with content-based techniques or other recommender systems. While many researchers have focused their efforts in improving CF performance, the theoretical justification of CF algorithms is lacking. That is, we do not know many things about how CF is done. Furthermore, the relative performances of CF algorithms are known to be domain and data dependent. It is very time-consuming and expensive to implement and launce a CF recommender system, and also the system unsuited for the given domain provides customers with poor quality recommendations that make them easily annoyed. Therefore, predicting the performances of CF algorithms in advance is practically important and needed. In this study, we propose an efficient approach to predict the performance of CF. Social Network Analysis (SNA) and Artificial Neural Network (ANN) are applied to develop our prediction model. CF can be modeled as a social network in which customers are nodes and purchase relationships between customers are links. SNA facilitates an exploration of the topological properties of the network structure that are implicit in data for CF recommendations. An ANN model is developed through an analysis of network topology, such as network density, inclusiveness, clustering coefficient, network centralization, and Krackhardt's efficiency. While network density, expressed as a proportion of the maximum possible number of links, captures the density of the whole network, the clustering coefficient captures the degree to which the overall network contains localized pockets of dense connectivity. Inclusiveness refers to the number of nodes which are included within the various connected parts of the social network. Centralization reflects the extent to which connections are concentrated in a small number of nodes rather than distributed equally among all nodes. Krackhardt's efficiency characterizes how dense the social network is beyond that barely needed to keep the social group even indirectly connected to one another. We use these social network measures as input variables of the ANN model. As an output variable, we use the recommendation accuracy measured by F1-measure. In order to evaluate the effectiveness of the ANN model, sales transaction data from H department store, one of the well-known department stores in Korea, was used. Total 396 experimental samples were gathered, and we used 40%, 40%, and 20% of them, for training, test, and validation, respectively. The 5-fold cross validation was also conducted to enhance the reliability of our experiments. The input variable measuring process consists of following three steps; analysis of customer similarities, construction of a social network, and analysis of social network patterns. We used Net Miner 3 and UCINET 6.0 for SNA, and Clementine 11.1 for ANN modeling. The experiments reported that the ANN model has 92.61% estimated accuracy and 0.0049 RMSE. Thus, we can know that our prediction model helps decide whether CF is useful for a given application with certain data characteristics.