• Title/Summary/Keyword: target problem

Search Result 1,781, Processing Time 0.032 seconds

The Effect of Price Promotional Information about Brand on Consumer's Quality Perception: Conditioning on Pretrial Brand (품패개격촉소신식대소비자질량인지적영향(品牌价格促销信息对消费者质量认知的影响))

  • Lee, Min-Hoon;Lim, Hang-Seop
    • Journal of Global Scholars of Marketing Science
    • /
    • v.19 no.3
    • /
    • pp.17-27
    • /
    • 2009
  • Price promotion typically reduces the price for a given quantity or increases the quantity available at the same price, thereby enhancing value and creating an economic incentive to purchase. It often is used to encourage product or service trial among nonusers of products or services. Thus, it is important to understand the effects of price promotions on quality perception made by consumer who do not have prior experience with the promoted brand. However, if consumers associate a price promotion itself with inferior brand quality, the promotion may not achieve the sales increase the economic incentives otherwise might have produced. More specifically, low qualitative perception through price promotion will undercut the economic and psychological incentives and reduce the likelihood of purchase. Thus, it is important for marketers to understand how price promotional informations about a brand have impact on consumer's unfavorable quality perception of the brand. Previous literatures on the effects of price promotions on quality perception reveal inconsistent explanations. Some focused on the unfavorable effect of price promotion on consumer's perception. But others showed that price promotions didn't raise unfavorable perception on the brand. Prior researches found these inconsistent results related to the timing of the price promotion's exposure and quality evaluation relative to trial. And, whether the consumer has been experienced with the product promotions in the past or not may moderate the effects. A few studies considered differences among product categories as fundamental factors. The purpose of this research is to investigate the effect of price promotional informations on consumer's unfavorable quality perception under the different conditions. The author controlled the timing of the promotional exposure and varied past promotional patterns and information presenting patterns. Unlike previous researches, the author examined the effects of price promotions setting limit to pretrial situation by controlling potentially moderating effects of prior personal experience with the brand. This manipulations enable to resolve possible controversies in relation to this issue. And this manipulation is meaningful for the work sector. Price promotion is not only used to target existing consumers but also to encourage product or service trial among nonusers of products or services. Thus, it is important for marketers to understand how price promotional informations about a brand have impact on consumer's unfavorable quality perception of the brand. If consumers associate a price promotion itself with inferior quality about unused brand, the promotion may not achieve the sales increase the economic incentives otherwise might have produced. In addition, if the price promotion ends, the consumer that have purchased that certain brand will likely to display sharply decreased repurchasing behavior. Through a literature review, hypothesis 1 was set as follows to investigate the adjustive effect of past price promotion on quality perception made by consumers; The influence that price promotion of unused brand have on quality perception made by consumers will be adjusted by past price promotion activity of the brand. In other words, a price promotion of an unused brand that have not done a price promotion in the past will have a unfavorable effect on quality perception made by consumer. Hypothesis 2-1 was set as follows : When an unused brand undertakes price promotion for the first time, the information presenting pattern of price promotion will have an effect on the consumer's attribution for the cause of the price promotion. Hypothesis 2-2 was set as follows : The more consumer dispositionally attribute the cause of price promotion, the more unfavorable the quality perception made by consumer will be. Through test 1, the subjects were given a brief explanation of the product and the brand before they were provided with a $2{\times}2$ factorial design that has 4 patterns of price promotion (presence or absence of past price promotion * presence or absence of current price promotion) and the explanation describing the price promotion pattern of each cell. Then the perceived quality of imaginary brand WAVEX was evaluated in the scale of 7. The reason tennis racket was chosen is because the selected product group must have had almost no past price promotions to eliminate the influence of average frequency of promotion on the value of price promotional information as Raghubir and Corfman (1999) pointed out. Test 2 was also carried out on students of the same management faculty of test 1 with tennis racket as the product group. As with test 1, subjects with average familiarity for the product group and low familiarity for the brand was selected. Each subjects were assigned to one of the two cells representing two different information presenting patterns of price promotion of WAVEX (case where the reason behind price promotion was provided/case where the reason behind price promotion was not provided). Subjects looked at each promotional information before evaluating the perceived quality of the brand WAVEX in the scale of 7. The effect of price promotion for unfamiliar pretrial brand on consumer's perceived quality was proved to be moderated with the presence or absence of past price promotion. The consistency with past promotional behavior is important variable that makes unfavorable effect on brand evaluations get worse. If the price promotion for the brand has never been carried out before, price promotion activity may have more unfavorable effects on consumer's quality perception. Second, when the price promotion of unfamiliar pretrial brand was executed for the first time, presenting method of informations has impact on consumer's attribution for the cause of firm's promotion. And the unfavorable effect of quality perception is higher when the consumer does dispositional attribution comparing with situational attribution. Unlike the previous studies where the main focus was the absence or presence of favorable or unfavorable motivation from situational/dispositional attribution, the focus of this study was exaus ing the fact that a situational attribution can be inferred even if the consumer employs a dispositional attribution on the price promotional behavior, if the company provides a persuasive reason. Such approach, in academic perspectih sis a large significance in that it explained the anchoring and adjng ch approcedures by applying it to a non-mathematical problem unlike the previous studies where it wis ionaly explained by applying it to a mathematical problem. In other wordn, there is a highrspedency tmatispositionally attribute other's behaviors according to the fuedach aal attribution errors and when this is applied to the situation of price promotions, we can infer that consumers are likely tmatispositionally attribute the company's price promotion behaviors. Ha ever, even ueder these circumstances, the company can adjng the consumer's anchoring tmareduce the po wibiliute thdispositional attribution. Furthermore, unlike majority of previous researches on short/long-term effects of price promotion that only considered the effect of price promotions on consumer's purchasing behaviors, this research measured the effect on perceived quality, one of man elements that affects the purchasing behavior of consumers. These results carry useful implications for the work sector. A guideline of effectively providing promotional informations for a new brand can be suggested through the outcomes of this research. If the brand is to avoid false implications such as inferior quality while implementing a price promotion strategy, it must provide a clear and acceptable reasons behind the promotion. Especially it is more important for the company with no past price promotion to provide a clear reason. An inconsistent behavior can be the cause of consumer's distrust and anxiety. This is also one of the most important factor of risk of endless price wars. Price promotions without prior notice can buy doubt from consumers not market share.

  • PDF

Clickstream Big Data Mining for Demographics based Digital Marketing (인구통계특성 기반 디지털 마케팅을 위한 클릭스트림 빅데이터 마이닝)

  • Park, Jiae;Cho, Yoonho
    • Journal of Intelligence and Information Systems
    • /
    • v.22 no.3
    • /
    • pp.143-163
    • /
    • 2016
  • The demographics of Internet users are the most basic and important sources for target marketing or personalized advertisements on the digital marketing channels which include email, mobile, and social media. However, it gradually has become difficult to collect the demographics of Internet users because their activities are anonymous in many cases. Although the marketing department is able to get the demographics using online or offline surveys, these approaches are very expensive, long processes, and likely to include false statements. Clickstream data is the recording an Internet user leaves behind while visiting websites. As the user clicks anywhere in the webpage, the activity is logged in semi-structured website log files. Such data allows us to see what pages users visited, how long they stayed there, how often they visited, when they usually visited, which site they prefer, what keywords they used to find the site, whether they purchased any, and so forth. For such a reason, some researchers tried to guess the demographics of Internet users by using their clickstream data. They derived various independent variables likely to be correlated to the demographics. The variables include search keyword, frequency and intensity for time, day and month, variety of websites visited, text information for web pages visited, etc. The demographic attributes to predict are also diverse according to the paper, and cover gender, age, job, location, income, education, marital status, presence of children. A variety of data mining methods, such as LSA, SVM, decision tree, neural network, logistic regression, and k-nearest neighbors, were used for prediction model building. However, this research has not yet identified which data mining method is appropriate to predict each demographic variable. Moreover, it is required to review independent variables studied so far and combine them as needed, and evaluate them for building the best prediction model. The objective of this study is to choose clickstream attributes mostly likely to be correlated to the demographics from the results of previous research, and then to identify which data mining method is fitting to predict each demographic attribute. Among the demographic attributes, this paper focus on predicting gender, age, marital status, residence, and job. And from the results of previous research, 64 clickstream attributes are applied to predict the demographic attributes. The overall process of predictive model building is compose of 4 steps. In the first step, we create user profiles which include 64 clickstream attributes and 5 demographic attributes. The second step performs the dimension reduction of clickstream variables to solve the curse of dimensionality and overfitting problem. We utilize three approaches which are based on decision tree, PCA, and cluster analysis. We build alternative predictive models for each demographic variable in the third step. SVM, neural network, and logistic regression are used for modeling. The last step evaluates the alternative models in view of model accuracy and selects the best model. For the experiments, we used clickstream data which represents 5 demographics and 16,962,705 online activities for 5,000 Internet users. IBM SPSS Modeler 17.0 was used for our prediction process, and the 5-fold cross validation was conducted to enhance the reliability of our experiments. As the experimental results, we can verify that there are a specific data mining method well-suited for each demographic variable. For example, age prediction is best performed when using the decision tree based dimension reduction and neural network whereas the prediction of gender and marital status is the most accurate by applying SVM without dimension reduction. We conclude that the online behaviors of the Internet users, captured from the clickstream data analysis, could be well used to predict their demographics, thereby being utilized to the digital marketing.

The Reserch on Actual Condition of Crime of Arson Which Occurs in Korea and Its Countermeasures (방화범죄의 실태와 그 대책 - 관심도와 동기의 다양화에 대한 대응 -)

  • Choi, Jong-Tae
    • Korean Security Journal
    • /
    • no.1
    • /
    • pp.371-408
    • /
    • 1997
  • This article is the reserch on actual condition of crime of arson which occurs in Korea and its countermeasures. The the presented problem in this article are that (1) we have generally very low rate concern about the crime of arson contrary to realistic problems of rapid increase of crime of arson (2) as such criminal motives became so diverse as to the economic or criminal purpose unlike characteristic and mental deficiency of old days, and to countermeasure these problems effectively it presentation the necessity of systemantic research. Based on analysis of reality of arson, the tendency of this arson in Korea in the ratio of increase is said to be higher than those in violence crime or general fire rate. and further its rate is far more greater than those of the U.S.A. and Japan. Arson is considered to be a method of using fire as crime and in case of presently residence to be the abject, it is a public offense crime which aqccompany fatality in human life. This is the well It now fact to all of us. And further in order to presentation to the crime of arson, strictness of criminal law (criminal law No, 164 and 169, and fire protection law No. 110 and 111) and classification of arsonist as felony are institutionary reinforced to punish with certainty of possibility, Therefore, as tendency of arson has been increased compared to other nations, it is necessary to supplement strategical policy to bring out overall concerns of the seriousness of risk and damage of arson, which have been resulted from the lack of understanding. In characteristics analysis of crime of arson, (1) It is now reveald that, in the past such crime rate appeared far more within the boundary of town or city areas in the past, presently increased rate of arsons in rural areas are far more than in the town or small city areas, thereby showing characteristics of crime of arson extending nation wide. (2) general timetable of arson shows that night more than day time rate, and reveald that is trait behavior in secrecy.(3) arsonists are usually arrested at site or by victim or report of third person(82,9%).Investigation activities or self surrenders rate only 11.2%. The time span of arrest is normally the same day of arson and at times it takes more than one year to arrest. This reveals its necessity to prepare for long period of time for arrest, (4) age rate of arson is in their thirties mostly as compared to homicide, robbery and adultery, and considerable numbers of arsons are in old age of over fifties. It reveals age rate is increased (5) Over half of the arsonists are below the junior high school (6) the rate of convicts by thier records is based on first offenders primarily and secondly more than 4 time convicts. This apparently shows necessity of effective correctional education policy for their social assimilation together with re-investigation of human education at the primary and secondary education system in thier life. The examples of motivation for arosnits, such as personal animosity, fury, monetary swindle, luscious purpose and other aims of destroying of proof, and other social resistance, violence including ways of threatening, beside the motives of individual defects, are diverse and arsonic suicide and specifically suicidal accompany together keenly manifested. When we take this fact with the criminal theory, it really reveals arsons of crime are increasing and its casualities are serious and a point as a way of suicide is the anomie theory of Durkheim and comensurate with the theory of that of Merton, Specifically in the arson of industrial complex, it is revealed that one with revolutionary motive or revolting motive would do the arsonic act. For the policy of prevention of arsons, professional research work in organizational cooperation for preventive activities is conducted in municipal or city wise functions in the name of Parson Taskforces and beside a variety of research institutes in federal government have been operating effectively to countermeasure in many fields of research. Franch and Sweden beside the U.S. set up a overall operation of fire prevention research funtions and have obtained very successful result. Japan also put their research likewise for countermeasure. In this research as a way of preventive fire policy, first, it is necessary to accomodate the legal preventitive activities for fire prevention in judicial side and as an administrative side, (1) precise statistic management of crime of arson (2) establishment of professional research functions or a corporate (3) improvement of system for cooperative structural team for investigation of fires and menpower organization of professional members. Secondly, social mentality in individual prospect, recognition of fires by arson and youth education of such effect, educational program for development and practical promotion. Thirdly, in view of environmental side, the ways of actual performance by programming with the establishment of cooperative advancement in local social function elements with administrative office, habitants, school facilities and newspapers measures (2) establishment of personal protection where weak menpowers are displayed in special fire prevention measures. These measures are presented for prevention of crime of arson. The control of crime and prevention shall be prepared as a means of self defence by the principle of self responsibility Specifically arsonists usually aims at the comparatively weak control of fire prevention is prevalent and it is therefore necessary to prepare individual facilities with their spontaneous management of fire prevention instead of public municipal funtures of local geverment. As Clifford L. Karchmer asserted instead of concerns about who would commit arson, what portion of area would be the target of the arson. It is effective to minister spontaveously the fire prevention measure in his facility with the consideration of characteristics of arson. On the other hand, it is necessary for the concerned personnel of local goverment and groups to distribute to the local society in timely manner for new information about the fire prevention, thus contribute to effective result of fire prevention result. In consideration of these factors, it is inevitable to never let coincide with the phemonemon of arsons in similar or mimic features as recognized that these could prevail just an epedemic as a strong imitational attitude. In processing of policy to encounter these problems, it is necessary to place priority of city policy to enhancement of overall concerns toward the definitive essense of crime of arson.

  • PDF

Case Analysis of the Promotion Methodologies in the Smart Exhibition Environment (스마트 전시 환경에서 프로모션 적용 사례 및 분석)

  • Moon, Hyun Sil;Kim, Nam Hee;Kim, Jae Kyeong
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.171-183
    • /
    • 2012
  • In the development of technologies, the exhibition industry has received much attention from governments and companies as an important way of marketing activities. Also, the exhibitors have considered the exhibition as new channels of marketing activities. However, the growing size of exhibitions for net square feet and the number of visitors naturally creates the competitive environment for them. Therefore, to make use of the effective marketing tools in these environments, they have planned and implemented many promotion technics. Especially, through smart environment which makes them provide real-time information for visitors, they can implement various kinds of promotion. However, promotions ignoring visitors' various needs and preferences can lose the original purposes and functions of them. That is, as indiscriminate promotions make visitors feel like spam, they can't achieve their purposes. Therefore, they need an approach using STP strategy which segments visitors through right evidences (Segmentation), selects the target visitors (Targeting), and give proper services to them (Positioning). For using STP Strategy in the smart exhibition environment, we consider these characteristics of it. First, an exhibition is defined as market events of a specific duration, which are held at intervals. According to this, exhibitors who plan some promotions should different events and promotions in each exhibition. Therefore, when they adopt traditional STP strategies, a system can provide services using insufficient information and of existing visitors, and should guarantee the performance of it. Second, to segment automatically, cluster analysis which is generally used as data mining technology can be adopted. In the smart exhibition environment, information of visitors can be acquired in real-time. At the same time, services using this information should be also provided in real-time. However, many clustering algorithms have scalability problem which they hardly work on a large database and require for domain knowledge to determine input parameters. Therefore, through selecting a suitable methodology and fitting, it should provide real-time services. Finally, it is needed to make use of data in the smart exhibition environment. As there are useful data such as booth visit records and participation records for events, the STP strategy for the smart exhibition is based on not only demographical segmentation but also behavioral segmentation. Therefore, in this study, we analyze a case of the promotion methodology which exhibitors can provide a differentiated service to segmented visitors in the smart exhibition environment. First, considering characteristics of the smart exhibition environment, we draw evidences of segmentation and fit the clustering methodology for providing real-time services. There are many studies for classify visitors, but we adopt a segmentation methodology based on visitors' behavioral traits. Through the direct observation, Veron and Levasseur classify visitors into four groups to liken visitors' traits to animals (Butterfly, fish, grasshopper, and ant). Especially, because variables of their classification like the number of visits and the average time of a visit can estimate in the smart exhibition environment, it can provide theoretical and practical background for our system. Next, we construct a pilot system which automatically selects suitable visitors along the objectives of promotions and instantly provide promotion messages to them. That is, based on the segmentation of our methodology, our system automatically selects suitable visitors along the characteristics of promotions. We adopt this system to real exhibition environment, and analyze data from results of adaptation. As a result, as we classify visitors into four types through their behavioral pattern in the exhibition, we provide some insights for researchers who build the smart exhibition environment and can gain promotion strategies fitting each cluster. First, visitors of ANT type show high response rate for promotion messages except experience promotion. So they are fascinated by actual profits in exhibition area, and dislike promotions requiring a long time. Contrastively, visitors of GRASSHOPPER type show high response rate only for experience promotion. Second, visitors of FISH type appear favors to coupon and contents promotions. That is, although they don't look in detail, they prefer to obtain further information such as brochure. Especially, exhibitors that want to give much information for limited time should give attention to visitors of this type. Consequently, these promotion strategies are expected to give exhibitors some insights when they plan and organize their activities, and grow the performance of them.

A Hybrid SVM Classifier for Imbalanced Data Sets (불균형 데이터 집합의 분류를 위한 하이브리드 SVM 모델)

  • Lee, Jae Sik;Kwon, Jong Gu
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.125-140
    • /
    • 2013
  • We call a data set in which the number of records belonging to a certain class far outnumbers the number of records belonging to the other class, 'imbalanced data set'. Most of the classification techniques perform poorly on imbalanced data sets. When we evaluate the performance of a certain classification technique, we need to measure not only 'accuracy' but also 'sensitivity' and 'specificity'. In a customer churn prediction problem, 'retention' records account for the majority class, and 'churn' records account for the minority class. Sensitivity measures the proportion of actual retentions which are correctly identified as such. Specificity measures the proportion of churns which are correctly identified as such. The poor performance of the classification techniques on imbalanced data sets is due to the low value of specificity. Many previous researches on imbalanced data sets employed 'oversampling' technique where members of the minority class are sampled more than those of the majority class in order to make a relatively balanced data set. When a classification model is constructed using this oversampled balanced data set, specificity can be improved but sensitivity will be decreased. In this research, we developed a hybrid model of support vector machine (SVM), artificial neural network (ANN) and decision tree, that improves specificity while maintaining sensitivity. We named this hybrid model 'hybrid SVM model.' The process of construction and prediction of our hybrid SVM model is as follows. By oversampling from the original imbalanced data set, a balanced data set is prepared. SVM_I model and ANN_I model are constructed using the imbalanced data set, and SVM_B model is constructed using the balanced data set. SVM_I model is superior in sensitivity and SVM_B model is superior in specificity. For a record on which both SVM_I model and SVM_B model make the same prediction, that prediction becomes the final solution. If they make different prediction, the final solution is determined by the discrimination rules obtained by ANN and decision tree. For a record on which SVM_I model and SVM_B model make different predictions, a decision tree model is constructed using ANN_I output value as input and actual retention or churn as target. We obtained the following two discrimination rules: 'IF ANN_I output value <0.285, THEN Final Solution = Retention' and 'IF ANN_I output value ${\geq}0.285$, THEN Final Solution = Churn.' The threshold 0.285 is the value optimized for the data used in this research. The result we present in this research is the structure or framework of our hybrid SVM model, not a specific threshold value such as 0.285. Therefore, the threshold value in the above discrimination rules can be changed to any value depending on the data. In order to evaluate the performance of our hybrid SVM model, we used the 'churn data set' in UCI Machine Learning Repository, that consists of 85% retention customers and 15% churn customers. Accuracy of the hybrid SVM model is 91.08% that is better than that of SVM_I model or SVM_B model. The points worth noticing here are its sensitivity, 95.02%, and specificity, 69.24%. The sensitivity of SVM_I model is 94.65%, and the specificity of SVM_B model is 67.00%. Therefore the hybrid SVM model developed in this research improves the specificity of SVM_B model while maintaining the sensitivity of SVM_I model.

Validity and Pertinence of Administrative Capital City Proposal (행정수도 건설안의 타당성과 시의성)

  • 김형국
    • Journal of the Korean Geographical Society
    • /
    • v.38 no.2
    • /
    • pp.312-323
    • /
    • 2003
  • This writer absolutely agrees with the government that regional disequilibrium is severe enough to consider moving the administrative capital. Pursuing this course solely to establish a balanced development, however, is not a convincing enough reason. The capital city is directly related to not only the social and economic situation but, much more importantly, to the domestic political situation as well. In the mid-1970s, the proposal by the Third Republic to move the capital city temporarily was based completely on security reasons. At e time, the then opposition leader Kim, Dae-jung said that establishing a safe distance from the demilitarized zone(DMZ) reflected a typically military decision. His view was that retaining the capital city close to the DMZ would show more consideration for the will of the people to defend their own country. In fact, independent Pakistan moved its capital city from Karachi to Islamabad, situated dose to Kashmir the subject of hot territorial dispute with India. It is regrettable that no consideration has been given to the urgent political situation in the Korean peninsula, which is presently enveloped in a dense nuclear fog. As a person requires health to pursue his/her dream, a country must have security to implement a balanced territorial development. According to current urban theories, the fate of a country depends on its major cities. A negligently guarded capital city runs the risk of becoming hostage and bringing ruin to the whole country. In this vein, North Koreas undoubted main target of attack in the armed communist reunification of Korea is Seoul. For the preservation of our state, therefore, it is only right that Seoul must be shielded to prevent becoming hostage to North Korea. The location of the US Armed Forces to the north of the capital city is based on the judgment that defense of Seoul is of absolute importance. At the same time, regardless of their different standpoints, South and North Korea agree that division of the Korean people into two separate countries is abnormal. Reunification, which so far has defied all predictions, may be realized earlier than anyone expects. The day of reunification seems to be the best day for the relocation of the capital city. Building a proper capital city would take at least twenty years, and a capital city cannot be dragged from one place to another. On the day of a free and democratic reunification, a national agreement will be reached naturally to find a nationally symbolic city as in Brazil or Australia. Even if security does not pose a problem, the governments way of thinking would not greatly contribute to the balanced development of the country. The Chungcheon region, which is earmarked as the new location of the capital city, has been the greatest beneficiary of its proximity to the capital region. Not being a disadvantaged region, locating the capital city there would not help alleviate regional disparity. If it is absolutely necessary to find a candidate region at present, considering security, balanced regional development and post-reunification scenario of the future, Cheolwon area located in the middle of the Korean peninsula may be a plausible choice. Even if the transfer of capital is delayed in consideration of the present political conflict between the South and the North Koreas, there is a definite shortcut to realizing a balanced regional development. It can be found not in the geographical dispersal of the central government, but in the decentralization of power to the provinces. If the government has surplus money to build a new symbolic capital city, it is only right that it should improve, for instance, the quality of drinking water which now everyone eschews, and to help the regional subway authority whose chronic deficit state resoled in a recent disastrous accident. And it is proper to time the transfer of capital city to coincide with that of the reunification of Korea whenever Providence intends.

A Study on the Effect of Network Centralities on Recommendation Performance (네트워크 중심성 척도가 추천 성능에 미치는 영향에 대한 연구)

  • Lee, Dongwon
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.23-46
    • /
    • 2021
  • Collaborative filtering, which is often used in personalization recommendations, is recognized as a very useful technique to find similar customers and recommend products to them based on their purchase history. However, the traditional collaborative filtering technique has raised the question of having difficulty calculating the similarity for new customers or products due to the method of calculating similaritiesbased on direct connections and common features among customers. For this reason, a hybrid technique was designed to use content-based filtering techniques together. On the one hand, efforts have been made to solve these problems by applying the structural characteristics of social networks. This applies a method of indirectly calculating similarities through their similar customers placed between them. This means creating a customer's network based on purchasing data and calculating the similarity between the two based on the features of the network that indirectly connects the two customers within this network. Such similarity can be used as a measure to predict whether the target customer accepts recommendations. The centrality metrics of networks can be utilized for the calculation of these similarities. Different centrality metrics have important implications in that they may have different effects on recommended performance. In this study, furthermore, the effect of these centrality metrics on the performance of recommendation may vary depending on recommender algorithms. In addition, recommendation techniques using network analysis can be expected to contribute to increasing recommendation performance even if they apply not only to new customers or products but also to entire customers or products. By considering a customer's purchase of an item as a link generated between the customer and the item on the network, the prediction of user acceptance of recommendation is solved as a prediction of whether a new link will be created between them. As the classification models fit the purpose of solving the binary problem of whether the link is engaged or not, decision tree, k-nearest neighbors (KNN), logistic regression, artificial neural network, and support vector machine (SVM) are selected in the research. The data for performance evaluation used order data collected from an online shopping mall over four years and two months. Among them, the previous three years and eight months constitute social networks composed of and the experiment was conducted by organizing the data collected into the social network. The next four months' records were used to train and evaluate recommender models. Experiments with the centrality metrics applied to each model show that the recommendation acceptance rates of the centrality metrics are different for each algorithm at a meaningful level. In this work, we analyzed only four commonly used centrality metrics: degree centrality, betweenness centrality, closeness centrality, and eigenvector centrality. Eigenvector centrality records the lowest performance in all models except support vector machines. Closeness centrality and betweenness centrality show similar performance across all models. Degree centrality ranking moderate across overall models while betweenness centrality always ranking higher than degree centrality. Finally, closeness centrality is characterized by distinct differences in performance according to the model. It ranks first in logistic regression, artificial neural network, and decision tree withnumerically high performance. However, it only records very low rankings in support vector machine and K-neighborhood with low-performance levels. As the experiment results reveal, in a classification model, network centrality metrics over a subnetwork that connects the two nodes can effectively predict the connectivity between two nodes in a social network. Furthermore, each metric has a different performance depending on the classification model type. This result implies that choosing appropriate metrics for each algorithm can lead to achieving higher recommendation performance. In general, betweenness centrality can guarantee a high level of performance in any model. It would be possible to consider the introduction of proximity centrality to obtain higher performance for certain models.

A Study of Equipment Accuracy and Test Precision in Dual Energy X-ray Absorptiometry (골밀도검사의 올바른 질 관리에 따른 임상적용과 해석 -이중 에너지 방사선 흡수법을 중심으로-)

  • Dong, Kyung-Rae;Kim, Ho-Sung;Jung, Woon-Kwan
    • Journal of radiological science and technology
    • /
    • v.31 no.1
    • /
    • pp.17-23
    • /
    • 2008
  • Purpose : Because there is a difference depending on the environment as for an inspection equipment the important part of bone density scan and the precision/accuracy of a tester, the management of quality must be made systematically. The equipment failure caused by overload effect due to the aged equipment and the increase of a patient was made frequently. Thus, the replacement of equipment and additional purchases of new bonedensity equipment caused a compatibility problem in tracking patients. This study wants to know whether the clinical changes of patient's bonedensity can be accurately and precisely reflected when used it compatiblly like the existing equipment after equipment replacement and expansion. Materials and methods : Two equipments of GE Lunar Prodigy Advance(P1 and P2) and the Phantom HOLOGIC Spine Road(HSP) were used to measure equipment precision. Each device scans 20 times so that precision data was acquired from the phantom(Group 1). The precision of a tester was measured by shooting twice the same patient, every 15 members from each of the target equipment in 120 women(average age 48.78, 20-60 years old)(Group 2). In addition, the measurement of the precision of a tester and the cross-calibration data were made by scanning 20 times in each of the equipment using HSP, based on the data obtained from the management of quality using phantom(ASP) every morning (Group 3). The same patient was shot only once in one equipment alternately to make the measurement of the precision of a tester and the cross-calibration data in 120 women(average age 48.78, 20-60 years old)(Group 4). Results : It is steady equipment according to daily Q.C Data with $0.996\;g/cm^2$, change value(%CV) 0.08. The mean${\pm}$SD and a %CV price are ALP in Group 1(P1 : $1.064{\pm}0.002\;g/cm^2$, $%CV=0.190\;g/cm^2$, P2 : $1.061{\pm}0.003\;g/cm^2$, %CV=0.192). The mean${\pm}$SD and a %CV price are P1 : $1.187{\pm}0.002\;g/cm^2$, $%CV=0.164\;g/cm^2$, P2 : $1.198{\pm}0.002\;g/cm^2$, %CV=0.163 in Group 2. The average error${\pm}$2SD and %CV are P1 - (spine: $0.001{\pm}0.03\;g/cm^2$, %CV=0.94, Femur: $0.001{\pm}0.019\;g/cm^2$, %CV=0.96), P2 - (spine: $0.002{\pm}0.018\;g/cm^2$, %CV=0.55, Femur: $0.001{\pm}0.013\;g/cm^2$, %CV=0.48) in Group 3. The average error${\pm}2SD$, %CV, and r value was spine : $0.006{\pm}0.024\;g/cm^2$, %CV=0.86, r=0.995, Femur: $0{\pm}0.014\;g/cm^2$, %CV=0.54, r=0.998 in Group 4. Conclusion: Both LUNAR ASP CV% and HOLOGIC Spine Phantom are included in the normal range of error of ${\pm}2%$ defined in ISCD. BMD measurement keeps a relatively constant value, so showing excellent repeatability. The Phantom has homogeneous characteristics, but it has limitations to reflect the clinical part including variations in patient's body weight or body fat. As a result, it is believed that quality control using Phantom will be useful to check mis-calibration of the equipment used. A value measured a patient two times with one equipment, and that of double-crossed two equipment are all included within 2SD Value in the Bland - Altman Graph compared results of Group 3 with Group 4. The r value of 0.99 or higher in Linear regression analysis(Regression Analysis) indicated high precision and correlation. Therefore, it revealed that two compatible equipment did not affect in tracking the patients. Regular testing equipment and capabilities of a tester, then appropriate calibration will have to be achieved in order to calculate confidential BMD.

  • PDF

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.

DEVELOPMENT OF STATEWIDE TRUCK TRAFFIC FORECASTING METHOD BY USING LIMITED O-D SURVEY DATA (한정된 O-D조사자료를 이용한 주 전체의 트럭교통예측방법 개발)

  • 박만배
    • Proceedings of the KOR-KST Conference
    • /
    • 1995.02a
    • /
    • pp.101-113
    • /
    • 1995
  • The objective of this research is to test the feasibility of developing a statewide truck traffic forecasting methodology for Wisconsin by using Origin-Destination surveys, traffic counts, classification counts, and other data that are routinely collected by the Wisconsin Department of Transportation (WisDOT). Development of a feasible model will permit estimation of future truck traffic for every major link in the network. This will provide the basis for improved estimation of future pavement deterioration. Pavement damage rises exponentially as axle weight increases, and trucks are responsible for most of the traffic-induced damage to pavement. Consequently, forecasts of truck traffic are critical to pavement management systems. The pavement Management Decision Supporting System (PMDSS) prepared by WisDOT in May 1990 combines pavement inventory and performance data with a knowledge base consisting of rules for evaluation, problem identification and rehabilitation recommendation. Without a r.easonable truck traffic forecasting methodology, PMDSS is not able to project pavement performance trends in order to make assessment and recommendations in the future years. However, none of WisDOT's existing forecasting methodologies has been designed specifically for predicting truck movements on a statewide highway network. For this research, the Origin-Destination survey data avaiiable from WisDOT, including two stateline areas, one county, and five cities, are analyzed and the zone-to'||'&'||'not;zone truck trip tables are developed. The resulting Origin-Destination Trip Length Frequency (00 TLF) distributions by trip type are applied to the Gravity Model (GM) for comparison with comparable TLFs from the GM. The gravity model is calibrated to obtain friction factor curves for the three trip types, Internal-Internal (I-I), Internal-External (I-E), and External-External (E-E). ~oth "macro-scale" calibration and "micro-scale" calibration are performed. The comparison of the statewide GM TLF with the 00 TLF for the macro-scale calibration does not provide suitable results because the available 00 survey data do not represent an unbiased sample of statewide truck trips. For the "micro-scale" calibration, "partial" GM trip tables that correspond to the 00 survey trip tables are extracted from the full statewide GM trip table. These "partial" GM trip tables are then merged and a partial GM TLF is created. The GM friction factor curves are adjusted until the partial GM TLF matches the 00 TLF. Three friction factor curves, one for each trip type, resulting from the micro-scale calibration produce a reasonable GM truck trip model. A key methodological issue for GM. calibration involves the use of multiple friction factor curves versus a single friction factor curve for each trip type in order to estimate truck trips with reasonable accuracy. A single friction factor curve for each of the three trip types was found to reproduce the 00 TLFs from the calibration data base. Given the very limited trip generation data available for this research, additional refinement of the gravity model using multiple mction factor curves for each trip type was not warranted. In the traditional urban transportation planning studies, the zonal trip productions and attractions and region-wide OD TLFs are available. However, for this research, the information available for the development .of the GM model is limited to Ground Counts (GC) and a limited set ofOD TLFs. The GM is calibrated using the limited OD data, but the OD data are not adequate to obtain good estimates of truck trip productions and attractions .. Consequently, zonal productions and attractions are estimated using zonal population as a first approximation. Then, Selected Link based (SELINK) analyses are used to adjust the productions and attractions and possibly recalibrate the GM. The SELINK adjustment process involves identifying the origins and destinations of all truck trips that are assigned to a specified "selected link" as the result of a standard traffic assignment. A link adjustment factor is computed as the ratio of the actual volume for the link (ground count) to the total assigned volume. This link adjustment factor is then applied to all of the origin and destination zones of the trips using that "selected link". Selected link based analyses are conducted by using both 16 selected links and 32 selected links. The result of SELINK analysis by u~ing 32 selected links provides the least %RMSE in the screenline volume analysis. In addition, the stability of the GM truck estimating model is preserved by using 32 selected links with three SELINK adjustments, that is, the GM remains calibrated despite substantial changes in the input productions and attractions. The coverage of zones provided by 32 selected links is satisfactory. Increasing the number of repetitions beyond four is not reasonable because the stability of GM model in reproducing the OD TLF reaches its limits. The total volume of truck traffic captured by 32 selected links is 107% of total trip productions. But more importantly, ~ELINK adjustment factors for all of the zones can be computed. Evaluation of the travel demand model resulting from the SELINK adjustments is conducted by using screenline volume analysis, functional class and route specific volume analysis, area specific volume analysis, production and attraction analysis, and Vehicle Miles of Travel (VMT) analysis. Screenline volume analysis by using four screenlines with 28 check points are used for evaluation of the adequacy of the overall model. The total trucks crossing the screenlines are compared to the ground count totals. L V/GC ratios of 0.958 by using 32 selected links and 1.001 by using 16 selected links are obtained. The %RM:SE for the four screenlines is inversely proportional to the average ground count totals by screenline .. The magnitude of %RM:SE for the four screenlines resulting from the fourth and last GM run by using 32 and 16 selected links is 22% and 31 % respectively. These results are similar to the overall %RMSE achieved for the 32 and 16 selected links themselves of 19% and 33% respectively. This implies that the SELINICanalysis results are reasonable for all sections of the state.Functional class and route specific volume analysis is possible by using the available 154 classification count check points. The truck traffic crossing the Interstate highways (ISH) with 37 check points, the US highways (USH) with 50 check points, and the State highways (STH) with 67 check points is compared to the actual ground count totals. The magnitude of the overall link volume to ground count ratio by route does not provide any specific pattern of over or underestimate. However, the %R11SE for the ISH shows the least value while that for the STH shows the largest value. This pattern is consistent with the screenline analysis and the overall relationship between %RMSE and ground count volume groups. Area specific volume analysis provides another broad statewide measure of the performance of the overall model. The truck traffic in the North area with 26 check points, the West area with 36 check points, the East area with 29 check points, and the South area with 64 check points are compared to the actual ground count totals. The four areas show similar results. No specific patterns in the L V/GC ratio by area are found. In addition, the %RMSE is computed for each of the four areas. The %RMSEs for the North, West, East, and South areas are 92%, 49%, 27%, and 35% respectively, whereas, the average ground counts are 481, 1383, 1532, and 3154 respectively. As for the screenline and volume range analyses, the %RMSE is inversely related to average link volume. 'The SELINK adjustments of productions and attractions resulted in a very substantial reduction in the total in-state zonal productions and attractions. The initial in-state zonal trip generation model can now be revised with a new trip production's trip rate (total adjusted productions/total population) and a new trip attraction's trip rate. Revised zonal production and attraction adjustment factors can then be developed that only reflect the impact of the SELINK adjustments that cause mcreases or , decreases from the revised zonal estimate of productions and attractions. Analysis of the revised production adjustment factors is conducted by plotting the factors on the state map. The east area of the state including the counties of Brown, Outagamie, Shawano, Wmnebago, Fond du Lac, Marathon shows comparatively large values of the revised adjustment factors. Overall, both small and large values of the revised adjustment factors are scattered around Wisconsin. This suggests that more independent variables beyond just 226; population are needed for the development of the heavy truck trip generation model. More independent variables including zonal employment data (office employees and manufacturing employees) by industry type, zonal private trucks 226; owned and zonal income data which are not available currently should be considered. A plot of frequency distribution of the in-state zones as a function of the revised production and attraction adjustment factors shows the overall " adjustment resulting from the SELINK analysis process. Overall, the revised SELINK adjustments show that the productions for many zones are reduced by, a factor of 0.5 to 0.8 while the productions for ~ relatively few zones are increased by factors from 1.1 to 4 with most of the factors in the 3.0 range. No obvious explanation for the frequency distribution could be found. The revised SELINK adjustments overall appear to be reasonable. The heavy truck VMT analysis is conducted by comparing the 1990 heavy truck VMT that is forecasted by the GM truck forecasting model, 2.975 billions, with the WisDOT computed data. This gives an estimate that is 18.3% less than the WisDOT computation of 3.642 billions of VMT. The WisDOT estimates are based on the sampling the link volumes for USH, 8TH, and CTH. This implies potential error in sampling the average link volume. The WisDOT estimate of heavy truck VMT cannot be tabulated by the three trip types, I-I, I-E ('||'&'||'pound;-I), and E-E. In contrast, the GM forecasting model shows that the proportion ofE-E VMT out of total VMT is 21.24%. In addition, tabulation of heavy truck VMT by route functional class shows that the proportion of truck traffic traversing the freeways and expressways is 76.5%. Only 14.1% of total freeway truck traffic is I-I trips, while 80% of total collector truck traffic is I-I trips. This implies that freeways are traversed mainly by I-E and E-E truck traffic while collectors are used mainly by I-I truck traffic. Other tabulations such as average heavy truck speed by trip type, average travel distance by trip type and the VMT distribution by trip type, route functional class and travel speed are useful information for highway planners to understand the characteristics of statewide heavy truck trip patternS. Heavy truck volumes for the target year 2010 are forecasted by using the GM truck forecasting model. Four scenarios are used. Fo~ better forecasting, ground count- based segment adjustment factors are developed and applied. ISH 90 '||'&'||' 94 and USH 41 are used as example routes. The forecasting results by using the ground count-based segment adjustment factors are satisfactory for long range planning purposes, but additional ground counts would be useful for USH 41. Sensitivity analysis provides estimates of the impacts of the alternative growth rates including information about changes in the trip types using key routes. The network'||'&'||'not;based GMcan easily model scenarios with different rates of growth in rural versus . . urban areas, small versus large cities, and in-state zones versus external stations. cities, and in-state zones versus external stations.

  • PDF