• Title/Summary/Keyword: Value problem

Search Result 4,682, Processing Time 0.041 seconds

A Study on the Nutritive Value and Utilization of Powdered Seaweeds (해조의 식용분말화에 관한 연구)

  • Yu, Jong-Yull;Lee, Ki-Yull;Kim, Sook-Hee
    • Journal of Nutrition and Health
    • /
    • v.8 no.1
    • /
    • pp.15-37
    • /
    • 1975
  • I. Subject of the study A study on the nutritive value and utilization of powdered seaweeds. II. Purpose and Importance of the study A. In Korea the shortage of food will be inevitable by the rapidly growing population. It will be very important study to develop a new food from the seaweeds which were not used hitherto for human consumption. B. The several kinds of seaweeds have been used by man in Korea mainly as side-dishes. However, a properly powdered seaweed will enable itself to be a good supplement or mixture to certain cereal flours. C. By adding the powdered seaweed to any cereals which have long been staple foods in this country the two fold benefits; saving of cereals and change of dietary pattern, will be secured. III. Objects and scope of the study A. Objects of the study The objects will come under four items. 1. To develop a powdered seaweed as a new food from the seaweeds which have been not used for human consumption. 2. To evaluate the nutritional quality of the products the analysis for chemical composition and animal feeding experiment will be conducted. 3. Experimental cocking and accepability test will be conducted for the powdered products to evaluate the value as food stuff. 4. Sanitary test and also economical analysis will be conducted for the powdered products. B. Scope of the study 1. Production of seaweed powders Sargassum fulvellum growing in eastern coast and Sargassum patens C.A. in southern coast were used as the material for the powders. These algae, which have been not used for human consumption, were pulverized through the processes of washing, drying, pulverization, etc. 2. Nutritional experiments a. Chemical composition Proximate components (water, protein, fat, cellulose, sugar, ash, salt), minerals (calcium, phosphorus, iron, iodine), vitamins (A, $B_1,\;B_2$ niacin, C) and amino acids were analyzed for the seaweed powders. b. Animal feeding experiment Weaning 160 rats (80 male and 80 female rats) were used as experimental animals, dividing them into 16 groups, 10 rats each group. Each group was fed for 12 weeks on cereal diet (Wheat flour, rice powder, barley powder, potato powder, corn flour) with the supplementary levels of 5%, 10%, 15%, 20% and 30% of the seaweed powder. After the feeding the growth, feed efficiency ratio, protain efficiency ratio and ,organs weights were checked and urine analysis, feces analysis and serum analysis were also conducted. 3. Experimental cooking and acceptability test a. Several basic studies were conducted to find the characteristics of the seaweed powder. b. 17 kinds of Korean dishes and 9 kinds of foreign dishes were prepared with cereal flours (wheat, rice, barley, potato, corn) with the supplementary levels of 5%, 10%, 15%, 20% and 30% of the seaweed powder. c. Acceptability test for the dishes was conducted according to plank's Form. 4. Sanitary test The heavy metals (Cd, Pb, As, Hg) in the seaweed powders were determined. 5. Economical analysis The retail price of the seaweed powder was compared with those of other cereals in the market. And also economical analysis was made from the nutritional point of view, calculating the body weight gained in grams per unit price of each feeding diet. IV. Results of the study and the suggestion for application A. Chemical composition 1. There is no any big difference in proximate components between powders of Sargassum fulvellum in eastern coast and Sargassum patens C.A. in southern coast. Seasonal difference is also not significant. Higher levels of protein, cellulose, ash and salt were found in the powders compared with common cereal foods. 2. The levels of calcium (Ca) and iron (Fe) in the powders were significantly higher than common cereal foods and also rich in iodine (I). Existence of vitamin A and vitamin C in the Powders is different point from cereal foods. Vitamin $B_1\;and\;B_2$ are also relatively rich in the powders.'Vitamin A in ·Sargassum fulvellum is high and the levels of some minerals and vitamins are seemed4 to be some influenced by seasons. 3. In the amino acid composition methionine, isoleucine, Iysine and valine are limiting amino acids. The protein qualities of Sargassum fulvellum and Sargassum patens C.A. are seemed to be .almost same and generally ·good. Seasonal difference in amino acid composition was found. B. Animal feeding experiment 1. The best growth was found at.10% supplemental level of the seaweed Powder and lower growth rate was shown at 30% level. 2. It was shown that 15% supplemental level of the Seaweed powder seems to fulfil, to some extent the mineral requirement of the animals. 3. No any changes were found in organs development except that, in kidney, there found decreasing in weight by increasing the supplemental level of the seaweed powder. 4. There is no any significant changes in nitrogen retention, serum cholesterol, serum calcium and urinary calcium in each supplemental level of the seaweed powder. 5. In animal feeding experiment it was concluded that $5%{\sim}15%$ levels supplementation of the seaweed powder are possible. C. Experimental cooking and acceptability test 1. The seaweed powder showed to be utilized more excellently in foreign cookings than in Korean cookings. Higher supplemental level of seaweed was passible in foreign cookings. 2. Hae-Jo-Kang and Jeon-Byung were more excellent than Song-Pyun, wheat cake, Soo-Je-Bee and wheat noodle. Hae-Je-Kang was excellent in its quality even as high as 5% supplemental level. 3. The higher levels of supplementation were used the more sticky cooking products were obtained. Song-Pyun and wheat cake were palatable and lustrous in 2% supplementation level. 4. In drop cookie the higher levels of supplementation, the more crisp product was obtained, compared with other cookies. 5. Corn cake, thin rice gruel, rice gruel and potato Jeon-Byung were more excellent in their quality than potato Man-Doo and potato noodle. Corn cake, thin rice gruel and rice gruel were excellent even as high as 5% supplementation level. 6. In several cooking Porducts some seaweed-oder was perceived in case of 3% or more levels of supplementation. This may be much diminished by the use of proper condiments. D. Sanitary test It seems that there is no any heavy metals (Cd, Pb, As, Hg) problem in these seaweed Powders in case these Powders are used as supplements to any cereal flours E. Economical analysis The price of the seaweed powder is lower than those of other cereals and that may be more lowered when mass production of the seaweed powder is made in future. The supplement of the seaweed powder to any cereals is also economical with the criterion of animal growth rate. F. It is recommended that these seaweed powders should be developed and used as supplement to any cereal flours or used as other food material. By doing so, both saving of cereals and improvement of individual's nutrition will greatly be achieved. It is also recommended that the feeding experiment for men would be conducted in future.

  • PDF

Analysis of HBeAg and HBV DNA Detection in Hepatitis B Patients Treated with Antiviral Therapy (항 바이러스 치료중인 B형 간염환자에서 HBeAg 및 HBV DNA 검출에 관한 분석)

  • Cheon, Jun Hong;Chae, Hong Ju;Park, Mi Sun;Lim, Soo Yeon;Yoo, Seon Hee;Lee, Sun Ho
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.23 no.1
    • /
    • pp.35-39
    • /
    • 2019
  • Purpose Hepatitis B virus (hepatitis B virus, HBV) infection is a worldwide major public health problem and it is known as a major cause of chronic hepatitis, liver cirrhosis and liver cancer. And serologic tests of hepatitis B virus is essential for diagnosing and treating these diseases. In addition, with the development of molecular diagnostics, the detection of HBV DNA in serum diagnoses HBV infection and is recognized as an important indicator for the antiviral agent treatment response assessment. We performed HBeAg assay using Immunoradiometric assay (IRMA) and Chemiluminescent Microparticle Immunoassay (CMIA) in hepatitis B patients treated with antiviral agents. The detection rate of HBV DNA in serum was measured and compared by RT-PCR (Real Time - Polymerase Chain Reaction) method Materials and Methods HBeAg serum examination and HBV DNA quantification test were conducted on 270 hepatitis B patients undergoing anti-virus treatment after diagnosis of hepatitis B virus infection. Two serologic tests (IRMA, CMIA) with different detection principles were applied for the HBeAg serum test. Serum HBV DNA was quantitatively measured by real-time polymerase chain reaction (RT-PCR) using the Abbott m2000 System. Results The detection rate of HBeAg was 24.1% (65/270) for IRMA and 82.2% (222/270) for CMIA. Detection rate of serum HBV DNA by real-time RT-PCR is 29.3% (79/270). The measured amount of serum HBV DNA concentration is $4.8{\times}10^7{\pm}1.9{\times}10^8IU/mL$($mean{\pm}SD$). The minimum value is 16IU/mL, the maximum value is $1.0{\times}10^9IU/mL$, and the reference value for quantitative detection limit is 15IU/mL. The detection rates and concentrations of HBV DNA by group according to the results of HBeAg serological (IRMA, CMIA)tests were as follows. 1) Group I (IRMA negative, CMIA positive, N = 169), HBV DNA detection rate of 17.7% (30/169), $6.8{\times}10^5{\pm}1.9{\times}10^6IU/mL$ 2) Group II (IRMA positive, CMIA positive, N = 53), HBV DNA detection rate 62.3% (33/53), $1.1{\times}10^8{\pm}2.8{\times}10^8IU/mL$ 3) Group III (IRMA negative, CMIA negative, N = 36), HBV DNA detection rate 36.1% (13/36), $3.0{\times}10^5{\pm}1.1{\times}10^6IU/mL$ 4) Group IV(IRMA positive, CMIA negative, N = 12), HBV DNA detection rate 25% (3/12), $1.3{\times}10^3{\pm}1.1{\times}10^3IU/mL$ Conclusion HBeAg detection rate according to the serological test showed a large difference. This difference is considered for a number of reasons such as characteristics of the Ab used for assay kit and epitope, HBV of genotype. Detection rate and the concentration of the group-specific HBV DNA classified serologic results confirmed the high detection rate and the concentration in Group II (IRMA-positive, CMIA positive, N = 53).

An Examination into the Illegal Trade of Cultural Properties (문화재(文化財)의 국제적 불법 거래(不法 去來)에 관한 고찰)

  • Cho, Boo-Keun
    • Korean Journal of Heritage: History & Science
    • /
    • v.37
    • /
    • pp.371-405
    • /
    • 2004
  • International circulation of cultural assets involves numerous countries thereby making an approach based on international law essential to resolving this problem. Since the end of the $2^{nd}$ World War, as the value of cultural assets evolved from material value to moral and ethical values, with emphasis on establishing national identities, newly independent nations and former colonial states took issue with ownership of cultural assets which led to the need for international cooperation and statutory provisions for the return of cultural assets. UNESCO's 1954 "Convention for the Protection of Cultural Property in the Event of Armed Conflict" as preparatory measures for the protection of cultural assets, the 1970 "Convention on the Means of Prohibiting and Preventing the Illicit Import and Transfer of Ownership of Cultural Property" to regulate transfer of cultural assets, and the 1995 "Unidroit Convention on Stolen or Illegally Exported Cultural Objects" which required the return of illegally acquired cultural property are examples of international agreements established on illegal transfers of cultural assets. In addition, the UN agency UNESCO established the Division of Cultural Heritage to oversee cultural assets related matters, and the UN since its 1973 resolution 3187, has continued to demonstrate interest in protection of cultural assets. The resolution 3187 affirms the return of cultural assets to the country of origin, advises on preventing illegal transfers of works of art and cultural assets, advises cataloguing cultural assets within the respective countries and, conclusively, recommends becoming a member of UNESCO, composing a forum for international cooperation. Differences in defining cultural assets pose a limitation on international agreements. While the 1954 Convention states that cultural assets are not limited to movable property and includes immovable property, the 1970 Convention's objective of 'Prohibiting and preventing the illicit import, export and transfer of ownership of cultural property' effectively limits the subject to tangible movable cultural property. The 1995 Convention also has tangible movable cultural property as its subject. On this point, the two conventions demonstrate distinction from the 1954 Convention and the 1972 Convention that focuses on immovable cultural property and natural property. The disparity in defining cultural property is due to the object and purpose of the convention and does not reflect an inherent divergence. In the case of Korea, beginning with the 1866 French invasion, 36 years of Japanese colonial rule, military rule and period of economic development caused outflow of numerous cultural assets to foreign countries. Of course, it is neither possible nor necessary to have all of these cultural properties returned, but among those that have significant value in establishing cultural and historical identity or those that have been taken symbolically as a demonstration of occupational rule can cause issues in their return. In these cases, the 1954 Convention and the ratification of the first legislation must be actively considered. In the return of cultural property, if the illicit acquisition is the core issue, it is a simple matter of following the international accords, while if it rises to the level of diplomatic discussions, it will become a political issue. In that case, the country requesting the return must convince the counterpart country. Realizing a response to the earnest need for preventing illicit trading of cultural assets will require extensive national and civic societal efforts in the East Asian area to overcome its current deficiencies. The most effective way to prevent illicit trading of cultural property is rapid circulation of information between Interpol member countries, which will require development of an internet based communication system as well as more effective deployment of legislation to prevent trading of illicitly acquired cultural property, subscription to international conventions and cataloguing collections.

A Study of Anomaly Detection for ICT Infrastructure using Conditional Multimodal Autoencoder (ICT 인프라 이상탐지를 위한 조건부 멀티모달 오토인코더에 관한 연구)

  • Shin, Byungjin;Lee, Jonghoon;Han, Sangjin;Park, Choong-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.57-73
    • /
    • 2021
  • Maintenance and prevention of failure through anomaly detection of ICT infrastructure is becoming important. System monitoring data is multidimensional time series data. When we deal with multidimensional time series data, we have difficulty in considering both characteristics of multidimensional data and characteristics of time series data. When dealing with multidimensional data, correlation between variables should be considered. Existing methods such as probability and linear base, distance base, etc. are degraded due to limitations called the curse of dimensions. In addition, time series data is preprocessed by applying sliding window technique and time series decomposition for self-correlation analysis. These techniques are the cause of increasing the dimension of data, so it is necessary to supplement them. The anomaly detection field is an old research field, and statistical methods and regression analysis were used in the early days. Currently, there are active studies to apply machine learning and artificial neural network technology to this field. Statistically based methods are difficult to apply when data is non-homogeneous, and do not detect local outliers well. The regression analysis method compares the predictive value and the actual value after learning the regression formula based on the parametric statistics and it detects abnormality. Anomaly detection using regression analysis has the disadvantage that the performance is lowered when the model is not solid and the noise or outliers of the data are included. There is a restriction that learning data with noise or outliers should be used. The autoencoder using artificial neural networks is learned to output as similar as possible to input data. It has many advantages compared to existing probability and linear model, cluster analysis, and map learning. It can be applied to data that does not satisfy probability distribution or linear assumption. In addition, it is possible to learn non-mapping without label data for teaching. However, there is a limitation of local outlier identification of multidimensional data in anomaly detection, and there is a problem that the dimension of data is greatly increased due to the characteristics of time series data. In this study, we propose a CMAE (Conditional Multimodal Autoencoder) that enhances the performance of anomaly detection by considering local outliers and time series characteristics. First, we applied Multimodal Autoencoder (MAE) to improve the limitations of local outlier identification of multidimensional data. Multimodals are commonly used to learn different types of inputs, such as voice and image. The different modal shares the bottleneck effect of Autoencoder and it learns correlation. In addition, CAE (Conditional Autoencoder) was used to learn the characteristics of time series data effectively without increasing the dimension of data. In general, conditional input mainly uses category variables, but in this study, time was used as a condition to learn periodicity. The CMAE model proposed in this paper was verified by comparing with the Unimodal Autoencoder (UAE) and Multi-modal Autoencoder (MAE). The restoration performance of Autoencoder for 41 variables was confirmed in the proposed model and the comparison model. The restoration performance is different by variables, and the restoration is normally well operated because the loss value is small for Memory, Disk, and Network modals in all three Autoencoder models. The process modal did not show a significant difference in all three models, and the CPU modal showed excellent performance in CMAE. ROC curve was prepared for the evaluation of anomaly detection performance in the proposed model and the comparison model, and AUC, accuracy, precision, recall, and F1-score were compared. In all indicators, the performance was shown in the order of CMAE, MAE, and AE. Especially, the reproduction rate was 0.9828 for CMAE, which can be confirmed to detect almost most of the abnormalities. The accuracy of the model was also improved and 87.12%, and the F1-score was 0.8883, which is considered to be suitable for anomaly detection. In practical aspect, the proposed model has an additional advantage in addition to performance improvement. The use of techniques such as time series decomposition and sliding windows has the disadvantage of managing unnecessary procedures; and their dimensional increase can cause a decrease in the computational speed in inference.The proposed model has characteristics that are easy to apply to practical tasks such as inference speed and model management.

The Effect of Price Promotional Information about Brand on Consumer's Quality Perception: Conditioning on Pretrial Brand (품패개격촉소신식대소비자질량인지적영향(品牌价格促销信息对消费者质量认知的影响))

  • Lee, Min-Hoon;Lim, Hang-Seop
    • Journal of Global Scholars of Marketing Science
    • /
    • v.19 no.3
    • /
    • pp.17-27
    • /
    • 2009
  • Price promotion typically reduces the price for a given quantity or increases the quantity available at the same price, thereby enhancing value and creating an economic incentive to purchase. It often is used to encourage product or service trial among nonusers of products or services. Thus, it is important to understand the effects of price promotions on quality perception made by consumer who do not have prior experience with the promoted brand. However, if consumers associate a price promotion itself with inferior brand quality, the promotion may not achieve the sales increase the economic incentives otherwise might have produced. More specifically, low qualitative perception through price promotion will undercut the economic and psychological incentives and reduce the likelihood of purchase. Thus, it is important for marketers to understand how price promotional informations about a brand have impact on consumer's unfavorable quality perception of the brand. Previous literatures on the effects of price promotions on quality perception reveal inconsistent explanations. Some focused on the unfavorable effect of price promotion on consumer's perception. But others showed that price promotions didn't raise unfavorable perception on the brand. Prior researches found these inconsistent results related to the timing of the price promotion's exposure and quality evaluation relative to trial. And, whether the consumer has been experienced with the product promotions in the past or not may moderate the effects. A few studies considered differences among product categories as fundamental factors. The purpose of this research is to investigate the effect of price promotional informations on consumer's unfavorable quality perception under the different conditions. The author controlled the timing of the promotional exposure and varied past promotional patterns and information presenting patterns. Unlike previous researches, the author examined the effects of price promotions setting limit to pretrial situation by controlling potentially moderating effects of prior personal experience with the brand. This manipulations enable to resolve possible controversies in relation to this issue. And this manipulation is meaningful for the work sector. Price promotion is not only used to target existing consumers but also to encourage product or service trial among nonusers of products or services. Thus, it is important for marketers to understand how price promotional informations about a brand have impact on consumer's unfavorable quality perception of the brand. If consumers associate a price promotion itself with inferior quality about unused brand, the promotion may not achieve the sales increase the economic incentives otherwise might have produced. In addition, if the price promotion ends, the consumer that have purchased that certain brand will likely to display sharply decreased repurchasing behavior. Through a literature review, hypothesis 1 was set as follows to investigate the adjustive effect of past price promotion on quality perception made by consumers; The influence that price promotion of unused brand have on quality perception made by consumers will be adjusted by past price promotion activity of the brand. In other words, a price promotion of an unused brand that have not done a price promotion in the past will have a unfavorable effect on quality perception made by consumer. Hypothesis 2-1 was set as follows : When an unused brand undertakes price promotion for the first time, the information presenting pattern of price promotion will have an effect on the consumer's attribution for the cause of the price promotion. Hypothesis 2-2 was set as follows : The more consumer dispositionally attribute the cause of price promotion, the more unfavorable the quality perception made by consumer will be. Through test 1, the subjects were given a brief explanation of the product and the brand before they were provided with a $2{\times}2$ factorial design that has 4 patterns of price promotion (presence or absence of past price promotion * presence or absence of current price promotion) and the explanation describing the price promotion pattern of each cell. Then the perceived quality of imaginary brand WAVEX was evaluated in the scale of 7. The reason tennis racket was chosen is because the selected product group must have had almost no past price promotions to eliminate the influence of average frequency of promotion on the value of price promotional information as Raghubir and Corfman (1999) pointed out. Test 2 was also carried out on students of the same management faculty of test 1 with tennis racket as the product group. As with test 1, subjects with average familiarity for the product group and low familiarity for the brand was selected. Each subjects were assigned to one of the two cells representing two different information presenting patterns of price promotion of WAVEX (case where the reason behind price promotion was provided/case where the reason behind price promotion was not provided). Subjects looked at each promotional information before evaluating the perceived quality of the brand WAVEX in the scale of 7. The effect of price promotion for unfamiliar pretrial brand on consumer's perceived quality was proved to be moderated with the presence or absence of past price promotion. The consistency with past promotional behavior is important variable that makes unfavorable effect on brand evaluations get worse. If the price promotion for the brand has never been carried out before, price promotion activity may have more unfavorable effects on consumer's quality perception. Second, when the price promotion of unfamiliar pretrial brand was executed for the first time, presenting method of informations has impact on consumer's attribution for the cause of firm's promotion. And the unfavorable effect of quality perception is higher when the consumer does dispositional attribution comparing with situational attribution. Unlike the previous studies where the main focus was the absence or presence of favorable or unfavorable motivation from situational/dispositional attribution, the focus of this study was exaus ing the fact that a situational attribution can be inferred even if the consumer employs a dispositional attribution on the price promotional behavior, if the company provides a persuasive reason. Such approach, in academic perspectih sis a large significance in that it explained the anchoring and adjng ch approcedures by applying it to a non-mathematical problem unlike the previous studies where it wis ionaly explained by applying it to a mathematical problem. In other wordn, there is a highrspedency tmatispositionally attribute other's behaviors according to the fuedach aal attribution errors and when this is applied to the situation of price promotions, we can infer that consumers are likely tmatispositionally attribute the company's price promotion behaviors. Ha ever, even ueder these circumstances, the company can adjng the consumer's anchoring tmareduce the po wibiliute thdispositional attribution. Furthermore, unlike majority of previous researches on short/long-term effects of price promotion that only considered the effect of price promotions on consumer's purchasing behaviors, this research measured the effect on perceived quality, one of man elements that affects the purchasing behavior of consumers. These results carry useful implications for the work sector. A guideline of effectively providing promotional informations for a new brand can be suggested through the outcomes of this research. If the brand is to avoid false implications such as inferior quality while implementing a price promotion strategy, it must provide a clear and acceptable reasons behind the promotion. Especially it is more important for the company with no past price promotion to provide a clear reason. An inconsistent behavior can be the cause of consumer's distrust and anxiety. This is also one of the most important factor of risk of endless price wars. Price promotions without prior notice can buy doubt from consumers not market share.

  • PDF

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.

Methods for Integration of Documents using Hierarchical Structure based on the Formal Concept Analysis (FCA 기반 계층적 구조를 이용한 문서 통합 기법)

  • Kim, Tae-Hwan;Jeon, Ho-Cheol;Choi, Joong-Min
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.3
    • /
    • pp.63-77
    • /
    • 2011
  • The World Wide Web is a very large distributed digital information space. From its origins in 1991, the web has grown to encompass diverse information resources as personal home pasges, online digital libraries and virtual museums. Some estimates suggest that the web currently includes over 500 billion pages in the deep web. The ability to search and retrieve information from the web efficiently and effectively is an enabling technology for realizing its full potential. With powerful workstations and parallel processing technology, efficiency is not a bottleneck. In fact, some existing search tools sift through gigabyte.syze precompiled web indexes in a fraction of a second. But retrieval effectiveness is a different matter. Current search tools retrieve too many documents, of which only a small fraction are relevant to the user query. Furthermore, the most relevant documents do not nessarily appear at the top of the query output order. Also, current search tools can not retrieve the documents related with retrieved document from gigantic amount of documents. The most important problem for lots of current searching systems is to increase the quality of search. It means to provide related documents or decrease the number of unrelated documents as low as possible in the results of search. For this problem, CiteSeer proposed the ACI (Autonomous Citation Indexing) of the articles on the World Wide Web. A "citation index" indexes the links between articles that researchers make when they cite other articles. Citation indexes are very useful for a number of purposes, including literature search and analysis of the academic literature. For details of this work, references contained in academic articles are used to give credit to previous work in the literature and provide a link between the "citing" and "cited" articles. A citation index indexes the citations that an article makes, linking the articleswith the cited works. Citation indexes were originally designed mainly for information retrieval. The citation links allow navigating the literature in unique ways. Papers can be located independent of language, and words in thetitle, keywords or document. A citation index allows navigation backward in time (the list of cited articles) and forwardin time (which subsequent articles cite the current article?) But CiteSeer can not indexes the links between articles that researchers doesn't make. Because it indexes the links between articles that only researchers make when they cite other articles. Also, CiteSeer is not easy to scalability. Because CiteSeer can not indexes the links between articles that researchers doesn't make. All these problems make us orient for designing more effective search system. This paper shows a method that extracts subject and predicate per each sentence in documents. A document will be changed into the tabular form that extracted predicate checked value of possible subject and object. We make a hierarchical graph of a document using the table and then integrate graphs of documents. The graph of entire documents calculates the area of document as compared with integrated documents. We mark relation among the documents as compared with the area of documents. Also it proposes a method for structural integration of documents that retrieves documents from the graph. It makes that the user can find information easier. We compared the performance of the proposed approaches with lucene search engine using the formulas for ranking. As a result, the F.measure is about 60% and it is better as about 15%.

A Time Series Graph based Convolutional Neural Network Model for Effective Input Variable Pattern Learning : Application to the Prediction of Stock Market (효과적인 입력변수 패턴 학습을 위한 시계열 그래프 기반 합성곱 신경망 모형: 주식시장 예측에의 응용)

  • Lee, Mo-Se;Ahn, Hyunchul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.167-181
    • /
    • 2018
  • Over the past decade, deep learning has been in spotlight among various machine learning algorithms. In particular, CNN(Convolutional Neural Network), which is known as the effective solution for recognizing and classifying images or voices, has been popularly applied to classification and prediction problems. In this study, we investigate the way to apply CNN in business problem solving. Specifically, this study propose to apply CNN to stock market prediction, one of the most challenging tasks in the machine learning research. As mentioned, CNN has strength in interpreting images. Thus, the model proposed in this study adopts CNN as the binary classifier that predicts stock market direction (upward or downward) by using time series graphs as its inputs. That is, our proposal is to build a machine learning algorithm that mimics an experts called 'technical analysts' who examine the graph of past price movement, and predict future financial price movements. Our proposed model named 'CNN-FG(Convolutional Neural Network using Fluctuation Graph)' consists of five steps. In the first step, it divides the dataset into the intervals of 5 days. And then, it creates time series graphs for the divided dataset in step 2. The size of the image in which the graph is drawn is $40(pixels){\times}40(pixels)$, and the graph of each independent variable was drawn using different colors. In step 3, the model converts the images into the matrices. Each image is converted into the combination of three matrices in order to express the value of the color using R(red), G(green), and B(blue) scale. In the next step, it splits the dataset of the graph images into training and validation datasets. We used 80% of the total dataset as the training dataset, and the remaining 20% as the validation dataset. And then, CNN classifiers are trained using the images of training dataset in the final step. Regarding the parameters of CNN-FG, we adopted two convolution filters ($5{\times}5{\times}6$ and $5{\times}5{\times}9$) in the convolution layer. In the pooling layer, $2{\times}2$ max pooling filter was used. The numbers of the nodes in two hidden layers were set to, respectively, 900 and 32, and the number of the nodes in the output layer was set to 2(one is for the prediction of upward trend, and the other one is for downward trend). Activation functions for the convolution layer and the hidden layer were set to ReLU(Rectified Linear Unit), and one for the output layer set to Softmax function. To validate our model - CNN-FG, we applied it to the prediction of KOSPI200 for 2,026 days in eight years (from 2009 to 2016). To match the proportions of the two groups in the independent variable (i.e. tomorrow's stock market movement), we selected 1,950 samples by applying random sampling. Finally, we built the training dataset using 80% of the total dataset (1,560 samples), and the validation dataset using 20% (390 samples). The dependent variables of the experimental dataset included twelve technical indicators popularly been used in the previous studies. They include Stochastic %K, Stochastic %D, Momentum, ROC(rate of change), LW %R(Larry William's %R), A/D oscillator(accumulation/distribution oscillator), OSCP(price oscillator), CCI(commodity channel index), and so on. To confirm the superiority of CNN-FG, we compared its prediction accuracy with the ones of other classification models. Experimental results showed that CNN-FG outperforms LOGIT(logistic regression), ANN(artificial neural network), and SVM(support vector machine) with the statistical significance. These empirical results imply that converting time series business data into graphs and building CNN-based classification models using these graphs can be effective from the perspective of prediction accuracy. Thus, this paper sheds a light on how to apply deep learning techniques to the domain of business problem solving.

A Study on the Clustering Method of Row and Multiplex Housing in Seoul Using K-Means Clustering Algorithm and Hedonic Model (K-Means Clustering 알고리즘과 헤도닉 모형을 활용한 서울시 연립·다세대 군집분류 방법에 관한 연구)

  • Kwon, Soonjae;Kim, Seonghyeon;Tak, Onsik;Jeong, Hyeonhee
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.3
    • /
    • pp.95-118
    • /
    • 2017
  • Recent centrally the downtown area, the transaction between the row housing and multiplex housing is activated and platform services such as Zigbang and Dabang are growing. The row housing and multiplex housing is a blind spot for real estate information. Because there is a social problem, due to the change in market size and information asymmetry due to changes in demand. Also, the 5 or 25 districts used by the Seoul Metropolitan Government or the Korean Appraisal Board(hereafter, KAB) were established within the administrative boundaries and used in existing real estate studies. This is not a district classification for real estate researches because it is zoned urban planning. Based on the existing study, this study found that the city needs to reset the Seoul Metropolitan Government's spatial structure in estimating future housing prices. So, This study attempted to classify the area without spatial heterogeneity by the reflected the property price characteristics of row housing and Multiplex housing. In other words, There has been a problem that an inefficient side has arisen due to the simple division by the existing administrative district. Therefore, this study aims to cluster Seoul as a new area for more efficient real estate analysis. This study was applied to the hedonic model based on the real transactions price data of row housing and multiplex housing. And the K-Means Clustering algorithm was used to cluster the spatial structure of Seoul. In this study, data onto real transactions price of the Seoul Row housing and Multiplex Housing from January 2014 to December 2016, and the official land value of 2016 was used and it provided by Ministry of Land, Infrastructure and Transport(hereafter, MOLIT). Data preprocessing was followed by the following processing procedures: Removal of underground transaction, Price standardization per area, Removal of Real transaction case(above 5 and below -5). In this study, we analyzed data from 132,707 cases to 126,759 data through data preprocessing. The data analysis tool used the R program. After data preprocessing, data model was constructed. Priority, the K-means Clustering was performed. In addition, a regression analysis was conducted using Hedonic model and it was conducted a cosine similarity analysis. Based on the constructed data model, we clustered on the basis of the longitude and latitude of Seoul and conducted comparative analysis of existing area. The results of this study indicated that the goodness of fit of the model was above 75 % and the variables used for the Hedonic model were significant. In other words, 5 or 25 districts that is the area of the existing administrative area are divided into 16 districts. So, this study derived a clustering method of row housing and multiplex housing in Seoul using K-Means Clustering algorithm and hedonic model by the reflected the property price characteristics. Moreover, they presented academic and practical implications and presented the limitations of this study and the direction of future research. Academic implication has clustered by reflecting the property price characteristics in order to improve the problems of the areas used in the Seoul Metropolitan Government, KAB, and Existing Real Estate Research. Another academic implications are that apartments were the main study of existing real estate research, and has proposed a method of classifying area in Seoul using public information(i.e., real-data of MOLIT) of government 3.0. Practical implication is that it can be used as a basic data for real estate related research on row housing and multiplex housing. Another practical implications are that is expected the activation of row housing and multiplex housing research and, that is expected to increase the accuracy of the model of the actual transaction. The future research direction of this study involves conducting various analyses to overcome the limitations of the threshold and indicates the need for deeper research.

Sentiment Analysis of Movie Review Using Integrated CNN-LSTM Mode (CNN-LSTM 조합모델을 이용한 영화리뷰 감성분석)

  • Park, Ho-yeon;Kim, Kyoung-jae
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.141-154
    • /
    • 2019
  • Rapid growth of internet technology and social media is progressing. Data mining technology has evolved to enable unstructured document representations in a variety of applications. Sentiment analysis is an important technology that can distinguish poor or high-quality content through text data of products, and it has proliferated during text mining. Sentiment analysis mainly analyzes people's opinions in text data by assigning predefined data categories as positive and negative. This has been studied in various directions in terms of accuracy from simple rule-based to dictionary-based approaches using predefined labels. In fact, sentiment analysis is one of the most active researches in natural language processing and is widely studied in text mining. When real online reviews aren't available for others, it's not only easy to openly collect information, but it also affects your business. In marketing, real-world information from customers is gathered on websites, not surveys. Depending on whether the website's posts are positive or negative, the customer response is reflected in the sales and tries to identify the information. However, many reviews on a website are not always good, and difficult to identify. The earlier studies in this research area used the reviews data of the Amazon.com shopping mal, but the research data used in the recent studies uses the data for stock market trends, blogs, news articles, weather forecasts, IMDB, and facebook etc. However, the lack of accuracy is recognized because sentiment calculations are changed according to the subject, paragraph, sentiment lexicon direction, and sentence strength. This study aims to classify the polarity analysis of sentiment analysis into positive and negative categories and increase the prediction accuracy of the polarity analysis using the pretrained IMDB review data set. First, the text classification algorithm related to sentiment analysis adopts the popular machine learning algorithms such as NB (naive bayes), SVM (support vector machines), XGboost, RF (random forests), and Gradient Boost as comparative models. Second, deep learning has demonstrated discriminative features that can extract complex features of data. Representative algorithms are CNN (convolution neural networks), RNN (recurrent neural networks), LSTM (long-short term memory). CNN can be used similarly to BoW when processing a sentence in vector format, but does not consider sequential data attributes. RNN can handle well in order because it takes into account the time information of the data, but there is a long-term dependency on memory. To solve the problem of long-term dependence, LSTM is used. For the comparison, CNN and LSTM were chosen as simple deep learning models. In addition to classical machine learning algorithms, CNN, LSTM, and the integrated models were analyzed. Although there are many parameters for the algorithms, we examined the relationship between numerical value and precision to find the optimal combination. And, we tried to figure out how the models work well for sentiment analysis and how these models work. This study proposes integrated CNN and LSTM algorithms to extract the positive and negative features of text analysis. The reasons for mixing these two algorithms are as follows. CNN can extract features for the classification automatically by applying convolution layer and massively parallel processing. LSTM is not capable of highly parallel processing. Like faucets, the LSTM has input, output, and forget gates that can be moved and controlled at a desired time. These gates have the advantage of placing memory blocks on hidden nodes. The memory block of the LSTM may not store all the data, but it can solve the CNN's long-term dependency problem. Furthermore, when LSTM is used in CNN's pooling layer, it has an end-to-end structure, so that spatial and temporal features can be designed simultaneously. In combination with CNN-LSTM, 90.33% accuracy was measured. This is slower than CNN, but faster than LSTM. The presented model was more accurate than other models. In addition, each word embedding layer can be improved when training the kernel step by step. CNN-LSTM can improve the weakness of each model, and there is an advantage of improving the learning by layer using the end-to-end structure of LSTM. Based on these reasons, this study tries to enhance the classification accuracy of movie reviews using the integrated CNN-LSTM model.