• Title/Summary/Keyword: Initial set

Search Result 1,300, Processing Time 0.028 seconds

Establishment of Test Conditions and Interlaboratory Comparison Study of Neuro-2a Assay for Saxitoxin Detection (Saxitoxin 검출을 위한 Neuro-2a 시험법 조건 확립 및 실험실 간 변동성 비교 연구)

  • Youngjin Kim;Jooree Seo;Jun Kim;Jeong-In Park;Jong Hee Kim;Hyun Park;Young-Seok Han;Youn-Jung Kim
    • Journal of Marine Life Science
    • /
    • v.9 no.1
    • /
    • pp.9-21
    • /
    • 2024
  • Paralytic shellfish poisoning (PSP) including Saxitoxin (STX) is caused by harmful algae, and poisoning occurs when the contaminated seafood is consumed. The mouse bioassay (MBA), a standard test method for detecting PSP, is being sanctioned in many countries due to its low detection limit and the animal concerns. An alternative to the MBA is the Neuro-2a cell-based assay. This study aimed to establish various test conditions for Neuro-2a assay, including cell density, culture conditions, and STX treatment conditions, to suit the domestic laboratory environment. As a result, the initial cell density was set to 40,000 cells/well and the incubation time to 24 hours. Additionally, the concentration of Ouabain and Veratridine (O/V) was set to 500/50 μM, at which most cells died. In this study, we identified eight concentrations of STX, ranging from 368 to 47,056 fg/μl, which produced an S-shaped dose-response curve when treated with O/V. Through inter-laboratory variability comparison of the Neuro-2a assay, we established five Quality Control Criteria to verify the appropriateness of the experiments and six Data Criteria (Top and Bottom OD, EC50, EC20, Hill slop, and R2 of graph) to determine the reliability of the experimental data. The Neuro-2a assay conducted under the established conditions showed an EC50 value of approximately 1,800~3,500 fg/μl. The intra- & inter-lab variability comparison results showed that the coefficients of variation (CVs) for the Quality Control and Data values ranged from 1.98% to 29.15%, confirming the reproducibility of the experiments. This study presented Quality Control Criteria and Data Criteria to assess the appropriateness of the experiments and confirmed the excellent repeatability and reproducibility of the Neuro-2a assay. To apply the Neuro-2a assay as an alternative method for detecting PSP in domestic seafood, it is essential to establish a toxin extraction method from seafood and toxin quantification methods, and perform correlation analysis with MBA and instrumental analysis methods.

Stock Price Prediction by Utilizing Category Neutral Terms: Text Mining Approach (카테고리 중립 단어 활용을 통한 주가 예측 방안: 텍스트 마이닝 활용)

  • Lee, Minsik;Lee, Hong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.123-138
    • /
    • 2017
  • Since the stock market is driven by the expectation of traders, studies have been conducted to predict stock price movements through analysis of various sources of text data. In order to predict stock price movements, research has been conducted not only on the relationship between text data and fluctuations in stock prices, but also on the trading stocks based on news articles and social media responses. Studies that predict the movements of stock prices have also applied classification algorithms with constructing term-document matrix in the same way as other text mining approaches. Because the document contains a lot of words, it is better to select words that contribute more for building a term-document matrix. Based on the frequency of words, words that show too little frequency or importance are removed. It also selects words according to their contribution by measuring the degree to which a word contributes to correctly classifying a document. The basic idea of constructing a term-document matrix was to collect all the documents to be analyzed and to select and use the words that have an influence on the classification. In this study, we analyze the documents for each individual item and select the words that are irrelevant for all categories as neutral words. We extract the words around the selected neutral word and use it to generate the term-document matrix. The neutral word itself starts with the idea that the stock movement is less related to the existence of the neutral words, and that the surrounding words of the neutral word are more likely to affect the stock price movements. And apply it to the algorithm that classifies the stock price fluctuations with the generated term-document matrix. In this study, we firstly removed stop words and selected neutral words for each stock. And we used a method to exclude words that are included in news articles for other stocks among the selected words. Through the online news portal, we collected four months of news articles on the top 10 market cap stocks. We split the news articles into 3 month news data as training data and apply the remaining one month news articles to the model to predict the stock price movements of the next day. We used SVM, Boosting and Random Forest for building models and predicting the movements of stock prices. The stock market opened for four months (2016/02/01 ~ 2016/05/31) for a total of 80 days, using the initial 60 days as a training set and the remaining 20 days as a test set. The proposed word - based algorithm in this study showed better classification performance than the word selection method based on sparsity. This study predicted stock price volatility by collecting and analyzing news articles of the top 10 stocks in market cap. We used the term - document matrix based classification model to estimate the stock price fluctuations and compared the performance of the existing sparse - based word extraction method and the suggested method of removing words from the term - document matrix. The suggested method differs from the word extraction method in that it uses not only the news articles for the corresponding stock but also other news items to determine the words to extract. In other words, it removed not only the words that appeared in all the increase and decrease but also the words that appeared common in the news for other stocks. When the prediction accuracy was compared, the suggested method showed higher accuracy. The limitation of this study is that the stock price prediction was set up to classify the rise and fall, and the experiment was conducted only for the top ten stocks. The 10 stocks used in the experiment do not represent the entire stock market. In addition, it is difficult to show the investment performance because stock price fluctuation and profit rate may be different. Therefore, it is necessary to study the research using more stocks and the yield prediction through trading simulation.

Problems in the Korean National Family Planning Program (한국가족계획사업(韓國家族計劃事業)의 문제점(問題點))

  • Hong, Jong-Kwan
    • Clinical and Experimental Reproductive Medicine
    • /
    • v.2 no.2
    • /
    • pp.27-36
    • /
    • 1975
  • The success of the family planning program in Korea is reflected in the decrease in the growth rate from 3.0% in 1962 to 2.0% in 1971, and in the decrease in the fertility rate from 43/1,000 in 1960 to 29/1,000 in 1970. However, it would be erroneous to attribute these reductions entirely to the family planning program. Other socio-economic factors, such as the increasing age at marriage and the increasing use of induced abortions, definitely had an impact on the lowered growth and fertility rate. Despite the relative success of the program to data in meeting its goals, there is no room for complacency. Meeting the goal of a further reduction in the population growth rate to 1.3% by 1981 is a much more difficult task than any one faced in the past. Not only must fertility be lowered further, but the size of the target population itself will expand tremendously in the late seventies; due to the post-war baby boom of the 1950's reaching reproductive ages. Furthermore, it is doubtful that the age at marriage will continue to rise as in the past or that the incidence of induced abortion will continue to increase. Consequently, future reductions in fertility will be more dependent on the performance of the national family planning program, with less assistance from these non-program factors. This paper will describe various approaches to help to the solution of these current problems. 1. PRACTICE RATE IN FAMILY PLANNING In 1973, the attitude (approval) and knowledge rates were quite high; 94% and 98% respectively. But a large gap exists between that and the actual practice rate, which is only 3695. Two factors must be considered in attempting to close the KAP-gap. The first is to change social norms, which still favor a larger family, increasing the practice rate cannot be done very quickly. The second point to consider is that the family planning program has not yet reached all the eligible women. A 1973 study determineded that a large portion, 3096 in fact, of all eligible women do not want more children, but are not practicing family planning. Thus, future efforts to help close the KAP-gap must focus attention and services on this important large group of potential acceptors. 2. CONTINUATION RATES Dissatisfaction with the loop and pill has resulted in high discontinuation rates. For example, a 1973 survey revealed that within the first six months initial loop acceptance. nearly 50% were dropouts, and that within the first four months of inital pill acceptance. nearly 50% were dropouts. These discontinuation rates have risen over the past few years. The high rate of discontinuance obviously decreases the contraceptive effectiveness. and has resulted in many unwanted births which is directly related to the increase of induced abortions. In the future, the family planning program must emphasize the improved quality of initial and follow-up services. rather than more quantity, in order to insure higher continuation rates and thus more effective contraceptive protection. 3. INDUCED ABORTION As noted earlier. the use of induced abortions has been increase yearly. For example, in 1960, the average number of abortions was 0.6 abortions per women in the 15-44 age range. By 1970. that had increased to 2 abortions per women. In 1966. 13% of all women between 15-44 had experienced at least one abortion. By 1971, that figure jumped to 28%. In 1973 alone, the total number of abortions was 400,000. Besides the ever incre.sing number of induced abortions, another change has that those who use abortions have shifted since 1965 to include- not. only the middle class, but also rural and low-income women. In the future. in response to the demand for abortion services among rural and low-income w~men, the government must provide and support abortion services for these women as a part of the national family planning program. 4. TARGET SYSTIi:M Since 1962, the nationwide target system has been used to set a target for each method, and the target number of acceptors is then apportioned out to various sub-areas according to the number of eligible couples in each area. Because these targets are set without consideration for demographic factors, particular tastes, prejudices, and previous patterns of acceptance in the area, a high discontinuation rate for all methods and a high wastage rate for the oral pill and condom results. In the future. to alleviate these problems of the methodbased target system. an alternative. such as the weighted-credit system, should be adopted on a nation wide basis. In this system. each contraceptive method is. assigned a specific number of points based upon the couple-years of protection (CYP) provided by the method. and no specific targets for each method are given. 5. INCREASE OF STERILIZA.TION TARGET Two special projects. the hospital-based family planning program and the armed forces program, has greatly contributed to the increasing acceptance in female and male sterilization respectively. From January-September 1974, 28,773 sterilizations were performed. During the same time in 1975, 46,894 were performed; a 63% increase. If this trend continues, by the end of 1975. approximately 70,000 sterilizations will have been performed. Sterilization is a much better method than both the loop and pill, in terms of more effective contraceptive protection and the almost zero dropout rate. In the future, the. family planning program should continue to stress the special programs which make more sterilizations possible. In particular, it should seek to add the laparoscope techniques to facilitate female sterilization acceptance rates. 6. INCREASE NUMBER OF PRIVATE ACCEPTORS Among the current family planning users, approximately 1/3 are in the private sector and thus do not- require government subsidy. The number of private acceptors increases with increasing urbanization and economic growth. To speed this process, the government initiated the special hospital based family planning program which is utilized mostly by the private sector. However, in the future, to further hasten the increase of private acceptors, the government should encourage doctors in private practice to provide family planning services, and provide the contraceptive supplies. This way, those do utilize the private medical system will also be able to receive family planning services and pay for it. Another means of increasing the number of private acceptors, IS to greatly expand the commercial outlets for pills and condoms beyond the existing service points of drugstores, hospitals, and health centers. 7. IE&C PROGRAM The current preferred family size is nearly twice as high as needed to achieve a stable poplation. Also, a strong boy preference hinders a small family size as nearly all couples fuel they must have at least one or more sons. The IE&C program must, in the future, strive to emphasize the values of the small family and equality of the sexes. A second problem for the IE&C program to work. with in the: future is the large group of people who approves family planning, want no more children, but do not practice. The IE&C program must work to motivate these people to accept family planning And finally, for those who already practice, an IE&C program in the future must stress continuation of use. The IE&C campaign, to insure highest effectiveness, should be based on a detailed factor analysis of contraceptive discontinuance. In conclusion, Korea faces a serious unfavorable sociodemographic situation- in the future unless the population growth rate can be curtailed. And in the future, the decrease in fertility will depend solely on the family planning program, as the effect of other socio-economic factors has already been maximumally felt. A second serious factor to consider is the increasing number of eligible women due to the 1950's baby boom. Thus, to meet these challenges, the program target must be increased and the program must improve the effectiveness of its current activities and develop new programs.

  • PDF

A Data-based Sales Forecasting Support System for New Businesses (데이터기반의 신규 사업 매출추정방법 연구: 지능형 사업평가 시스템을 중심으로)

  • Jun, Seung-Pyo;Sung, Tae-Eung;Choi, San
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.1-22
    • /
    • 2017
  • Analysis of future business or investment opportunities, such as business feasibility analysis and company or technology valuation, necessitate objective estimation on the relevant market and expected sales. While there are various ways to classify the estimation methods of these new sales or market size, they can be broadly divided into top-down and bottom-up approaches by benchmark references. Both methods, however, require a lot of resources and time. Therefore, we propose a data-based intelligent demand forecasting system to support evaluation of new business. This study focuses on analogical forecasting, one of the traditional quantitative forecasting methods, to develop sales forecasting intelligence systems for new businesses. Instead of simply estimating sales for a few years, we hereby propose a method of estimating the sales of new businesses by using the initial sales and the sales growth rate of similar companies. To demonstrate the appropriateness of this method, it is examined whether the sales performance of recently established companies in the same industry category in Korea can be utilized as a reference variable for the analogical forecasting. In this study, we examined whether the phenomenon of "mean reversion" was observed in the sales of start-up companies in order to identify errors in estimating sales of new businesses based on industry sales growth rate and whether the differences in business environment resulting from the different timing of business launch affects growth rate. We also conducted analyses of variance (ANOVA) and latent growth model (LGM) to identify differences in sales growth rates by industry category. Based on the results, we proposed industry-specific range and linear forecasting models. This study analyzed the sales of only 150,000 start-up companies in Korea in the last 10 years, and identified that the average growth rate of start-ups in Korea is higher than the industry average in the first few years, but it shortly shows the phenomenon of mean-reversion. In addition, although the start-up founding juncture affects the sales growth rate, it is not high significantly and the sales growth rate can be different according to the industry classification. Utilizing both this phenomenon and the performance of start-up companies in relevant industries, we have proposed two models of new business sales based on the sales growth rate. The method proposed in this study makes it possible to objectively and quickly estimate the sales of new business by industry, and it is expected to provide reference information to judge whether sales estimated by other methods (top-down/bottom-up approach) pass the bounds from ordinary cases in relevant industry. In particular, the results of this study can be practically used as useful reference information for business feasibility analysis or technical valuation for entering new business. When using the existing top-down method, it can be used to set the range of market size or market share. As well, when using the bottom-up method, the estimation period may be set in accordance of the mean reverting period information for the growth rate. The two models proposed in this study will enable rapid and objective sales estimation of new businesses, and are expected to improve the efficiency of business feasibility analysis and technology valuation process by developing intelligent information system. In academic perspectives, it is a very important discovery that the phenomenon of 'mean reversion' is found among start-up companies out of general small-and-medium enterprises (SMEs) as well as stable companies such as listed companies. In particular, there exists the significance of this study in that over the large-scale data the mean reverting phenomenon of the start-up firms' sales growth rate is different from that of the listed companies, and that there is a difference in each industry. If a linear model, which is useful for estimating the sales of a specific company, is highly likely to be utilized in practical aspects, it can be explained that the range model, which can be used for the estimation method of the sales of the unspecified firms, is highly likely to be used in political aspects. It implies that when analyzing the business activities and performance of a specific industry group or enterprise group there is political usability in that the range model enables to provide references and compare them by data based start-up sales forecasting system.

Visualizing the Results of Opinion Mining from Social Media Contents: Case Study of a Noodle Company (소셜미디어 콘텐츠의 오피니언 마이닝결과 시각화: N라면 사례 분석 연구)

  • Kim, Yoosin;Kwon, Do Young;Jeong, Seung Ryul
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.89-105
    • /
    • 2014
  • After emergence of Internet, social media with highly interactive Web 2.0 applications has provided very user friendly means for consumers and companies to communicate with each other. Users have routinely published contents involving their opinions and interests in social media such as blogs, forums, chatting rooms, and discussion boards, and the contents are released real-time in the Internet. For that reason, many researchers and marketers regard social media contents as the source of information for business analytics to develop business insights, and many studies have reported results on mining business intelligence from Social media content. In particular, opinion mining and sentiment analysis, as a technique to extract, classify, understand, and assess the opinions implicit in text contents, are frequently applied into social media content analysis because it emphasizes determining sentiment polarity and extracting authors' opinions. A number of frameworks, methods, techniques and tools have been presented by these researchers. However, we have found some weaknesses from their methods which are often technically complicated and are not sufficiently user-friendly for helping business decisions and planning. In this study, we attempted to formulate a more comprehensive and practical approach to conduct opinion mining with visual deliverables. First, we described the entire cycle of practical opinion mining using Social media content from the initial data gathering stage to the final presentation session. Our proposed approach to opinion mining consists of four phases: collecting, qualifying, analyzing, and visualizing. In the first phase, analysts have to choose target social media. Each target media requires different ways for analysts to gain access. There are open-API, searching tools, DB2DB interface, purchasing contents, and so son. Second phase is pre-processing to generate useful materials for meaningful analysis. If we do not remove garbage data, results of social media analysis will not provide meaningful and useful business insights. To clean social media data, natural language processing techniques should be applied. The next step is the opinion mining phase where the cleansed social media content set is to be analyzed. The qualified data set includes not only user-generated contents but also content identification information such as creation date, author name, user id, content id, hit counts, review or reply, favorite, etc. Depending on the purpose of the analysis, researchers or data analysts can select a suitable mining tool. Topic extraction and buzz analysis are usually related to market trends analysis, while sentiment analysis is utilized to conduct reputation analysis. There are also various applications, such as stock prediction, product recommendation, sales forecasting, and so on. The last phase is visualization and presentation of analysis results. The major focus and purpose of this phase are to explain results of analysis and help users to comprehend its meaning. Therefore, to the extent possible, deliverables from this phase should be made simple, clear and easy to understand, rather than complex and flashy. To illustrate our approach, we conducted a case study on a leading Korean instant noodle company. We targeted the leading company, NS Food, with 66.5% of market share; the firm has kept No. 1 position in the Korean "Ramen" business for several decades. We collected a total of 11,869 pieces of contents including blogs, forum contents and news articles. After collecting social media content data, we generated instant noodle business specific language resources for data manipulation and analysis using natural language processing. In addition, we tried to classify contents in more detail categories such as marketing features, environment, reputation, etc. In those phase, we used free ware software programs such as TM, KoNLP, ggplot2 and plyr packages in R project. As the result, we presented several useful visualization outputs like domain specific lexicons, volume and sentiment graphs, topic word cloud, heat maps, valence tree map, and other visualized images to provide vivid, full-colored examples using open library software packages of the R project. Business actors can quickly detect areas by a swift glance that are weak, strong, positive, negative, quiet or loud. Heat map is able to explain movement of sentiment or volume in categories and time matrix which shows density of color on time periods. Valence tree map, one of the most comprehensive and holistic visualization models, should be very helpful for analysts and decision makers to quickly understand the "big picture" business situation with a hierarchical structure since tree-map can present buzz volume and sentiment with a visualized result in a certain period. This case study offers real-world business insights from market sensing which would demonstrate to practical-minded business users how they can use these types of results for timely decision making in response to on-going changes in the market. We believe our approach can provide practical and reliable guide to opinion mining with visualized results that are immediately useful, not just in food industry but in other industries as well.

Electronic Roll Book using Electronic Bracelet.Child Safe-Guarding Device System (전자 팔찌를 이용한 전자 출석부.어린이 보호 장치 시스템)

  • Moon, Seung-Jin;Kim, Tae-Nam;Kim, Pan-Su
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.4
    • /
    • pp.143-155
    • /
    • 2011
  • Lately electronic tagging policy for the sexual offenders was introduced in order to reduce and prevent sexual offences. However, most sexual offences against children happening these days are committed by the tagged offenders whose identities have been released. So, for the crime prevention, we need measures with which we could minimize the suffers more promptly and actively. This paper suggests a new system to relieve the sexual abuse related anxiety of the children and solve the problems that electronic bracelet has. Existing bracelets are only worn by serious criminals, and it's only for risk management and positioning, there is no way to protect the children who are the potential victims of sexual abuse and there actually happened some cases. So we suggest also letting the students(children) wear the LBS(Location Based Service) and USN(Ubiquitous Sensor Network) technology based electronic bracelets to monitor and figure out dangerous situations intelligently, so that we could prevent sexual offences against children beforehand, and while a crime is happening, we could judge the situation of the crime intelligently and take swift action to minimize the suffer. And by checking students' attendance and position, guardians could know where their children are in real time and could protect the children from not only sexual offences but also violent crimes against children like kidnapping. The overall system is like follows : RFID Tag for children monitors the approach of offenders. While an offender's RFID tag is approaching, it will transmit the situation and position as the first warning message to the control center and the guardians. When the offender is going far away, it turns to monitoring mode, and if the tag of the child or the offender is taken off or the child and offender stay at one position for 3~5 minutes or longer, then it will consider this as a dangerous situation, then transmit the emergency situations and position as the second warning message to the control center and the guardians, and ask for the dispatch of police to prevent the crime at the initial stage. The RFID module of criminals' electronic bracelets is RFID TAG, and the RFID module for the children is RFID receiver(reader), so wherever the offenders are, if an offender is at a place within 20m from a child, RFID module for children will transmit the situation every certain periods to the control center by the automatic response of the receiver. As for the positioning module, outdoors GPS or mobile communications module(CELL module)is used and UWB, WI-FI based module is used indoors. The sensor is set under the purpose of making it possible to measure the position coordinates even indoors, so that one could send his real time situation and position to the server of central control center. By using the RFID electronic roll book system of educational institutions and safety system installed at home, children's position and situation can be checked. When the child leaves for school, attendance can be checked through the electronic roll book, and when school is over the information is sent to the guardians. And using RFID access control turnstiles installed at the apartment or entrance of the house, the arrival of the children could be checked and the information is transmitted to the guardians. If the student is absent or didn't arrive at home, the information of the child is sent to the central control center from the electronic roll book or access control turnstiles, and look for the position of the child's electronic bracelet using GPS or mobile communications module, then send the information to the guardians and teacher so that they could report to the police immediately if necessary. Central management and control system is built under the purpose of monitoring dangerous situations and guardians' checking. It saves the warning and pattern data to figure out the areas with dangerous situation, and could help introduce crime prevention systems like CCTV with the highest priority. And by DB establishment personal data could be saved, the frequency of first and second warnings made, the terminal ID of the specific child and offender, warning made position, situation (like approaching, taken off of the electronic bracelet, same position for a certain time) and so on could be recorded, and the data is going to be used for preventing crimes. Even though we've already introduced electronic tagging to prevent recurrence of child sexual offences, but the crimes continuously occur. So I suggest this system to prevent crimes beforehand concerning the children's safety. If we make electronic bracelets easy to use and carry, and set the price reasonably so that many children can use, then lots of criminals could be prevented and we can protect the children easily. By preventing criminals before happening, it is going to be a helpful system for our safe life.

A Study on the Application of Outlier Analysis for Fraud Detection: Focused on Transactions of Auction Exception Agricultural Products (부정 탐지를 위한 이상치 분석 활용방안 연구 : 농수산 상장예외품목 거래를 대상으로)

  • Kim, Dongsung;Kim, Kitae;Kim, Jongwoo;Park, Steve
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.93-108
    • /
    • 2014
  • To support business decision making, interests and efforts to analyze and use transaction data in different perspectives are increasing. Such efforts are not only limited to customer management or marketing, but also used for monitoring and detecting fraud transactions. Fraud transactions are evolving into various patterns by taking advantage of information technology. To reflect the evolution of fraud transactions, there are many efforts on fraud detection methods and advanced application systems in order to improve the accuracy and ease of fraud detection. As a case of fraud detection, this study aims to provide effective fraud detection methods for auction exception agricultural products in the largest Korean agricultural wholesale market. Auction exception products policy exists to complement auction-based trades in agricultural wholesale market. That is, most trades on agricultural products are performed by auction; however, specific products are assigned as auction exception products when total volumes of products are relatively small, the number of wholesalers is small, or there are difficulties for wholesalers to purchase the products. However, auction exception products policy makes several problems on fairness and transparency of transaction, which requires help of fraud detection. In this study, to generate fraud detection rules, real huge agricultural products trade transaction data from 2008 to 2010 in the market are analyzed, which increase more than 1 million transactions and 1 billion US dollar in transaction volume. Agricultural transaction data has unique characteristics such as frequent changes in supply volumes and turbulent time-dependent changes in price. Since this was the first trial to identify fraud transactions in this domain, there was no training data set for supervised learning. So, fraud detection rules are generated using outlier detection approach. We assume that outlier transactions have more possibility of fraud transactions than normal transactions. The outlier transactions are identified to compare daily average unit price, weekly average unit price, and quarterly average unit price of product items. Also quarterly averages unit price of product items of the specific wholesalers are used to identify outlier transactions. The reliability of generated fraud detection rules are confirmed by domain experts. To determine whether a transaction is fraudulent or not, normal distribution and normalized Z-value concept are applied. That is, a unit price of a transaction is transformed to Z-value to calculate the occurrence probability when we approximate the distribution of unit prices to normal distribution. The modified Z-value of the unit price in the transaction is used rather than using the original Z-value of it. The reason is that in the case of auction exception agricultural products, Z-values are influenced by outlier fraud transactions themselves because the number of wholesalers is small. The modified Z-values are called Self-Eliminated Z-scores because they are calculated excluding the unit price of the specific transaction which is subject to check whether it is fraud transaction or not. To show the usefulness of the proposed approach, a prototype of fraud transaction detection system is developed using Delphi. The system consists of five main menus and related submenus. First functionalities of the system is to import transaction databases. Next important functions are to set up fraud detection parameters. By changing fraud detection parameters, system users can control the number of potential fraud transactions. Execution functions provide fraud detection results which are found based on fraud detection parameters. The potential fraud transactions can be viewed on screen or exported as files. The study is an initial trial to identify fraud transactions in Auction Exception Agricultural Products. There are still many remained research topics of the issue. First, the scope of analysis data was limited due to the availability of data. It is necessary to include more data on transactions, wholesalers, and producers to detect fraud transactions more accurately. Next, we need to extend the scope of fraud transaction detection to fishery products. Also there are many possibilities to apply different data mining techniques for fraud detection. For example, time series approach is a potential technique to apply the problem. Even though outlier transactions are detected based on unit prices of transactions, however it is possible to derive fraud detection rules based on transaction volumes.

A Deep Learning Based Approach to Recognizing Accompanying Status of Smartphone Users Using Multimodal Data (스마트폰 다종 데이터를 활용한 딥러닝 기반의 사용자 동행 상태 인식)

  • Kim, Kilho;Choi, Sangwoo;Chae, Moon-jung;Park, Heewoong;Lee, Jaehong;Park, Jonghun
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.163-177
    • /
    • 2019
  • As smartphones are getting widely used, human activity recognition (HAR) tasks for recognizing personal activities of smartphone users with multimodal data have been actively studied recently. The research area is expanding from the recognition of the simple body movement of an individual user to the recognition of low-level behavior and high-level behavior. However, HAR tasks for recognizing interaction behavior with other people, such as whether the user is accompanying or communicating with someone else, have gotten less attention so far. And previous research for recognizing interaction behavior has usually depended on audio, Bluetooth, and Wi-Fi sensors, which are vulnerable to privacy issues and require much time to collect enough data. Whereas physical sensors including accelerometer, magnetic field and gyroscope sensors are less vulnerable to privacy issues and can collect a large amount of data within a short time. In this paper, a method for detecting accompanying status based on deep learning model by only using multimodal physical sensor data, such as an accelerometer, magnetic field and gyroscope, was proposed. The accompanying status was defined as a redefinition of a part of the user interaction behavior, including whether the user is accompanying with an acquaintance at a close distance and the user is actively communicating with the acquaintance. A framework based on convolutional neural networks (CNN) and long short-term memory (LSTM) recurrent networks for classifying accompanying and conversation was proposed. First, a data preprocessing method which consists of time synchronization of multimodal data from different physical sensors, data normalization and sequence data generation was introduced. We applied the nearest interpolation to synchronize the time of collected data from different sensors. Normalization was performed for each x, y, z axis value of the sensor data, and the sequence data was generated according to the sliding window method. Then, the sequence data became the input for CNN, where feature maps representing local dependencies of the original sequence are extracted. The CNN consisted of 3 convolutional layers and did not have a pooling layer to maintain the temporal information of the sequence data. Next, LSTM recurrent networks received the feature maps, learned long-term dependencies from them and extracted features. The LSTM recurrent networks consisted of two layers, each with 128 cells. Finally, the extracted features were used for classification by softmax classifier. The loss function of the model was cross entropy function and the weights of the model were randomly initialized on a normal distribution with an average of 0 and a standard deviation of 0.1. The model was trained using adaptive moment estimation (ADAM) optimization algorithm and the mini batch size was set to 128. We applied dropout to input values of the LSTM recurrent networks to prevent overfitting. The initial learning rate was set to 0.001, and it decreased exponentially by 0.99 at the end of each epoch training. An Android smartphone application was developed and released to collect data. We collected smartphone data for a total of 18 subjects. Using the data, the model classified accompanying and conversation by 98.74% and 98.83% accuracy each. Both the F1 score and accuracy of the model were higher than the F1 score and accuracy of the majority vote classifier, support vector machine, and deep recurrent neural network. In the future research, we will focus on more rigorous multimodal sensor data synchronization methods that minimize the time stamp differences. In addition, we will further study transfer learning method that enables transfer of trained models tailored to the training data to the evaluation data that follows a different distribution. It is expected that a model capable of exhibiting robust recognition performance against changes in data that is not considered in the model learning stage will be obtained.

Effect of Light Intensity on the Growth Responses of Three Woody Plants for Indoor Landscaping (실내녹화용 목본식물 3종의 초기 생육반응에 미치는 광량의 영향)

  • Kwon, Kei-Jung;Park, Bong-Ju
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.46 no.1
    • /
    • pp.1-8
    • /
    • 2018
  • The purpose of this study was to investigate the effects of light intensity on the initial growth response of three woody plants for indoor landscaping; Ardisia pusilla, Clusia rosea and Fatsia japonica. The plants were planted in 10cm pots, the light intensities used were of four levels-15, 30, 60, $120{\mu}mol{\cdot}m^{-2}{\cdot}s^{-1}$ PPFD-and light irradiation time was set to 12/12 (day/night). Growth responses including plant height, leaf length, leaf width, chlorophyll fluorescence (Fv/Fm), SPAD and Hunter values were measured at 4-week intervals, and shoot weight and root weight of fresh and dry plants were measured after completion of the experiment. Fatsia japonica tended to show greater leaf length and leaf width as light intensity became greater, while other plants did not show any significant differences at different light intensities. The Fv/Fm value of the Ardisia pusilla was found to be stressed at 60 and $120{\mu}mol{\cdot}m^{-2}{\cdot}s^{-1}$, while the Fv/Fm values were within normal range with other plants or at other light intensity levels to show no stress. Only Clusia rosea showed significantly different SPAD values at $120{\mu}mol{\cdot}m^{-2}{\cdot}s^{-1}$, and there was no significant SPAD value difference found with other plants or at other light intensity levels. While Hunter values of the Ardisia pusilla did not show any significant differences at any light intensity levels, Clusia rosea and Fatsia japonica showed specificity in L, a and b values at 60 and $120{\mu}mol{\cdot}m^{-2}{\cdot}s^{-1}$, respectively. Ardisia pusilla showed a big stem growth at $120{\mu}mol{\cdot}m^{-2}{\cdot}s^{-1}$, and Clusia rosea showed a steady growth at 60 and $120{\mu}mol{\cdot}m^{-2}{\cdot}s^{-1}$.

A Study of the Effect of the Permeability and Selectivity on the Performance of Membrane System Design (분리막 투과도와 분리도 인자의 시스템 설계 효과 연구)

  • Shin, Mi-Soo;Jang, Dongsoon;Lee, Yongguk
    • Journal of Korean Society of Environmental Engineers
    • /
    • v.38 no.12
    • /
    • pp.656-661
    • /
    • 2016
  • Manufacturing membrane materials with high selectivity and permeability is quite desirable but practically not possible, since the permeability and selectivity are usually inversely proportional. From the viewpoint of reducing the cost of $CO_2$ capture, module performance is even more important than the performance of membrane materials itself, which is affected by the permeance of the membrane (P, stagecut) and selectivity (S). As a typical example, when the mixture with a composition of 13% $CO_2$ and 87% of $N_2$ is fed into the module with 10% stage cut and selectivity 5, in the 10 parts of the permeate, $CO_2$ represents 4.28 parts and $N_2$ represents 5.72 parts. In this case, the $CO_2$ concentration in the permeate is 42.8% and the recovery rate of $CO_2$ in this first separation appears as 4.28/13 = 32.9%. When permeance and selectivity are doubled, however, from 10% to 20% and from 5 to 10, respectively, the $CO_2$ concentration in the permeant becomes 64.5% and the recovery rate is 12.9/13 = 99.2%. Since in this case, most of the $CO_2$ is separated, this may be the ideal condition. For a given feed concentration, the $CO_2$ concentration in the separated gas decreases if permeance is larger than the threshold value for complete recovery at a given selectivity. Conversely, for a given permeance, increasing the selectivity over the threshold value does not improve the process further. For a given initial feed gas concentration, if permeance or selectivity is larger than that required for the complete separation of $CO_2$, the process becomes less efficient. From all these considerations, we can see that there exists an optimum design for a given set of conditions.