• Title/Summary/Keyword: Feature-based Model

Search Result 2,040, Processing Time 0.032 seconds

Effect of SPR Chip with Nano-structured Surface on Sensitivity in SPR Sensor (나노형상을 가진 표면플라즈몬공명 센서칩의 감도 개선 효과)

  • Cho, Yong-Jin;Kim, Chul-Jin;Kim, Namsoo;Kim, Chong-Tai;Kim, Tae-Eun;Kim, Hyo-Sop;Kim, Jae-Ho
    • Food Engineering Progress
    • /
    • v.14 no.1
    • /
    • pp.49-53
    • /
    • 2010
  • Surface plasmon resonance (SPR) which is utilized in thin film refractometry-based sensors has been concerned on measurement of physical, chemical and biological quantities because of its high sensitivity and label-free feature. In this paper, an application of SPR to detection of alcohol content in wine and liquor was investigated. The result showed that SPR sensor had high potential to evaluate alcohol content. Nevertheless, food industry may need SPR sensor with higher sensitivity. Herein, we introduced a nano-technique into fabrication of SPR chip to enhance SPR sensitivity. Using Langmuir-Blodgett (LB) method, gold film with nano-structured surface was devised. In order to make a new SPR chip, firstly, a single layer of nano-scaled silica particles adhered to plain surface of gold film. Thereafter, gold was deposited on the template by an e-beam evaporator. Finally, the nano-structured surface with basin-like shape was obtained after removing the silica particles by sonication. In this study, two types of silica particles, or 130 nm and 300 nm, were used as template beads and sensitivity of the new SPR chip was tested with ethanol solution, respectively. Applying the new developed SPR sensor to a model food of alcoholic beverage, the sensitivity showed improvement of 95% over the conventional one.

White striping degree assessment using computer vision system and consumer acceptance test

  • Kato, Talita;Mastelini, Saulo Martiello;Campos, Gabriel Fillipe Centini;Barbon, Ana Paula Ayub da Costa;Prudencio, Sandra Helena;Shimokomaki, Massami;Soares, Adriana Lourenco;Barbon, Sylvio Jr.
    • Asian-Australasian Journal of Animal Sciences
    • /
    • v.32 no.7
    • /
    • pp.1015-1026
    • /
    • 2019
  • Objective: The objective of this study was to evaluate three different degrees of white striping (WS) addressing their automatic assessment and customer acceptance. The WS classification was performed based on a computer vision system (CVS), exploring different machine learning (ML) algorithms and the most important image features. Moreover, it was verified by consumer acceptance and purchase intent. Methods: The samples for image analysis were classified by trained specialists, according to severity degrees regarding visual and firmness aspects. Samples were obtained with a digital camera, and 25 features were extracted from these images. ML algorithms were applied aiming to induce a model capable of classifying the samples into three severity degrees. In addition, two sensory analyses were performed: 75 samples properly grilled were used for the first sensory test, and 9 photos for the second. All tests were performed using a 10-cm hybrid hedonic scale (acceptance test) and a 5-point scale (purchase intention). Results: The information gain metric ranked 13 attributes. However, just one type of image feature was not enough to describe the phenomenon. The classification models support vector machine, fuzzy-W, and random forest showed the best results with similar general accuracy (86.4%). The worst performance was obtained by multilayer perceptron (70.9%) with the high error rate in normal (NORM) sample predictions. The sensory analysis of acceptance verified that WS myopathy negatively affects the texture of the broiler breast fillets when grilled and the appearance attribute of the raw samples, which influenced the purchase intention scores of raw samples. Conclusion: The proposed system has proved to be adequate (fast and accurate) for the classification of WS samples. The sensory analysis of acceptance showed that WS myopathy negatively affects the tenderness of the broiler breast fillets when grilled, while the appearance attribute of the raw samples eventually influenced purchase intentions.

Statistical analysis of estimating incubation period distribution and case fatality rate of COVID-19 (COVID-19 바이러스 잠복 시간 분포 추정과 치사율 추정을 위한 생존 분석의 적용)

  • Ki, Han Jeong;Kim, Jieun;Kim, Sohee;Park, Juwon;Lee, Joohaeng;Kim, Yang-Jin
    • The Korean Journal of Applied Statistics
    • /
    • v.33 no.6
    • /
    • pp.777-789
    • /
    • 2020
  • COVID-19 has been rapidly spread world wide since late December 2019. In this paper, our interest is to estimate distribution of incubation time defined as period between infection of virus and the onset. Due to the limit of accessibility and asymptomatic feature of COVID-19 virus, the exact infection and onset time are not always observable. For estimation of incubation time, interval censoring technique is implemented. Furthermore, a competing risk model is applied to estimate the case fatality and cure fraction. Based on the result, the mean incubation time is about 5.4 days and the fatality rate is higher for older and male patient and the cure rate is higher at younger,female and asymptomatic patient.

Comparative Analysis by Batch Size when Diagnosing Pneumonia on Chest X-Ray Image using Xception Modeling (Xception 모델링을 이용한 흉부 X선 영상 폐렴(pneumonia) 진단 시 배치 사이즈별 비교 분석)

  • Kim, Ji-Yul;Ye, Soo-Young
    • Journal of the Korean Society of Radiology
    • /
    • v.15 no.4
    • /
    • pp.547-554
    • /
    • 2021
  • In order to quickly and accurately diagnose pneumonia on a chest X-ray image, different batch sizes of 4, 8, 16, and 32 were applied to the same Xception deep learning model, and modeling was performed 3 times, respectively. As a result of the performance evaluation of deep learning modeling, in the case of modeling to which batch size 32 was applied, the results of accuracy, loss function value, mean square error, and learning time per epoch showed the best results. And in the accuracy evaluation of the Test Metric, the modeling applied with batch size 8 showed the best results, and the precision evaluation showed excellent results in all batch sizes. In the recall evaluation, modeling applied with batch size 16 showed the best results, and for F1-score, modeling applied with batch size 16 showed the best results. And the AUC score evaluation was the same for all batch sizes. Based on these results, deep learning modeling with batch size 32 showed high accuracy, stable artificial neural network learning, and excellent speed. It is thought that accurate and rapid lesion detection will be possible if a batch size of 32 is applied in an automatic diagnosis study for feature extraction and classification of pneumonia in chest X-ray images using deep learning in the future.

Efficiency evaluation of nursing homes in China's eastern areas Based on DEA-Malmquist Model (DEA-Malmquist를 활용한 중국 동부지역 요양원의 효율성 평가에 관한 연구)

  • Chu, Ting;Sim, Jae-yeon
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.7
    • /
    • pp.273-282
    • /
    • 2021
  • Nursing home plays a role in providing elderly care in the context of China's rapid population aging, but little understanding of the efficiency of the nursing homes. In this paper, we investigated the efficiency in nursing homes using Data Envelopment Analysis (DEA) and Malmquist index (MPI) for the modeling of the number of nursing home beds, fixed assets, and medical personnel as input variables, and the number of elderly people of self-care, the number of elderly people of partial self-care, the number of bed-ridden elderly people and the income of nursing homes as output variables. Stratification analysis showed that the top two provinces in the DEA-CCR yield were Beijing and Shanghai in the five-year survey period. Four provinces (Beijing, Jiangsu, Shandong, and Shanghai) scored 1.00 in terms of DEA-BCC yield. The MPI analysis showed that Hainan ranked the highest five-year average in the included provinces. In terms of resource utilization, internal management, operation scale, and other aspects, the nursing homes in the provinces with high-efficiency evaluation results show high efficiency and technological progress, whereas the areas with low-efficiency evaluation showed a feature of the improving technical efficiency.

Denoising Self-Attention Network for Mixed-type Data Imputation (혼합형 데이터 보간을 위한 디노이징 셀프 어텐션 네트워크)

  • Lee, Do-Hoon;Kim, Han-Joon;Chun, Joonghoon
    • The Journal of the Korea Contents Association
    • /
    • v.21 no.11
    • /
    • pp.135-144
    • /
    • 2021
  • Recently, data-driven decision-making technology has become a key technology leading the data industry, and machine learning technology for this requires high-quality training datasets. However, real-world data contains missing values for various reasons, which degrades the performance of prediction models learned from the poor training data. Therefore, in order to build a high-performance model from real-world datasets, many studies on automatically imputing missing values in initial training data have been actively conducted. Many of conventional machine learning-based imputation techniques for handling missing data involve very time-consuming and cumbersome work because they are applied only to numeric type of columns or create individual predictive models for each columns. Therefore, this paper proposes a new data imputation technique called 'Denoising Self-Attention Network (DSAN)', which can be applied to mixed-type dataset containing both numerical and categorical columns. DSAN can learn robust feature expression vectors by combining self-attention and denoising techniques, and can automatically interpolate multiple missing variables in parallel through multi-task learning. To verify the validity of the proposed technique, data imputation experiments has been performed after arbitrarily generating missing values for several mixed-type training data. Then we show the validity of the proposed technique by comparing the performance of the binary classification models trained on imputed data together with the errors between the original and imputed values.

A Case Study on Global Marketing of 'CJ O Shopping' (CJ오쇼핑의 글로벌 마케팅 사례)

  • Yeu, Minsun;Lee, Doo-Hee;Yeo, Jun Sang;Lee, Hyunjoung
    • Asia Marketing Journal
    • /
    • v.13 no.4
    • /
    • pp.253-264
    • /
    • 2012
  • A growing number of Korean companies are trying to expand their business area into global market due to saturation in the Korean domestic market. Home shopping industry arriving on mature stage is faced with less growth recently. CJ O Shopping which is a top ranked home shopping company in Korea, has been showing meaningful performances by earlier moving to global market with thorough preparations. CJ O Shopping's global marketing strategy focused on asian countries including China, India, Vietnam, and Japan is going successfully, which enables top ranked on-line retailing company in asia as well as in Korea. CJ O Shopping effectively penetrated into overseas market with both core competence based on Korean home shopping model and rigorous preliminary study on target market. Especially shoppertainment (Shopping+Entertainment) that is unique feature of globally competitive Korean home shopping created huge differentiations in target market. Also choosing the influential local partner, sharing the business goals, and building the joint venture could make stable operations, thereby easily earning of well-established awareness from target consumers. A step ahead entry of competitors and intensive localization of CJ O Shopping's core competence for arriving safe in target market were additional key factors for global marketing success. We can extract above key factors for success as implications of case study on CJ O Shopping's global marketing, and expect those factors to be spread into lots of Korean companies and utilized as successful strategies for global marketing.

  • PDF

Statistical Techniques to Detect Sensor Drifts (센서드리프트 판별을 위한 통계적 탐지기술 고찰)

  • Seo, In-Yong;Shin, Ho-Cheol;Park, Moon-Ghu;Kim, Seong-Jun
    • Journal of the Korea Society for Simulation
    • /
    • v.18 no.3
    • /
    • pp.103-112
    • /
    • 2009
  • In a nuclear power plant (NPP), periodic sensor calibrations are required to assure sensors are operating correctly. However, only a few faulty sensors are found to be calibrated. For the safe operation of an NPP and the reduction of unnecessary calibration, on-line calibration monitoring is needed. In this paper, principal component-based Auto-Associative support vector regression (PCSVR) was proposed for the sensor signal validation of the NPP. It utilizes the attractive merits of principal component analysis (PCA) for extracting predominant feature vectors and AASVR because it easily represents complicated processes that are difficult to model with analytical and mechanistic models. With the use of real plant startup data from the Kori Nuclear Power Plant Unit 3, SVR hyperparameters were optimized by the response surface methodology (RSM). Moreover the statistical techniques are integrated with PCSVR for the failure detection. The residuals between the estimated signals and the measured signals are tested by the Shewhart Control Chart, Exponentially Weighted Moving Average (EWMA), Cumulative Sum (CUSUM) and generalized likelihood ratio test (GLRT) to detect whether the sensors are failed or not. This study shows the GLRT can be a candidate for the detection of sensor drift.

Prediction of Key Variables Affecting NBA Playoffs Advancement: Focusing on 3 Points and Turnover Features (미국 프로농구(NBA)의 플레이오프 진출에 영향을 미치는 주요 변수 예측: 3점과 턴오버 속성을 중심으로)

  • An, Sehwan;Kim, Youngmin
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.263-286
    • /
    • 2022
  • This study acquires NBA statistical information for a total of 32 years from 1990 to 2022 using web crawling, observes variables of interest through exploratory data analysis, and generates related derived variables. Unused variables were removed through a purification process on the input data, and correlation analysis, t-test, and ANOVA were performed on the remaining variables. For the variable of interest, the difference in the mean between the groups that advanced to the playoffs and did not advance to the playoffs was tested, and then to compensate for this, the average difference between the three groups (higher/middle/lower) based on ranking was reconfirmed. Of the input data, only this year's season data was used as a test set, and 5-fold cross-validation was performed by dividing the training set and the validation set for model training. The overfitting problem was solved by comparing the cross-validation result and the final analysis result using the test set to confirm that there was no difference in the performance matrix. Because the quality level of the raw data is high and the statistical assumptions are satisfied, most of the models showed good results despite the small data set. This study not only predicts NBA game results or classifies whether or not to advance to the playoffs using machine learning, but also examines whether the variables of interest are included in the major variables with high importance by understanding the importance of input attribute. Through the visualization of SHAP value, it was possible to overcome the limitation that could not be interpreted only with the result of feature importance, and to compensate for the lack of consistency in the importance calculation in the process of entering/removing variables. It was found that a number of variables related to three points and errors classified as subjects of interest in this study were included in the major variables affecting advancing to the playoffs in the NBA. Although this study is similar in that it includes topics such as match results, playoffs, and championship predictions, which have been dealt with in the existing sports data analysis field, and comparatively analyzed several machine learning models for analysis, there is a difference in that the interest features are set in advance and statistically verified, so that it is compared with the machine learning analysis result. Also, it was differentiated from existing studies by presenting explanatory visualization results using SHAP, one of the XAI models.

Analysis and Evaluation of Frequent Pattern Mining Technique based on Landmark Window (랜드마크 윈도우 기반의 빈발 패턴 마이닝 기법의 분석 및 성능평가)

  • Pyun, Gwangbum;Yun, Unil
    • Journal of Internet Computing and Services
    • /
    • v.15 no.3
    • /
    • pp.101-107
    • /
    • 2014
  • With the development of online service, recent forms of databases have been changed from static database structures to dynamic stream database structures. Previous data mining techniques have been used as tools of decision making such as establishment of marketing strategies and DNA analyses. However, the capability to analyze real-time data more quickly is necessary in the recent interesting areas such as sensor network, robotics, and artificial intelligence. Landmark window-based frequent pattern mining, one of the stream mining approaches, performs mining operations with respect to parts of databases or each transaction of them, instead of all the data. In this paper, we analyze and evaluate the techniques of the well-known landmark window-based frequent pattern mining algorithms, called Lossy counting and hMiner. When Lossy counting mines frequent patterns from a set of new transactions, it performs union operations between the previous and current mining results. hMiner, which is a state-of-the-art algorithm based on the landmark window model, conducts mining operations whenever a new transaction occurs. Since hMiner extracts frequent patterns as soon as a new transaction is entered, we can obtain the latest mining results reflecting real-time information. For this reason, such algorithms are also called online mining approaches. We evaluate and compare the performance of the primitive algorithm, Lossy counting and the latest one, hMiner. As the criteria of our performance analysis, we first consider algorithms' total runtime and average processing time per transaction. In addition, to compare the efficiency of storage structures between them, their maximum memory usage is also evaluated. Lastly, we show how stably the two algorithms conduct their mining works with respect to the databases that feature gradually increasing items. With respect to the evaluation results of mining time and transaction processing, hMiner has higher speed than that of Lossy counting. Since hMiner stores candidate frequent patterns in a hash method, it can directly access candidate frequent patterns. Meanwhile, Lossy counting stores them in a lattice manner; thus, it has to search for multiple nodes in order to access the candidate frequent patterns. On the other hand, hMiner shows worse performance than that of Lossy counting in terms of maximum memory usage. hMiner should have all of the information for candidate frequent patterns to store them to hash's buckets, while Lossy counting stores them, reducing their information by using the lattice method. Since the storage of Lossy counting can share items concurrently included in multiple patterns, its memory usage is more efficient than that of hMiner. However, hMiner presents better efficiency than that of Lossy counting with respect to scalability evaluation due to the following reasons. If the number of items is increased, shared items are decreased in contrast; thereby, Lossy counting's memory efficiency is weakened. Furthermore, if the number of transactions becomes higher, its pruning effect becomes worse. From the experimental results, we can determine that the landmark window-based frequent pattern mining algorithms are suitable for real-time systems although they require a significant amount of memory. Hence, we need to improve their data structures more efficiently in order to utilize them additionally in resource-constrained environments such as WSN(Wireless sensor network).