• Title/Summary/Keyword: 가중치 모델

Search Result 930, Processing Time 0.033 seconds

Comparison of chronic disease risk by dietary carbohydrate energy ratio in Korean elderly: Using the 2007-2009 Korea National Health and Nutrition Examination Survey (한국 노인 식사의 탄수화물 에너지비에 따른 만성질환 위험성 비교: 2007~2009년 국민건강영양조사 자료 이용)

  • Park, Min Seon;Suh, Yoon Suk;Chung, Young-Jin
    • Journal of Nutrition and Health
    • /
    • v.47 no.4
    • /
    • pp.247-257
    • /
    • 2014
  • Purpose: It is reported that most senior people consume a high carbohydrate diet, while a high carbohydrate diet could contribute to the risk of chronic disease. The aim of this study is to determine whether a high carbohydrate diet can increase the risk of chronic disease in elderly Koreans. Methods: Using the 2007-2009 Korean National Health Nutrition Examination Survey data, out of a total of 3,917 individuals aged 65 and above, final 1,535 subjects were analyzed, divided by dietary carbohydrate energy ratio into two groups of moderate carbohydrate ratio (MCR, 55-70%) and excessive carbohydrate ratio (ECR, > 70%). All data were processed after the application of weighted value, using a general linear model or logistic regression. Results: Eighty one percent of elderly Koreans consumed diets with carbohydrate energy ratio above 70%. The ECR group included more female subjects, rural residents, lower income, and lower education level. The ECR group showed lower waist circumference, lower diastolic blood pressure, and lower frequency of consumption of meat and egg, milk, and alcohol. The intake of energy and most nutrients, with the exception of fiber, potassium, vitamin A, and carotene, was lower in the ECR group compared to the MCR group. When analyzed by gender, the ECR group showed lower risk of dyslipidemia in male and obesity in female subjects, even though the ECR group showed low intake of some nutrients. No difference in the risk of hypertension, diabetes, and anemia was observed between the two groups in male or female subjects. Conclusion: This result suggested that a high carbohydrate diet would not be a cause to increase the risk of chronic disease in the elderly. Further study is needed in order to determine an appropriate carbohydrate energy ratio for elderly Koreans to reduce the risk of chronic disease.

CALPUFF Modeling of Odor/suspended Particulate in the Vicinity of Poultry Farms (축사 주변의 악취 및 부유분진의 CALPUFF 모델링: 계사 중심으로)

  • Lim, Kwang-Hee
    • Korean Chemical Engineering Research
    • /
    • v.57 no.1
    • /
    • pp.90-104
    • /
    • 2019
  • In this study, CALPUFF modeling was performed, using a real surface and upper air meterological data to predict trustworthy modeling-results. Pollutant-releases from windscreen chambers of enclosed poultry farms, P1 and P2, and from a open poultry farm, P3, and their diffusing behavior were modeled by CALPUFF modeling with volume sources as well as by finally-adjusted CALPUFF modeling where a linear velocity of upward-exit gas averaged with the weight of each directional-emitting area was applied as a model-linear velocity ($u^M_y$) at a stack, with point sources. In addition, based upon the scenario of poultry farm-releasing odor and particulate matter (PM) removal efficiencies of 0, 20, 50 and 80% or their corresponding emission rates of 100, 80, 50 and 20%, respectively, CALPUFF modeling was performed and concentrations of odor and PM were predicted at the region as a discrete receptor where civil complaints had been frequently filed. The predicted concentrations of ammonia, hydrogen sulfide, $PM_{2.5}$ and $PM_{10}$ were compared with those required to meet according to the offensive odor control law or the atmospheric environmental law. Subsequently their required removal efficiencies at poultry farms of P1, P2 and P3 were estimated. As a result, a priori assumption that pollutant concentrations at their discrete receptors are reduced by the same fraction as pollutant concentrations at P1, P2 and P3 as volume source or point source, were controlled and reduced, was proven applicable in this study. In case of volume source-adopted CALPUFF modeling, its required removal efficiencies of P1 compared with those of point source-adopted CALPUFF modeling, were predicted similar each other. However, In case of volume source-adopted CALPUFF modeling, its required removal efficiencies of both ammonia and $PM_{10}$ at not only P2 but also P3 were predicted higher than those of point source-adopted CALPUFF modeling. Nonetheless, the volume source-adopted CALPUFF modeling was preferred as a safe approach to resolve civil complaints. Accordingly, the required degrees of pollution prevention against ammonia, hydrogen sulfide, $PM_{2.5}$ and $PM_{10}$ at P1 and P2, were estimated in a proper manner.

Semantic Visualization of Dynamic Topic Modeling (다이내믹 토픽 모델링의 의미적 시각화 방법론)

  • Yeon, Jinwook;Boo, Hyunkyung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.131-154
    • /
    • 2022
  • Recently, researches on unstructured data analysis have been actively conducted with the development of information and communication technology. In particular, topic modeling is a representative technique for discovering core topics from massive text data. In the early stages of topic modeling, most studies focused only on topic discovery. As the topic modeling field matured, studies on the change of the topic according to the change of time began to be carried out. Accordingly, interest in dynamic topic modeling that handle changes in keywords constituting the topic is also increasing. Dynamic topic modeling identifies major topics from the data of the initial period and manages the change and flow of topics in a way that utilizes topic information of the previous period to derive further topics in subsequent periods. However, it is very difficult to understand and interpret the results of dynamic topic modeling. The results of traditional dynamic topic modeling simply reveal changes in keywords and their rankings. However, this information is insufficient to represent how the meaning of the topic has changed. Therefore, in this study, we propose a method to visualize topics by period by reflecting the meaning of keywords in each topic. In addition, we propose a method that can intuitively interpret changes in topics and relationships between or among topics. The detailed method of visualizing topics by period is as follows. In the first step, dynamic topic modeling is implemented to derive the top keywords of each period and their weight from text data. In the second step, we derive vectors of top keywords of each topic from the pre-trained word embedding model. Then, we perform dimension reduction for the extracted vectors. Then, we formulate a semantic vector of each topic by calculating weight sum of keywords in each vector using topic weight of each keyword. In the third step, we visualize the semantic vector of each topic using matplotlib, and analyze the relationship between or among the topics based on the visualized result. The change of topic can be interpreted in the following manners. From the result of dynamic topic modeling, we identify rising top 5 keywords and descending top 5 keywords for each period to show the change of the topic. Existing many topic visualization studies usually visualize keywords of each topic, but our approach proposed in this study differs from previous studies in that it attempts to visualize each topic itself. To evaluate the practical applicability of the proposed methodology, we performed an experiment on 1,847 abstracts of artificial intelligence-related papers. The experiment was performed by dividing abstracts of artificial intelligence-related papers into three periods (2016-2017, 2018-2019, 2020-2021). We selected seven topics based on the consistency score, and utilized the pre-trained word embedding model of Word2vec trained with 'Wikipedia', an Internet encyclopedia. Based on the proposed methodology, we generated a semantic vector for each topic. Through this, by reflecting the meaning of keywords, we visualized and interpreted the themes by period. Through these experiments, we confirmed that the rising and descending of the topic weight of a keyword can be usefully used to interpret the semantic change of the corresponding topic and to grasp the relationship among topics. In this study, to overcome the limitations of dynamic topic modeling results, we used word embedding and dimension reduction techniques to visualize topics by era. The results of this study are meaningful in that they broadened the scope of topic understanding through the visualization of dynamic topic modeling results. In addition, the academic contribution can be acknowledged in that it laid the foundation for follow-up studies using various word embeddings and dimensionality reduction techniques to improve the performance of the proposed methodology.

Investment Priorities and Weight Differences of Impact Investors (임팩트 투자자의 투자 우선순위와 비중 차이에 관한 연구)

  • Yoo, Sung Ho;Hwangbo, Yun
    • Asia-Pacific Journal of Business Venturing and Entrepreneurship
    • /
    • v.18 no.3
    • /
    • pp.17-32
    • /
    • 2023
  • In recent years, the need for social ventures that aim to grow while solving social problems through the efficiency and effectiveness of commercial organizations in the market has increased, while there is a limit to how much the government and the public can do to solve social problems. Against this background, the number of social venture startups is increasing in the domestic startup ecosystem, and interest in impact investors, which are investors in social ventures, is also increasing. Therefore, this research utilized judgment analysis technology to objectively analyze the validity and weight of judgment information based on the cognitive process and decision-making environment in the investment decision-making of impact investors. We proceeded with the research by constructing three classifications; first, investment priorities at the initial investment stage for financial benefit and return on investment as an investor, second, the political skills of the entrepreneurs (teams) for the social impact and ripple power, and social venture coexistence and solidarity, third, the social mission of a social venture that meets the purpose of an impact investment fund. As a result of this research, first of all, the investment decision-making priorities of impact investors are the expertise of the entrepreneur (team), the potential rate of return when the entrepreneur (team) succeeds, and the social mission of the entrepreneur (team). Second, impact investors do not have a uniform understanding of the investment decision-making factors, and the factors that determine investment decisions are different, and there are differences in the degree of the weighting. Third, among the various investment decision-making factors of impact investment, "entrepreneur's (team's) networking ability", "entrepreneur's (team's) social insight", "entrepreneur's (team's) interpersonal influence" was relatively lower than the other four factors. The practical contribution through this research is to help social ventures understand the investment determinant factors of impact investors in the process of financing, and impact investors can be expected to improve the quality of investment decision-making by referring to the judgment cases and analysis of impact investors. The academic contribution is that it empirically investigated the investment priorities and weighting differences of impact investors.

  • PDF

Metabolic risk and nutritional state according to breakfast energy level of Korean adults: Using the 2007~2009 Korea National Health and Nutrition Examination Survey (한국 성인의 아침식사 에너지 수준에 따른 대사적 위험과 영양상태: 2007~2009년 국민건강영양조사 자료 이용)

  • Jang, So-Hyoun;Suh, Yoon Suk;Chung, Young-Jin
    • Journal of Nutrition and Health
    • /
    • v.48 no.1
    • /
    • pp.46-57
    • /
    • 2015
  • Purpose: The aim of this study was to determine an appropriate energy level of breakfast with less risk of chronic disease for Korean adults. Methods: Using data from the 2007~2009 Korean National Health & Nutrition Examination Survey, from a total of 12,238 adults aged 19~64, the final 7,769 subjects were analyzed except subjects who were undergoing treatment for cancer or metabolic disorder. According to the percent of breakfast energy intake versus their estimated energy requirement (EER), the subjects were divided into four groups: < 10% (very low, VL), 10~20% (low, L), 20~30% (moderate, M), ${\geq}30%$ (sufficient, S). All data were analyzed on the metabolic risk and nutritional state after application of weighted value and adjustment of sex, age, residential area, income, education, job or jobless, and energy intake using a general linear model or logistic regression. Results: The subjects of group S were 16.9% of total subjects, group M 39.2%, group L 37.6%, and group VL 6.3%. The VL group included more male subjects, younger-aged (19 to 40 years), urban residents, higher income, higher education, and fewer breakfasts eaters together with family members. Among the 4 groups, the VL group showed the highest waist circumference, while the S group showed the lowest waist circumference, body mass index, and serum total cholesterol. The groups of VL and L with lower intake of breakfast energy showed high percent of energy from protein and fat, and low percent of energy from carbohydrate. With the increase of breakfast energy level, intake of energy, most nutrients and food groups increased, and the percentage of subjects consuming nutrients below EAR decreased. The VL group showed relatively higher intake of snacks, sugar, meat and eggs, oil, and seasonings, and the lowest intake of vegetable. Risk of obesity by waist circumference was highest in the VL group by 1.90 times of the S group and the same trend was shown in obesity by BMI. Risk of dyslipidemia by serum total cholesterol was 1.84 times higher in the VL group compared to the S group. Risk of diabetes by Glu-FBS (fasting blood sugar) was 1.57 times higher in the VL group compared to the S group. Conclusion: The results indicate that higher breakfast energy level is positively related to lower metabolic risk and more desirable nutritional state in Korean adults. Therefore, breakfast energy intake more than 30% of their own EER would be highly recommended for Korean adults.

Business Application of Convolutional Neural Networks for Apparel Classification Using Runway Image (합성곱 신경망의 비지니스 응용: 런웨이 이미지를 사용한 의류 분류를 중심으로)

  • Seo, Yian;Shin, Kyung-shik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.1-19
    • /
    • 2018
  • Large amount of data is now available for research and business sectors to extract knowledge from it. This data can be in the form of unstructured data such as audio, text, and image data and can be analyzed by deep learning methodology. Deep learning is now widely used for various estimation, classification, and prediction problems. Especially, fashion business adopts deep learning techniques for apparel recognition, apparel search and retrieval engine, and automatic product recommendation. The core model of these applications is the image classification using Convolutional Neural Networks (CNN). CNN is made up of neurons which learn parameters such as weights while inputs come through and reach outputs. CNN has layer structure which is best suited for image classification as it is comprised of convolutional layer for generating feature maps, pooling layer for reducing the dimensionality of feature maps, and fully-connected layer for classifying the extracted features. However, most of the classification models have been trained using online product image, which is taken under controlled situation such as apparel image itself or professional model wearing apparel. This image may not be an effective way to train the classification model considering the situation when one might want to classify street fashion image or walking image, which is taken in uncontrolled situation and involves people's movement and unexpected pose. Therefore, we propose to train the model with runway apparel image dataset which captures mobility. This will allow the classification model to be trained with far more variable data and enhance the adaptation with diverse query image. To achieve both convergence and generalization of the model, we apply Transfer Learning on our training network. As Transfer Learning in CNN is composed of pre-training and fine-tuning stages, we divide the training step into two. First, we pre-train our architecture with large-scale dataset, ImageNet dataset, which consists of 1.2 million images with 1000 categories including animals, plants, activities, materials, instrumentations, scenes, and foods. We use GoogLeNet for our main architecture as it has achieved great accuracy with efficiency in ImageNet Large Scale Visual Recognition Challenge (ILSVRC). Second, we fine-tune the network with our own runway image dataset. For the runway image dataset, we could not find any previously and publicly made dataset, so we collect the dataset from Google Image Search attaining 2426 images of 32 major fashion brands including Anna Molinari, Balenciaga, Balmain, Brioni, Burberry, Celine, Chanel, Chloe, Christian Dior, Cividini, Dolce and Gabbana, Emilio Pucci, Ermenegildo, Fendi, Giuliana Teso, Gucci, Issey Miyake, Kenzo, Leonard, Louis Vuitton, Marc Jacobs, Marni, Max Mara, Missoni, Moschino, Ralph Lauren, Roberto Cavalli, Sonia Rykiel, Stella McCartney, Valentino, Versace, and Yve Saint Laurent. We perform 10-folded experiments to consider the random generation of training data, and our proposed model has achieved accuracy of 67.2% on final test. Our research suggests several advantages over previous related studies as to our best knowledge, there haven't been any previous studies which trained the network for apparel image classification based on runway image dataset. We suggest the idea of training model with image capturing all the possible postures, which is denoted as mobility, by using our own runway apparel image dataset. Moreover, by applying Transfer Learning and using checkpoint and parameters provided by Tensorflow Slim, we could save time spent on training the classification model as taking 6 minutes per experiment to train the classifier. This model can be used in many business applications where the query image can be runway image, product image, or street fashion image. To be specific, runway query image can be used for mobile application service during fashion week to facilitate brand search, street style query image can be classified during fashion editorial task to classify and label the brand or style, and website query image can be processed by e-commerce multi-complex service providing item information or recommending similar item.

The Ontology Based, the Movie Contents Recommendation Scheme, Using Relations of Movie Metadata (온톨로지 기반 영화 메타데이터간 연관성을 활용한 영화 추천 기법)

  • Kim, Jaeyoung;Lee, Seok-Won
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.3
    • /
    • pp.25-44
    • /
    • 2013
  • Accessing movie contents has become easier and increased with the advent of smart TV, IPTV and web services that are able to be used to search and watch movies. In this situation, there are increasing search for preference movie contents of users. However, since the amount of provided movie contents is too large, the user needs more effort and time for searching the movie contents. Hence, there are a lot of researches for recommendations of personalized item through analysis and clustering of the user preferences and user profiles. In this study, we propose recommendation system which uses ontology based knowledge base. Our ontology can represent not only relations between metadata of movies but also relations between metadata and profile of user. The relation of each metadata can show similarity between movies. In order to build, the knowledge base our ontology model is considered two aspects which are the movie metadata model and the user model. On the part of build the movie metadata model based on ontology, we decide main metadata that are genre, actor/actress, keywords and synopsis. Those affect that users choose the interested movie. And there are demographic information of user and relation between user and movie metadata in user model. In our model, movie ontology model consists of seven concepts (Movie, Genre, Keywords, Synopsis Keywords, Character, and Person), eight attributes (title, rating, limit, description, character name, character description, person job, person name) and ten relations between concepts. For our knowledge base, we input individual data of 14,374 movies for each concept in contents ontology model. This movie metadata knowledge base is used to search the movie that is related to interesting metadata of user. And it can search the similar movie through relations between concepts. We also propose the architecture for movie recommendation. The proposed architecture consists of four components. The first component search candidate movies based the demographic information of the user. In this component, we decide the group of users according to demographic information to recommend the movie for each group and define the rule to decide the group of users. We generate the query that be used to search the candidate movie for recommendation in this component. The second component search candidate movies based user preference. When users choose the movie, users consider metadata such as genre, actor/actress, synopsis, keywords. Users input their preference and then in this component, system search the movie based on users preferences. The proposed system can search the similar movie through relation between concepts, unlike existing movie recommendation systems. Each metadata of recommended candidate movies have weight that will be used for deciding recommendation order. The third component the merges results of first component and second component. In this step, we calculate the weight of movies using the weight value of metadata for each movie. Then we sort movies order by the weight value. The fourth component analyzes result of third component, and then it decides level of the contribution of metadata. And we apply contribution weight to metadata. Finally, we use the result of this step as recommendation for users. We test the usability of the proposed scheme by using web application. We implement that web application for experimental process by using JSP, Java Script and prot$\acute{e}$g$\acute{e}$ API. In our experiment, we collect results of 20 men and woman, ranging in age from 20 to 29. And we use 7,418 movies with rating that is not fewer than 7.0. In order to experiment, we provide Top-5, Top-10 and Top-20 recommended movies to user, and then users choose interested movies. The result of experiment is that average number of to choose interested movie are 2.1 in Top-5, 3.35 in Top-10, 6.35 in Top-20. It is better than results that are yielded by for each metadata.

Bone mineral density and nutritional state according to milk consumption in Korean postmenopausal women who drink coffee: Using the 2008~2009 Korea National Health and Nutrition Examination Survey (한국 폐경 후 여성 커피소비자에서 우유섭취여부에 따른 골밀도와 영양상태 비교 : 2008~2009년 국민건강영양조사 자료 이용)

  • Ryu, Sun-Hyoung;Suh, Yoon Suk
    • Journal of Nutrition and Health
    • /
    • v.49 no.5
    • /
    • pp.347-357
    • /
    • 2016
  • Purpose: This study investigated bone mineral density and nutritional state according to consumption of milk in Korean postmenopausal women who drink coffee. Methods: Using the 2008~2009 Korean National Health & Nutrition Examination Survey data, a total of 1,373 postmenopausal females aged 50 yrs and over were analyzed after excluding those with diseases related to bone health. According to coffee and/or milk consumption, subjects were divided into four groups: coffee only, both coffee & milk, milk only, and none of the above. All data were processed after application of weighted values and adjustment of age, body mass index, physical activity, drinking, and smoking using a general linear model. For analysis of nutrient intake and bone density, data were additionally adjusted by total energy and calcium intake. Results: The coffee & milk group had more subjects younger than 65 yrs and higher education, urban residents, and higher income than any other group. The coffee only group showed somewhat similar characteristics as the none of the above group, which showed the highest percentage of subjects older than 65 and in a lower education and socio-economic state. Body weight, height, body mass index, and lean mass were the highest in coffee & milk group and lowest in the none of the above group. On the other hand, the milk only group showed the lowest values for body mass index and waist circumference, whereas percent body fat did not show any difference among the groups. The coffee and milk group showed the highest bone mineral density in the total femur and lumbar spine as well as the highest nutritional state and most food group intakes, followed by the milk only group, coffee only group, and none of the above group. In the assessment of osteoporosis based on T-score of bone mineral density, although not significant, the coffee and milk group and milk only group, which showed a better nutritional state, included more subjects with a normal bone density, whereas the none of the above group included more subjects with osteoporosis than any other group. Conclusion: Bone mineral density in postmenopausal women might not be affected by coffee drinking if their diets are accompanied by balanced food and nutrient intake including milk.

Development of a complex failure prediction system using Hierarchical Attention Network (Hierarchical Attention Network를 이용한 복합 장애 발생 예측 시스템 개발)

  • Park, Youngchan;An, Sangjun;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.127-148
    • /
    • 2020
  • The data center is a physical environment facility for accommodating computer systems and related components, and is an essential foundation technology for next-generation core industries such as big data, smart factories, wearables, and smart homes. In particular, with the growth of cloud computing, the proportional expansion of the data center infrastructure is inevitable. Monitoring the health of these data center facilities is a way to maintain and manage the system and prevent failure. If a failure occurs in some elements of the facility, it may affect not only the relevant equipment but also other connected equipment, and may cause enormous damage. In particular, IT facilities are irregular due to interdependence and it is difficult to know the cause. In the previous study predicting failure in data center, failure was predicted by looking at a single server as a single state without assuming that the devices were mixed. Therefore, in this study, data center failures were classified into failures occurring inside the server (Outage A) and failures occurring outside the server (Outage B), and focused on analyzing complex failures occurring within the server. Server external failures include power, cooling, user errors, etc. Since such failures can be prevented in the early stages of data center facility construction, various solutions are being developed. On the other hand, the cause of the failure occurring in the server is difficult to determine, and adequate prevention has not yet been achieved. In particular, this is the reason why server failures do not occur singularly, cause other server failures, or receive something that causes failures from other servers. In other words, while the existing studies assumed that it was a single server that did not affect the servers and analyzed the failure, in this study, the failure occurred on the assumption that it had an effect between servers. In order to define the complex failure situation in the data center, failure history data for each equipment existing in the data center was used. There are four major failures considered in this study: Network Node Down, Server Down, Windows Activation Services Down, and Database Management System Service Down. The failures that occur for each device are sorted in chronological order, and when a failure occurs in a specific equipment, if a failure occurs in a specific equipment within 5 minutes from the time of occurrence, it is defined that the failure occurs simultaneously. After configuring the sequence for the devices that have failed at the same time, 5 devices that frequently occur simultaneously within the configured sequence were selected, and the case where the selected devices failed at the same time was confirmed through visualization. Since the server resource information collected for failure analysis is in units of time series and has flow, we used Long Short-term Memory (LSTM), a deep learning algorithm that can predict the next state through the previous state. In addition, unlike a single server, the Hierarchical Attention Network deep learning model structure was used in consideration of the fact that the level of multiple failures for each server is different. This algorithm is a method of increasing the prediction accuracy by giving weight to the server as the impact on the failure increases. The study began with defining the type of failure and selecting the analysis target. In the first experiment, the same collected data was assumed as a single server state and a multiple server state, and compared and analyzed. The second experiment improved the prediction accuracy in the case of a complex server by optimizing each server threshold. In the first experiment, which assumed each of a single server and multiple servers, in the case of a single server, it was predicted that three of the five servers did not have a failure even though the actual failure occurred. However, assuming multiple servers, all five servers were predicted to have failed. As a result of the experiment, the hypothesis that there is an effect between servers is proven. As a result of this study, it was confirmed that the prediction performance was superior when the multiple servers were assumed than when the single server was assumed. In particular, applying the Hierarchical Attention Network algorithm, assuming that the effects of each server will be different, played a role in improving the analysis effect. In addition, by applying a different threshold for each server, the prediction accuracy could be improved. This study showed that failures that are difficult to determine the cause can be predicted through historical data, and a model that can predict failures occurring in servers in data centers is presented. It is expected that the occurrence of disability can be prevented in advance using the results of this study.

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.