• Title/Summary/Keyword: Systems Improvement

Search Result 6,054, Processing Time 0.034 seconds

Query-based Answer Extraction using Korean Dependency Parsing (의존 구문 분석을 이용한 질의 기반 정답 추출)

  • Lee, Dokyoung;Kim, Mintae;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.161-177
    • /
    • 2019
  • In this paper, we study the performance improvement of the answer extraction in Question-Answering system by using sentence dependency parsing result. The Question-Answering (QA) system consists of query analysis, which is a method of analyzing the user's query, and answer extraction, which is a method to extract appropriate answers in the document. And various studies have been conducted on two methods. In order to improve the performance of answer extraction, it is necessary to accurately reflect the grammatical information of sentences. In Korean, because word order structure is free and omission of sentence components is frequent, dependency parsing is a good way to analyze Korean syntax. Therefore, in this study, we improved the performance of the answer extraction by adding the features generated by dependency parsing analysis to the inputs of the answer extraction model (Bidirectional LSTM-CRF). The process of generating the dependency graph embedding consists of the steps of generating the dependency graph from the dependency parsing result and learning the embedding of the graph. In this study, we compared the performance of the answer extraction model when inputting basic word features generated without the dependency parsing and the performance of the model when inputting the addition of the Eojeol tag feature and dependency graph embedding feature. Since dependency parsing is performed on a basic unit of an Eojeol, which is a component of sentences separated by a space, the tag information of the Eojeol can be obtained as a result of the dependency parsing. The Eojeol tag feature means the tag information of the Eojeol. The process of generating the dependency graph embedding consists of the steps of generating the dependency graph from the dependency parsing result and learning the embedding of the graph. From the dependency parsing result, a graph is generated from the Eojeol to the node, the dependency between the Eojeol to the edge, and the Eojeol tag to the node label. In this process, an undirected graph is generated or a directed graph is generated according to whether or not the dependency relation direction is considered. To obtain the embedding of the graph, we used Graph2Vec, which is a method of finding the embedding of the graph by the subgraphs constituting a graph. We can specify the maximum path length between nodes in the process of finding subgraphs of a graph. If the maximum path length between nodes is 1, graph embedding is generated only by direct dependency between Eojeol, and graph embedding is generated including indirect dependencies as the maximum path length between nodes becomes larger. In the experiment, the maximum path length between nodes is adjusted differently from 1 to 3 depending on whether direction of dependency is considered or not, and the performance of answer extraction is measured. Experimental results show that both Eojeol tag feature and dependency graph embedding feature improve the performance of answer extraction. In particular, considering the direction of the dependency relation and extracting the dependency graph generated with the maximum path length of 1 in the subgraph extraction process in Graph2Vec as the input of the model, the highest answer extraction performance was shown. As a result of these experiments, we concluded that it is better to take into account the direction of dependence and to consider only the direct connection rather than the indirect dependence between the words. The significance of this study is as follows. First, we improved the performance of answer extraction by adding features using dependency parsing results, taking into account the characteristics of Korean, which is free of word order structure and omission of sentence components. Second, we generated feature of dependency parsing result by learning - based graph embedding method without defining the pattern of dependency between Eojeol. Future research directions are as follows. In this study, the features generated as a result of the dependency parsing are applied only to the answer extraction model in order to grasp the meaning. However, in the future, if the performance is confirmed by applying the features to various natural language processing models such as sentiment analysis or name entity recognition, the validity of the features can be verified more accurately.

An Analytical Approach Using Topic Mining for Improving the Service Quality of Hotels (호텔 산업의 서비스 품질 향상을 위한 토픽 마이닝 기반 분석 방법)

  • Moon, Hyun Sil;Sung, David;Kim, Jae Kyeong
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.21-41
    • /
    • 2019
  • Thanks to the rapid development of information technologies, the data available on Internet have grown rapidly. In this era of big data, many studies have attempted to offer insights and express the effects of data analysis. In the tourism and hospitality industry, many firms and studies in the era of big data have paid attention to online reviews on social media because of their large influence over customers. As tourism is an information-intensive industry, the effect of these information networks on social media platforms is more remarkable compared to any other types of media. However, there are some limitations to the improvements in service quality that can be made based on opinions on social media platforms. Users on social media platforms represent their opinions as text, images, and so on. Raw data sets from these reviews are unstructured. Moreover, these data sets are too big to extract new information and hidden knowledge by human competences. To use them for business intelligence and analytics applications, proper big data techniques like Natural Language Processing and data mining techniques are needed. This study suggests an analytical approach to directly yield insights from these reviews to improve the service quality of hotels. Our proposed approach consists of topic mining to extract topics contained in the reviews and the decision tree modeling to explain the relationship between topics and ratings. Topic mining refers to a method for finding a group of words from a collection of documents that represents a document. Among several topic mining methods, we adopted the Latent Dirichlet Allocation algorithm, which is considered as the most universal algorithm. However, LDA is not enough to find insights that can improve service quality because it cannot find the relationship between topics and ratings. To overcome this limitation, we also use the Classification and Regression Tree method, which is a kind of decision tree technique. Through the CART method, we can find what topics are related to positive or negative ratings of a hotel and visualize the results. Therefore, this study aims to investigate the representation of an analytical approach for the improvement of hotel service quality from unstructured review data sets. Through experiments for four hotels in Hong Kong, we can find the strengths and weaknesses of services for each hotel and suggest improvements to aid in customer satisfaction. Especially from positive reviews, we find what these hotels should maintain for service quality. For example, compared with the other hotels, a hotel has a good location and room condition which are extracted from positive reviews for it. In contrast, we also find what they should modify in their services from negative reviews. For example, a hotel should improve room condition related to soundproof. These results mean that our approach is useful in finding some insights for the service quality of hotels. That is, from the enormous size of review data, our approach can provide practical suggestions for hotel managers to improve their service quality. In the past, studies for improving service quality relied on surveys or interviews of customers. However, these methods are often costly and time consuming and the results may be biased by biased sampling or untrustworthy answers. The proposed approach directly obtains honest feedback from customers' online reviews and draws some insights through a type of big data analysis. So it will be a more useful tool to overcome the limitations of surveys or interviews. Moreover, our approach easily obtains the service quality information of other hotels or services in the tourism industry because it needs only open online reviews and ratings as input data. Furthermore, the performance of our approach will be better if other structured and unstructured data sources are added.

A Study of Anomaly Detection for ICT Infrastructure using Conditional Multimodal Autoencoder (ICT 인프라 이상탐지를 위한 조건부 멀티모달 오토인코더에 관한 연구)

  • Shin, Byungjin;Lee, Jonghoon;Han, Sangjin;Park, Choong-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.57-73
    • /
    • 2021
  • Maintenance and prevention of failure through anomaly detection of ICT infrastructure is becoming important. System monitoring data is multidimensional time series data. When we deal with multidimensional time series data, we have difficulty in considering both characteristics of multidimensional data and characteristics of time series data. When dealing with multidimensional data, correlation between variables should be considered. Existing methods such as probability and linear base, distance base, etc. are degraded due to limitations called the curse of dimensions. In addition, time series data is preprocessed by applying sliding window technique and time series decomposition for self-correlation analysis. These techniques are the cause of increasing the dimension of data, so it is necessary to supplement them. The anomaly detection field is an old research field, and statistical methods and regression analysis were used in the early days. Currently, there are active studies to apply machine learning and artificial neural network technology to this field. Statistically based methods are difficult to apply when data is non-homogeneous, and do not detect local outliers well. The regression analysis method compares the predictive value and the actual value after learning the regression formula based on the parametric statistics and it detects abnormality. Anomaly detection using regression analysis has the disadvantage that the performance is lowered when the model is not solid and the noise or outliers of the data are included. There is a restriction that learning data with noise or outliers should be used. The autoencoder using artificial neural networks is learned to output as similar as possible to input data. It has many advantages compared to existing probability and linear model, cluster analysis, and map learning. It can be applied to data that does not satisfy probability distribution or linear assumption. In addition, it is possible to learn non-mapping without label data for teaching. However, there is a limitation of local outlier identification of multidimensional data in anomaly detection, and there is a problem that the dimension of data is greatly increased due to the characteristics of time series data. In this study, we propose a CMAE (Conditional Multimodal Autoencoder) that enhances the performance of anomaly detection by considering local outliers and time series characteristics. First, we applied Multimodal Autoencoder (MAE) to improve the limitations of local outlier identification of multidimensional data. Multimodals are commonly used to learn different types of inputs, such as voice and image. The different modal shares the bottleneck effect of Autoencoder and it learns correlation. In addition, CAE (Conditional Autoencoder) was used to learn the characteristics of time series data effectively without increasing the dimension of data. In general, conditional input mainly uses category variables, but in this study, time was used as a condition to learn periodicity. The CMAE model proposed in this paper was verified by comparing with the Unimodal Autoencoder (UAE) and Multi-modal Autoencoder (MAE). The restoration performance of Autoencoder for 41 variables was confirmed in the proposed model and the comparison model. The restoration performance is different by variables, and the restoration is normally well operated because the loss value is small for Memory, Disk, and Network modals in all three Autoencoder models. The process modal did not show a significant difference in all three models, and the CPU modal showed excellent performance in CMAE. ROC curve was prepared for the evaluation of anomaly detection performance in the proposed model and the comparison model, and AUC, accuracy, precision, recall, and F1-score were compared. In all indicators, the performance was shown in the order of CMAE, MAE, and AE. Especially, the reproduction rate was 0.9828 for CMAE, which can be confirmed to detect almost most of the abnormalities. The accuracy of the model was also improved and 87.12%, and the F1-score was 0.8883, which is considered to be suitable for anomaly detection. In practical aspect, the proposed model has an additional advantage in addition to performance improvement. The use of techniques such as time series decomposition and sliding windows has the disadvantage of managing unnecessary procedures; and their dimensional increase can cause a decrease in the computational speed in inference.The proposed model has characteristics that are easy to apply to practical tasks such as inference speed and model management.

Developing the Process and Characteristics of Preservation of Area-Based Heritage Sites in Japan (일본 면형 유산 보존제도의 확산과정과 특성)

  • Sung, Wonseok;Kang, Dongjin
    • Korean Journal of Heritage: History & Science
    • /
    • v.53 no.4
    • /
    • pp.32-59
    • /
    • 2020
  • South Korea's area-based heritage preservation system originates from the "Preservation of Traditional Buildings Act" enacted in 1984. However, this system was abolished in 1996. As there was a need for protection of ancient cities in the 1960s, Japan enacted the Historic City Preservation Act in 1966, and 'Preservation Areas for Historic Landscapes' and 'Special Preservation Districts for Historic Landscapes' were introduced. For the preservation of area-based heritage sites, the 'Important Preservation Districts for Groups of Traditional Buildings' system introduced as part of the revision of the Cultural Heritage Protection Act in 1975 was the beginning. Then, in the early-2000s, discussions on the preservation of area-based heritage sites began in earnest, and the 'Important Cultural Landscape' system was introduced for protection of the space and context between heritage sites. Also, '33 Groups of Modernization Industry Heritage Sites' were designated in 2007, covering various material and immaterial resources related to the modernization of Japan, and '100 Beautiful Historic Landscapes of Japan' were selected for protection of local landscapes with historic value in the same year. In 2015, the "Japanese Heritage" system was established for the integrated preservation and management of tangible and intangible heritage aspects located in specific areas; in 2016, the "Japanese Agricultural Heritage" system was established for the succession and fostering of the disappearing agriculture and fishery industries; and in 2017, "the 20th Century Heritage," was established, representing evidence of modern and contemporary Japanese technologies in the 20th century. As a result, presently (in September 2020), 30 'Historic Landscape Preservation Areas', 60 'Historic Landscape Special Districts,' 120 'Important Preservation Districts for Groups of Traditional Buildings," 65 'Important Cultural Landscapes,' 66 'Groups of Modernization Industry Heritage Sites,' 264 "100 Beautiful Historic Landscapes of Japan,' 104 'Japanese Heritage Sites,' and 15 'Japanese Agricultural Heritage Sites' have been designated. According to this perception of situations, the research process for this study with its basic purpose of extracting the general characteristics of Japan's area-based heritage preservation system, has sequentially spread since 1976 as follows. First, this study investigates Japan's area-based heritage site preservation system and sets the scope of research through discussions of literature and preceding studies. Second, this study investigates the process of the spread of the area-based heritage site preservation system and analyzes the relationship between the systems according to their development, in order to draw upon their characteristics. Third, to concretize content related to relationships and characteristics, this study involves in-depth analysis of three representative examples and sums them up to identify the characteristics of Japan's area-based heritage system. A noticeable characteristic of Japan's area-based heritage site preservation system drawn from this is that new heritage sites are born each year. Consequently, an overlapping phenomenon takes place between heritage sites, and such phenomena occur alongside revitalization of related industries, traditional industry, and cultural tourism and the improvement of localities as well as the preservation of area-based heritage. These characteristics can be applied as suggestions for the revitalization of the 'modern historical and cultural space' system implemented by South Korea.

A Methodology of Customer Churn Prediction based on Two-Dimensional Loyalty Segmentation (이차원 고객충성도 세그먼트 기반의 고객이탈예측 방법론)

  • Kim, Hyung Su;Hong, Seung Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.111-126
    • /
    • 2020
  • Most industries have recently become aware of the importance of customer lifetime value as they are exposed to a competitive environment. As a result, preventing customers from churn is becoming a more important business issue than securing new customers. This is because maintaining churn customers is far more economical than securing new customers, and in fact, the acquisition cost of new customers is known to be five to six times higher than the maintenance cost of churn customers. Also, Companies that effectively prevent customer churn and improve customer retention rates are known to have a positive effect on not only increasing the company's profitability but also improving its brand image by improving customer satisfaction. Predicting customer churn, which had been conducted as a sub-research area for CRM, has recently become more important as a big data-based performance marketing theme due to the development of business machine learning technology. Until now, research on customer churn prediction has been carried out actively in such sectors as the mobile telecommunication industry, the financial industry, the distribution industry, and the game industry, which are highly competitive and urgent to manage churn. In addition, These churn prediction studies were focused on improving the performance of the churn prediction model itself, such as simply comparing the performance of various models, exploring features that are effective in forecasting departures, or developing new ensemble techniques, and were limited in terms of practical utilization because most studies considered the entire customer group as a group and developed a predictive model. As such, the main purpose of the existing related research was to improve the performance of the predictive model itself, and there was a relatively lack of research to improve the overall customer churn prediction process. In fact, customers in the business have different behavior characteristics due to heterogeneous transaction patterns, and the resulting churn rate is different, so it is unreasonable to assume the entire customer as a single customer group. Therefore, it is desirable to segment customers according to customer classification criteria, such as loyalty, and to operate an appropriate churn prediction model individually, in order to carry out effective customer churn predictions in heterogeneous industries. Of course, in some studies, there are studies in which customers are subdivided using clustering techniques and applied a churn prediction model for individual customer groups. Although this process of predicting churn can produce better predictions than a single predict model for the entire customer population, there is still room for improvement in that clustering is a mechanical, exploratory grouping technique that calculates distances based on inputs and does not reflect the strategic intent of an entity such as loyalties. This study proposes a segment-based customer departure prediction process (CCP/2DL: Customer Churn Prediction based on Two-Dimensional Loyalty segmentation) based on two-dimensional customer loyalty, assuming that successful customer churn management can be better done through improvements in the overall process than through the performance of the model itself. CCP/2DL is a series of churn prediction processes that segment two-way, quantitative and qualitative loyalty-based customer, conduct secondary grouping of customer segments according to churn patterns, and then independently apply heterogeneous churn prediction models for each churn pattern group. Performance comparisons were performed with the most commonly applied the General churn prediction process and the Clustering-based churn prediction process to assess the relative excellence of the proposed churn prediction process. The General churn prediction process used in this study refers to the process of predicting a single group of customers simply intended to be predicted as a machine learning model, using the most commonly used churn predicting method. And the Clustering-based churn prediction process is a method of first using clustering techniques to segment customers and implement a churn prediction model for each individual group. In cooperation with a global NGO, the proposed CCP/2DL performance showed better performance than other methodologies for predicting churn. This churn prediction process is not only effective in predicting churn, but can also be a strategic basis for obtaining a variety of customer observations and carrying out other related performance marketing activities.

Deriving adoption strategies of deep learning open source framework through case studies (딥러닝 오픈소스 프레임워크의 사례연구를 통한 도입 전략 도출)

  • Choi, Eunjoo;Lee, Junyeong;Han, Ingoo
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.4
    • /
    • pp.27-65
    • /
    • 2020
  • Many companies on information and communication technology make public their own developed AI technology, for example, Google's TensorFlow, Facebook's PyTorch, Microsoft's CNTK. By releasing deep learning open source software to the public, the relationship with the developer community and the artificial intelligence (AI) ecosystem can be strengthened, and users can perform experiment, implementation and improvement of it. Accordingly, the field of machine learning is growing rapidly, and developers are using and reproducing various learning algorithms in each field. Although various analysis of open source software has been made, there is a lack of studies to help develop or use deep learning open source software in the industry. This study thus attempts to derive a strategy for adopting the framework through case studies of a deep learning open source framework. Based on the technology-organization-environment (TOE) framework and literature review related to the adoption of open source software, we employed the case study framework that includes technological factors as perceived relative advantage, perceived compatibility, perceived complexity, and perceived trialability, organizational factors as management support and knowledge & expertise, and environmental factors as availability of technology skills and services, and platform long term viability. We conducted a case study analysis of three companies' adoption cases (two cases of success and one case of failure) and revealed that seven out of eight TOE factors and several factors regarding company, team and resource are significant for the adoption of deep learning open source framework. By organizing the case study analysis results, we provided five important success factors for adopting deep learning framework: the knowledge and expertise of developers in the team, hardware (GPU) environment, data enterprise cooperation system, deep learning framework platform, deep learning framework work tool service. In order for an organization to successfully adopt a deep learning open source framework, at the stage of using the framework, first, the hardware (GPU) environment for AI R&D group must support the knowledge and expertise of the developers in the team. Second, it is necessary to support the use of deep learning frameworks by research developers through collecting and managing data inside and outside the company with a data enterprise cooperation system. Third, deep learning research expertise must be supplemented through cooperation with researchers from academic institutions such as universities and research institutes. Satisfying three procedures in the stage of using the deep learning framework, companies will increase the number of deep learning research developers, the ability to use the deep learning framework, and the support of GPU resource. In the proliferation stage of the deep learning framework, fourth, a company makes the deep learning framework platform that improves the research efficiency and effectiveness of the developers, for example, the optimization of the hardware (GPU) environment automatically. Fifth, the deep learning framework tool service team complements the developers' expertise through sharing the information of the external deep learning open source framework community to the in-house community and activating developer retraining and seminars. To implement the identified five success factors, a step-by-step enterprise procedure for adoption of the deep learning framework was proposed: defining the project problem, confirming whether the deep learning methodology is the right method, confirming whether the deep learning framework is the right tool, using the deep learning framework by the enterprise, spreading the framework of the enterprise. The first three steps (i.e. defining the project problem, confirming whether the deep learning methodology is the right method, and confirming whether the deep learning framework is the right tool) are pre-considerations to adopt a deep learning open source framework. After the three pre-considerations steps are clear, next two steps (i.e. using the deep learning framework by the enterprise and spreading the framework of the enterprise) can be processed. In the fourth step, the knowledge and expertise of developers in the team are important in addition to hardware (GPU) environment and data enterprise cooperation system. In final step, five important factors are realized for a successful adoption of the deep learning open source framework. This study provides strategic implications for companies adopting or using deep learning framework according to the needs of each industry and business.

A Study on the Characteristics of Enterprise R&D Capabilities Using Data Mining (데이터마이닝을 활용한 기업 R&D역량 특성에 관한 탐색 연구)

  • Kim, Sang-Gook;Lim, Jung-Sun;Park, Wan
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.1-21
    • /
    • 2021
  • As the global business environment changes, uncertainties in technology development and market needs increase, and competition among companies intensifies, interests and demands for R&D activities of individual companies are increasing. In order to cope with these environmental changes, R&D companies are strengthening R&D investment as one of the means to enhance the qualitative competitiveness of R&D while paying more attention to facility investment. As a result, facilities or R&D investment elements are inevitably a burden for R&D companies to bear future uncertainties. It is true that the management strategy of increasing investment in R&D as a means of enhancing R&D capability is highly uncertain in terms of corporate performance. In this study, the structural factors that influence the R&D capabilities of companies are explored in terms of technology management capabilities, R&D capabilities, and corporate classification attributes by utilizing data mining techniques, and the characteristics these individual factors present according to the level of R&D capabilities are analyzed. This study also showed cluster analysis and experimental results based on evidence data for all domestic R&D companies, and is expected to provide important implications for corporate management strategies to enhance R&D capabilities of individual companies. For each of the three viewpoints, detailed evaluation indexes were composed of 7, 2, and 4, respectively, to quantitatively measure individual levels in the corresponding area. In the case of technology management capability and R&D capability, the sub-item evaluation indexes that are being used by current domestic technology evaluation agencies were referenced, and the final detailed evaluation index was newly constructed in consideration of whether data could be obtained quantitatively. In the case of corporate classification attributes, the most basic corporate classification profile information is considered. In particular, in order to grasp the homogeneity of the R&D competency level, a comprehensive score for each company was given using detailed evaluation indicators of technology management capability and R&D capability, and the competency level was classified into five grades and compared with the cluster analysis results. In order to give the meaning according to the comparative evaluation between the analyzed cluster and the competency level grade, the clusters with high and low trends in R&D competency level were searched for each cluster. Afterwards, characteristics according to detailed evaluation indicators were analyzed in the cluster. Through this method of conducting research, two groups with high R&D competency and one with low level of R&D competency were analyzed, and the remaining two clusters were similar with almost high incidence. As a result, in this study, individual characteristics according to detailed evaluation indexes were analyzed for two clusters with high competency level and one cluster with low competency level. The implications of the results of this study are that the faster the replacement cycle of professional managers who can effectively respond to changes in technology and market demand, the more likely they will contribute to enhancing R&D capabilities. In the case of a private company, it is necessary to increase the intensity of input of R&D capabilities by enhancing the sense of belonging of R&D personnel to the company through conversion to a corporate company, and to provide the accuracy of responsibility and authority through the organization of the team unit. Since the number of technical commercialization achievements and technology certifications are occurring both in the case of contributing to capacity improvement and in case of not, it was confirmed that there is a limit in reviewing it as an important factor for enhancing R&D capacity from the perspective of management. Lastly, the experience of utility model filing was identified as a factor that has an important influence on R&D capability, and it was confirmed the need to provide motivation to encourage utility model filings in order to enhance R&D capability. As such, the results of this study are expected to provide important implications for corporate management strategies to enhance individual companies' R&D capabilities.

Transfer Learning using Multiple ConvNet Layers Activation Features with Principal Component Analysis for Image Classification (전이학습 기반 다중 컨볼류션 신경망 레이어의 활성화 특징과 주성분 분석을 이용한 이미지 분류 방법)

  • Byambajav, Batkhuu;Alikhanov, Jumabek;Fang, Yang;Ko, Seunghyun;Jo, Geun Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.1
    • /
    • pp.205-225
    • /
    • 2018
  • Convolutional Neural Network (ConvNet) is one class of the powerful Deep Neural Network that can analyze and learn hierarchies of visual features. Originally, first neural network (Neocognitron) was introduced in the 80s. At that time, the neural network was not broadly used in both industry and academic field by cause of large-scale dataset shortage and low computational power. However, after a few decades later in 2012, Krizhevsky made a breakthrough on ILSVRC-12 visual recognition competition using Convolutional Neural Network. That breakthrough revived people interest in the neural network. The success of Convolutional Neural Network is achieved with two main factors. First of them is the emergence of advanced hardware (GPUs) for sufficient parallel computation. Second is the availability of large-scale datasets such as ImageNet (ILSVRC) dataset for training. Unfortunately, many new domains are bottlenecked by these factors. For most domains, it is difficult and requires lots of effort to gather large-scale dataset to train a ConvNet. Moreover, even if we have a large-scale dataset, training ConvNet from scratch is required expensive resource and time-consuming. These two obstacles can be solved by using transfer learning. Transfer learning is a method for transferring the knowledge from a source domain to new domain. There are two major Transfer learning cases. First one is ConvNet as fixed feature extractor, and the second one is Fine-tune the ConvNet on a new dataset. In the first case, using pre-trained ConvNet (such as on ImageNet) to compute feed-forward activations of the image into the ConvNet and extract activation features from specific layers. In the second case, replacing and retraining the ConvNet classifier on the new dataset, then fine-tune the weights of the pre-trained network with the backpropagation. In this paper, we focus on using multiple ConvNet layers as a fixed feature extractor only. However, applying features with high dimensional complexity that is directly extracted from multiple ConvNet layers is still a challenging problem. We observe that features extracted from multiple ConvNet layers address the different characteristics of the image which means better representation could be obtained by finding the optimal combination of multiple ConvNet layers. Based on that observation, we propose to employ multiple ConvNet layer representations for transfer learning instead of a single ConvNet layer representation. Overall, our primary pipeline has three steps. Firstly, images from target task are given as input to ConvNet, then that image will be feed-forwarded into pre-trained AlexNet, and the activation features from three fully connected convolutional layers are extracted. Secondly, activation features of three ConvNet layers are concatenated to obtain multiple ConvNet layers representation because it will gain more information about an image. When three fully connected layer features concatenated, the occurring image representation would have 9192 (4096+4096+1000) dimension features. However, features extracted from multiple ConvNet layers are redundant and noisy since they are extracted from the same ConvNet. Thus, a third step, we will use Principal Component Analysis (PCA) to select salient features before the training phase. When salient features are obtained, the classifier can classify image more accurately, and the performance of transfer learning can be improved. To evaluate proposed method, experiments are conducted in three standard datasets (Caltech-256, VOC07, and SUN397) to compare multiple ConvNet layer representations against single ConvNet layer representation by using PCA for feature selection and dimension reduction. Our experiments demonstrated the importance of feature selection for multiple ConvNet layer representation. Moreover, our proposed approach achieved 75.6% accuracy compared to 73.9% accuracy achieved by FC7 layer on the Caltech-256 dataset, 73.1% accuracy compared to 69.2% accuracy achieved by FC8 layer on the VOC07 dataset, 52.2% accuracy compared to 48.7% accuracy achieved by FC7 layer on the SUN397 dataset. We also showed that our proposed approach achieved superior performance, 2.8%, 2.1% and 3.1% accuracy improvement on Caltech-256, VOC07, and SUN397 dataset respectively compare to existing work.

The effects of aqueous extracts of plant roots on germination of seeds and growth of seedings (식물근의 추출물질이 종자발아 및 유식물의 생장에 미치는 영향)

  • Chan-Ho Park
    • KOREAN JOURNAL OF CROP SCIENCE
    • /
    • v.4 no.1
    • /
    • pp.1-23
    • /
    • 1968
  • This study aimed at contributing to the improvement of cropping systems after finding out the effects of excrements and components of crop root influence on other crops as well as themselves. The following forage crops suitable for our country were selected for the present study. Aqueous extracts of fresh roots, aqueous extracts of rotting roots and aqueous solutions of excrements of red clover, orchard grass and brome grass were studied for the effects influencing the germination and growth of seedlings of red clover, ladino clover, lespedeza, soybean, orchard grass, Italian ryegrass, brome grass, barley, wheat, sorghum, corn and Hog-millet. In view of the possibility that the organic acid might be closely related to the excrements and components of crop root connected with soil sickness, the acid components of three species of roots were analysed by paper chromatography and gas chromatography method. The following results were obtained: 1. Effects of Aqueous Extracts of Fresh Roots : Aqueous extracts of red clover: The extracts inhibited the growth of seedlings of the ladino clover and lespedeza and also inhibited the development of most crops except that of sorghum among the Graminaceae. Aqueous extracts of orchard grass: The extracts promoted the seedlings growth of red clover and soybean, while it inhibited the germination and growth of orchard grass. There were no noticeable effects influencing other crops while it inhibited the growth of barley and Hog-millet. Aqueous extracts of brome grass: There was no effect on Italian ryegrass but there was an inhibiting effect on the other crops. 2. Effects of Aqueous Extracts of Rotting Roots : Aqueous extracts of red clover: The extracts promoted the seedling growth of red clover. But it reflected the inhibiting effects on other crops except sorghum. Aqueous extracts of orchard grass: The extracts promoted the growth of red clover, ladino clover, soybean and sorghun, while it inhibited the germination and rooting of barley and Hog-millet. Aqueous extracts of brome grass: The extracts gave the promotive effects to the growth of red clover, soybean and sorghum, but caused inhibiting effects on orchard grass, brome grass, barley and Hog-millet. 3. Effects of Aqueous Solutions of Excrements : The aqueous solution of excrements of red clover reflected the inhibition effects to the growth of Graminaceae, while the aqueous solutions of excrements of orchard grass and Italian ryegrass caused the promotive effects on the growth of red clover. 4. Results of Organic Acid Analysis : The oxalic acid, citric acid, tartaric acid, malonic acid, malic acid and succinic acid were included in the roots of red clover as unvolatile organic acid, and in the orchard grass and brome grass there were included the oxalic acid, citric acid, tartaric acid and malic acid. And formic acid was confirmed in the red clover, orchard grass and brome grass as volatile organic acid. In consideration of the results mentioned in above the effects of excrements and components of roots found in this studies may be summarized as follows. 1) The red clover generally gave a disadvantageous effect on the Graminaceae. Such trend was considered chiefly caused by the presence of many organic acids, namely oxalic, citric, tartaric, malonic, malic, succinic and formic acid. 2) The orchard grass generally gave an advantageous effect on the Leguminosae. This may be due to a few kinds of organic acid contained in the root, namely oxalic, citric, tartaric, malic and formic acid. Furthermore a certain of promotive materials for growth was noted. 3) As long as the root of brome grass are not rotten, it gave a disadvantageous effect on the Leguminosae and Graminaceae. This may be due to the fact that several unidentified volatile organic acid were also included besides the confirmed organic acid, namely oxalic, citric, tartaric, malic and formic acid. 5. Effects of Components in Roots to the Soil Sickness : 1) It was considered that the cause of alleged red clover's soil sickness did not result from the toxic components of the roots. 2) It was recognized that the toxic components of roots might be the cause of soil sickness in case the orchard grass and brome grass were put into the long-term single cropping. 6. Effects of Rooted Components to the Companion Crops in the Cropping System : a) In case of aqueous extracts of fresh roots and aqueous excrements (Inter cropping and mixed cropping) : 1) Advantageous combinations : Orchard grass->Red clover, Soybean, Italian ryegrass->Red clover, 2) Disadvantageous combinations : Red clover->Ladino clover, Lespedeza, Orchard grass, Italian ryegrass, Fescue Ky-31, Brome grass, Barley, Wheat, Corn and Hog.millet, Orchard grass->Lespedeza, Orchard grass, Barley and Hog-millet, Brome grass->Red clover, Ladino clover, Lespedeza, Soybean, Orchard grass, Brome grass, Barley, Wheat, Sorghum, Corn and Hog-millet, 3) Harmless combinations : Red clover->Red clover, Soybean and Sorghum, Orchard grass->Ladino clover, Italian ryegrass, Brome grass, Wheat, Sorghum and Corn, Brome grass->Italian ryegrass, b) In case of aquecus extracts of rotting roots(After cropping) : 1) Advantageous combinations : Red clover->Red clover and Sorghum, Orchard grass->Red clover, Ladino clover, Soybean, Sorghum, and Corn, Brome grass->Red clover, Soybean and Sorghum, 2) Disadvantageous combinations : Red clover->Lespedeza, Orchard grass, Italian ryegrass, Brome grass, Barley, Wheat, and Hog-millet Orchard grass->Barley and Hog-millet, Brome grass->Orchard grass, Brome grass, Barley and Hog-millet, 3) Harmless combinations : Red clover->Ladino clover, Soybean and Corn, Orchard grass->Lespedeza, Orchard grass, Italian ryegrass, Brome grass and Wheat Brome gass->Ladino clover, Lespedeza, Italian ryegrass and Wheat.

  • PDF

Comparison of CT based-CTV plan and CT based-ICRU38 plan in Brachytherapy Planning of Uterine Cervix Cancer (자궁경부암 강내조사 시 CT를 이용한 CTV에 근거한 치료계획과 ICRU 38에 근거한 치료계획의 비교)

  • Cho, Jung-Ken;Han, Tae-Jong
    • Journal of Radiation Protection and Research
    • /
    • v.32 no.3
    • /
    • pp.105-110
    • /
    • 2007
  • Purpose : In spite of recent remarkable improvement of diagnostic imaging modalities such as CT, MRI, and PET and radiation therapy planing systems, ICR plan of uterine cervix cancer, based on recommendation of ICRU38(2D film-based) such as Point A, is still used widely. A 3-dimensional ICR plan based on CT image provides dose-volume histogram(DVH) information of the tumor and normal tissue. In this study, we compared tumor-dose, rectal-dose and bladder-dose through an analysis of DVH between CTV plan and ICRU38 plan based on CT image. Method and Material : We analyzed 11 patients with a cervix cancer who received the ICR of Ir-192 HDR. After 40Gy of external beam radiation therapy, ICR plan was established using PLATO(Nucletron) v.14.2 planing system. CT scan was done to all the patients using CT-simulator(Ultra Z, Philips). We contoured CTV, rectum and bladder on the CT image and established CTV plan which delivers the 100% dose to CTV and ICRU plan which delivers the 100% dose to the point A. Result : The volume$(average{\pm}SD)$ of CTV, rectum and bladder in all of 11 patients is $21.8{\pm}6.6cm^3,\;60.9{\pm}25.0cm^3,\;111.6{\pm}40.1cm^3$ respectively. The volume covered by 100% isodose curve is $126.7{\pm}18.9cm^3$ in ICRU plan and $98.2{\pm}74.5cm^3$ in CTV plan(p=0.0001), respectively. In (On) ICRU planning, $22.0cm^3$ of CTV volume was not covered by 100% isodose curve in one patient whose residual tumor size is greater than 4cm, while more than 100% dose was irradiated unnecessarily to the normal organ of $62.2{\pm}4.8cm^3$ other than the tumor in the remaining 10 patients with a residual tumor less than 4cm in size. Bladder dose recommended by ICRU 38 was $90.1{\pm}21.3%$ and $68.7{\pm}26.6%$ in ICRU plan and in CTV plan respectively(p=0.001) while rectal dose recommended by ICRU 38 was $86.4{\pm}18.3%$ and $76.9{\pm}15.6%$ in ICRU plan and in CTV plan, respectively(p=0.08). Bladder and rectum maximum dose was $137.2{\pm}50.1%,\;101.1{\pm}41.8%$ in ICRU plan and $107.6{\pm}47.9%,\;86.9{\pm}30.8%$ in CTV plan, respectively. Therefore, the radiation dose to normal organ was lower in CTV plan than in ICRU plan. But the normal tissue dose was remarkably higher than a recommended dose in CTV plan in one patient whose residual tumor size was greater than 4cm. The volume of rectum receiving more than 80% isodose (V80rec) was $1.8{\pm}2.4cm^3$ in ICRU plan and $0.7{\pm}1.0cm^3$ in CTV plan(p=0.02). The volume of bladder receiving more than 80% isodose(V80bla) was $12.2{\pm}8.9cm^3$ in ICRU plan and $3.5{\pm}4.1cm^3$ in CTV plan(p=0.005). According to these parameters, CTV plan could also save more normal tissue compared to ICRU38 plan. Conclusion : An unnecessary excessive radiation dose is irradiated to normal tissues within 100% isodose area in the traditional ICRU plan in case of a small size of cervix cancer, but if we use CTV plan based on CT image, the normal tissue dose could be reduced remarkably without a compromise of tumor dose. However, in a large tumor case, we need more research on an effective 3D-planing to reduce the normal tissue dose.