• Title/Summary/Keyword: handling

Search Result 5,321, Processing Time 0.037 seconds

Development of Quantification Methods for the Myocardial Blood Flow Using Ensemble Independent Component Analysis for Dynamic $H_2^{15}O$ PET (동적 $H_2^{15}O$ PET에서 앙상블 독립성분분석법을 이용한 심근 혈류 정량화 방법 개발)

  • Lee, Byeong-Il;Lee, Jae-Sung;Lee, Dong-Soo;Kang, Won-Jun;Lee, Jong-Jin;Kim, Soo-Jin;Choi, Seung-Jin;Chung, June-Key;Lee, Myung-Chul
    • The Korean Journal of Nuclear Medicine
    • /
    • v.38 no.6
    • /
    • pp.486-491
    • /
    • 2004
  • Purpose: factor analysis and independent component analysis (ICA) has been used for handling dynamic image sequences. Theoretical advantages of a newly suggested ICA method, ensemble ICA, leaded us to consider applying this method to the analysis of dynamic myocardial $H_2^{15}O$ PET data. In this study, we quantified patients' blood flow using the ensemble ICA method. Materials and Methods: Twenty subjects underwent $H_2^{15}O$ PET scans using ECAT EXACT 47 scanner and myocardial perfusion SPECT using Vertex scanner. After transmission scanning, dynamic emission scans were initiated simultaneously with the injection of $555{\sim}740$ MBq $H_2^{15}O$. Hidden independent components can be extracted from the observed mixed data (PET image) by means of ICA algorithms. Ensemble learning is a variational Bayesian method that provides an analytical approximation to the parameter posterior using a tractable distribution. Variational approximation forms a lower bound on the ensemble likelihood and the maximization of the lower bound is achieved through minimizing the Kullback-Leibler divergence between the true posterior and the variational posterior. In this study, posterior pdf was approximated by a rectified Gaussian distribution to incorporate non-negativity constraint, which is suitable to dynamic images in nuclear medicine. Blood flow was measured in 9 regions - apex, four areas in mid wall, and four areas in base wall. Myocardial perfusion SPECT score and angiography results were compared with the regional blood flow. Results: Major cardiac components were separated successfully by the ensemble ICA method and blood flow could be estimated in 15 among 20 patients. Mean myocardial blood flow was $1.2{\pm}0.40$ ml/min/g in rest, $1.85{\pm}1.12$ ml/min/g in stress state. Blood flow values obtained by an operator in two different occasion were highly correlated (r=0.99). In myocardium component image, the image contrast between left ventricle and myocardium was 1:2.7 in average. Perfusion reserve was significantly different between the regions with and without stenosis detected by the coronary angiography (P<0.01). In 66 segment with stenosis confirmed by angiography, the segments with reversible perfusion decrease in perfusion SPECT showed lower perfusion reserve values in $H_2^{15}O$ PET. Conclusions: Myocardial blood flow could be estimated using an ICA method with ensemble learning. We suggest that the ensemble ICA incorporating non-negative constraint is a feasible method to handle dynamic image sequence obtained by the nuclear medicine techniques.

Export Control System based on Case Based Reasoning: Design and Evaluation (사례 기반 지능형 수출통제 시스템 : 설계와 평가)

  • Hong, Woneui;Kim, Uihyun;Cho, Sinhee;Kim, Sansung;Yi, Mun Yong;Shin, Donghoon
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.3
    • /
    • pp.109-131
    • /
    • 2014
  • As the demand of nuclear power plant equipment is continuously growing worldwide, the importance of handling nuclear strategic materials is also increasing. While the number of cases submitted for the exports of nuclear-power commodity and technology is dramatically increasing, preadjudication (or prescreening to be simple) of strategic materials has been done so far by experts of a long-time experience and extensive field knowledge. However, there is severe shortage of experts in this domain, not to mention that it takes a long time to develop an expert. Because human experts must manually evaluate all the documents submitted for export permission, the current practice of nuclear material export is neither time-efficient nor cost-effective. Toward alleviating the problem of relying on costly human experts only, our research proposes a new system designed to help field experts make their decisions more effectively and efficiently. The proposed system is built upon case-based reasoning, which in essence extracts key features from the existing cases, compares the features with the features of a new case, and derives a solution for the new case by referencing similar cases and their solutions. Our research proposes a framework of case-based reasoning system, designs a case-based reasoning system for the control of nuclear material exports, and evaluates the performance of alternative keyword extraction methods (full automatic, full manual, and semi-automatic). A keyword extraction method is an essential component of the case-based reasoning system as it is used to extract key features of the cases. The full automatic method was conducted using TF-IDF, which is a widely used de facto standard method for representative keyword extraction in text mining. TF (Term Frequency) is based on the frequency count of the term within a document, showing how important the term is within a document while IDF (Inverted Document Frequency) is based on the infrequency of the term within a document set, showing how uniquely the term represents the document. The results show that the semi-automatic approach, which is based on the collaboration of machine and human, is the most effective solution regardless of whether the human is a field expert or a student who majors in nuclear engineering. Moreover, we propose a new approach of computing nuclear document similarity along with a new framework of document analysis. The proposed algorithm of nuclear document similarity considers both document-to-document similarity (${\alpha}$) and document-to-nuclear system similarity (${\beta}$), in order to derive the final score (${\gamma}$) for the decision of whether the presented case is of strategic material or not. The final score (${\gamma}$) represents a document similarity between the past cases and the new case. The score is induced by not only exploiting conventional TF-IDF, but utilizing a nuclear system similarity score, which takes the context of nuclear system domain into account. Finally, the system retrieves top-3 documents stored in the case base that are considered as the most similar cases with regard to the new case, and provides them with the degree of credibility. With this final score and the credibility score, it becomes easier for a user to see which documents in the case base are more worthy of looking up so that the user can make a proper decision with relatively lower cost. The evaluation of the system has been conducted by developing a prototype and testing with field data. The system workflows and outcomes have been verified by the field experts. This research is expected to contribute the growth of knowledge service industry by proposing a new system that can effectively reduce the burden of relying on costly human experts for the export control of nuclear materials and that can be considered as a meaningful example of knowledge service application.

Deep Learning-based Professional Image Interpretation Using Expertise Transplant (전문성 이식을 통한 딥러닝 기반 전문 이미지 해석 방법론)

  • Kim, Taejin;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.2
    • /
    • pp.79-104
    • /
    • 2020
  • Recently, as deep learning has attracted attention, the use of deep learning is being considered as a method for solving problems in various fields. In particular, deep learning is known to have excellent performance when applied to applying unstructured data such as text, sound and images, and many studies have proven its effectiveness. Owing to the remarkable development of text and image deep learning technology, interests in image captioning technology and its application is rapidly increasing. Image captioning is a technique that automatically generates relevant captions for a given image by handling both image comprehension and text generation simultaneously. In spite of the high entry barrier of image captioning that analysts should be able to process both image and text data, image captioning has established itself as one of the key fields in the A.I. research owing to its various applicability. In addition, many researches have been conducted to improve the performance of image captioning in various aspects. Recent researches attempt to create advanced captions that can not only describe an image accurately, but also convey the information contained in the image more sophisticatedly. Despite many recent efforts to improve the performance of image captioning, it is difficult to find any researches to interpret images from the perspective of domain experts in each field not from the perspective of the general public. Even for the same image, the part of interests may differ according to the professional field of the person who has encountered the image. Moreover, the way of interpreting and expressing the image also differs according to the level of expertise. The public tends to recognize the image from a holistic and general perspective, that is, from the perspective of identifying the image's constituent objects and their relationships. On the contrary, the domain experts tend to recognize the image by focusing on some specific elements necessary to interpret the given image based on their expertise. It implies that meaningful parts of an image are mutually different depending on viewers' perspective even for the same image. So, image captioning needs to implement this phenomenon. Therefore, in this study, we propose a method to generate captions specialized in each domain for the image by utilizing the expertise of experts in the corresponding domain. Specifically, after performing pre-training on a large amount of general data, the expertise in the field is transplanted through transfer-learning with a small amount of expertise data. However, simple adaption of transfer learning using expertise data may invoke another type of problems. Simultaneous learning with captions of various characteristics may invoke so-called 'inter-observation interference' problem, which make it difficult to perform pure learning of each characteristic point of view. For learning with vast amount of data, most of this interference is self-purified and has little impact on learning results. On the contrary, in the case of fine-tuning where learning is performed on a small amount of data, the impact of such interference on learning can be relatively large. To solve this problem, therefore, we propose a novel 'Character-Independent Transfer-learning' that performs transfer learning independently for each character. In order to confirm the feasibility of the proposed methodology, we performed experiments utilizing the results of pre-training on MSCOCO dataset which is comprised of 120,000 images and about 600,000 general captions. Additionally, according to the advice of an art therapist, about 300 pairs of 'image / expertise captions' were created, and the data was used for the experiments of expertise transplantation. As a result of the experiment, it was confirmed that the caption generated according to the proposed methodology generates captions from the perspective of implanted expertise whereas the caption generated through learning on general data contains a number of contents irrelevant to expertise interpretation. In this paper, we propose a novel approach of specialized image interpretation. To achieve this goal, we present a method to use transfer learning and generate captions specialized in the specific domain. In the future, by applying the proposed methodology to expertise transplant in various fields, we expected that many researches will be actively conducted to solve the problem of lack of expertise data and to improve performance of image captioning.

Comparative Uptake of Tc-99m Sestamibi and Tc-99m Tetrofosmin in Cancer Cells and Tissue Expressing P-Glycoprotein or Multidrug Resistance Associated Protein (P-Glycoprotein과 Multidrug Resistance Associated Protein을 발현하는 암세포와 종양에서 Tc-99m Sestamibi와 Tc-99m Tetrofosmin의 섭취율 비교)

  • Cho, Jung-Ah;Lee, Jae-Tae;Yoo, Jung-Ah;Seo, Ji-Hyoung;Bae, Jin-Ho;Jeong, Shin-Young;Ahn, Byeong-Cheol;Sohn, Sang-Gyun;Ha, Jeoung-Hee;Lee, Kyu-Bo
    • The Korean Journal of Nuclear Medicine
    • /
    • v.39 no.1
    • /
    • pp.34-43
    • /
    • 2005
  • Purpose: $^{99m}Tc$-sestamibi(MIBI) and $^{99m}Tc$-tetrofosmin have been used as substrates for P-glycoprotein (Pgp) and multidrug resistance associated protein (MRP), which are closely associated with multidrug resistance of the tumors. To understand different handling of radiotracers in cancer cell lines expressing Pgp and MRP, we compared cellular uptakes of $^{99m}Tc$-MIBI and $^{99m}Tc$-tetrofosmin. The effects of cyclosporin A (CsA), well-known multidrug resistant reversing agent, on the uptake of both tracers were also compared. Materials and Methods: HCT15/CL02 human colorectal cancer cells for Pgp expressing cells, and human non-small cell lung cancer A549 cells for MRP expressing cells, were used for in vitro and in vivo studies. RT-PCR, western blot analysis and immunohistochemistry were used for detection of Pgp and MRP. MDR-reversal effect with CsA was evaluated at different drug concentrations after incubation with MIBI or tetrofosmin. Radioactivities of supernatant and pellet were measured with gamma well counter. Tumoral uptake of the tracers were measured from tumor bearing nude mice treated with or without CsA. Results: RT-PCR, western blot analysis of the cells and irnrnunochemical staining revealed selective expression of Pgp and MRP for HCY15/CL02 and A549 cells, respectively. There were no significant difference in cellular uptakes of both tracers in HCT15/CL02 cells, but MIBI uptake was slightly higher than that of tetrofosmin in A549 cells. Co-incubation with CsA resulted in a increase in cellular uptakes of MIBI and tetrofosmin. Uptake of MIBI or tetrofosmin in HCT15/CL02 cells was increased by 10- and 2.4-fold, and by 7.5 and 6.3-fold in A549 cells, respectively. Percentage increase of MIBI was higher than that of tetrofosmin with CsA for both cells (p<0.05). In vivo biodistribution study showed that MIBI (114% at 10 min, 257% at 60 min, 396% at 240 min) and tetrofosmin uptake (110% at 10 min, 205% at 60 min, 410% at 240 min) were progressively increased by the time, up to 240 min with CsA. But increases in tumoral uptake were not significantly different between MIBI and tetrofosmin for both tumors. Conclusion: MIBI seems to be a better tracer than tetrofosmin for evaluating MDR reversal effect of the modulators in vitro, but these differences were not evident in vivo tumoral uptake. Both MIBI and tetrofosmin seem to be suitable tracers for imaging Pgp- and MRP-mediated drug resistance in tumors.

Subject-Balanced Intelligent Text Summarization Scheme (주제 균형 지능형 텍스트 요약 기법)

  • Yun, Yeoil;Ko, Eunjung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.141-166
    • /
    • 2019
  • Recently, channels like social media and SNS create enormous amount of data. In all kinds of data, portions of unstructured data which represented as text data has increased geometrically. But there are some difficulties to check all text data, so it is important to access those data rapidly and grasp key points of text. Due to needs of efficient understanding, many studies about text summarization for handling and using tremendous amounts of text data have been proposed. Especially, a lot of summarization methods using machine learning and artificial intelligence algorithms have been proposed lately to generate summary objectively and effectively which called "automatic summarization". However almost text summarization methods proposed up to date construct summary focused on frequency of contents in original documents. Those summaries have a limitation for contain small-weight subjects that mentioned less in original text. If summaries include contents with only major subject, bias occurs and it causes loss of information so that it is hard to ascertain every subject documents have. To avoid those bias, it is possible to summarize in point of balance between topics document have so all subject in document can be ascertained, but still unbalance of distribution between those subjects remains. To retain balance of subjects in summary, it is necessary to consider proportion of every subject documents originally have and also allocate the portion of subjects equally so that even sentences of minor subjects can be included in summary sufficiently. In this study, we propose "subject-balanced" text summarization method that procure balance between all subjects and minimize omission of low-frequency subjects. For subject-balanced summary, we use two concept of summary evaluation metrics "completeness" and "succinctness". Completeness is the feature that summary should include contents of original documents fully and succinctness means summary has minimum duplication with contents in itself. Proposed method has 3-phases for summarization. First phase is constructing subject term dictionaries. Topic modeling is used for calculating topic-term weight which indicates degrees that each terms are related to each topic. From derived weight, it is possible to figure out highly related terms for every topic and subjects of documents can be found from various topic composed similar meaning terms. And then, few terms are selected which represent subject well. In this method, it is called "seed terms". However, those terms are too small to explain each subject enough, so sufficient similar terms with seed terms are needed for well-constructed subject dictionary. Word2Vec is used for word expansion, finds similar terms with seed terms. Word vectors are created after Word2Vec modeling, and from those vectors, similarity between all terms can be derived by using cosine-similarity. Higher cosine similarity between two terms calculated, higher relationship between two terms defined. So terms that have high similarity values with seed terms for each subjects are selected and filtering those expanded terms subject dictionary is finally constructed. Next phase is allocating subjects to every sentences which original documents have. To grasp contents of all sentences first, frequency analysis is conducted with specific terms that subject dictionaries compose. TF-IDF weight of each subjects are calculated after frequency analysis, and it is possible to figure out how much sentences are explaining about each subjects. However, TF-IDF weight has limitation that the weight can be increased infinitely, so by normalizing TF-IDF weights for every subject sentences have, all values are changed to 0 to 1 values. Then allocating subject for every sentences with maximum TF-IDF weight between all subjects, sentence group are constructed for each subjects finally. Last phase is summary generation parts. Sen2Vec is used to figure out similarity between subject-sentences, and similarity matrix can be formed. By repetitive sentences selecting, it is possible to generate summary that include contents of original documents fully and minimize duplication in summary itself. For evaluation of proposed method, 50,000 reviews of TripAdvisor are used for constructing subject dictionaries and 23,087 reviews are used for generating summary. Also comparison between proposed method summary and frequency-based summary is performed and as a result, it is verified that summary from proposed method can retain balance of all subject more which documents originally have.

A Study on the Tree Surgery Problem and Protection Measures in Monumental Old Trees (천연기념물 노거수 외과수술 문제점 및 보존 관리방안에 관한 연구)

  • Jung, Jong Soo
    • Korean Journal of Heritage: History & Science
    • /
    • v.42 no.1
    • /
    • pp.122-142
    • /
    • 2009
  • This study explored all domestic and international theories for maintenance and health enhancement of an old and big tree, and carried out the anatomical survey of the operation part of the tree toward he current status of domestic surgery and the perception survey of an expert group, and drew out following conclusion through the process of suggesting its reform plan. First, as a result of analyzing the correlation of the 67 subject trees with their ages, growth status. surroundings, it revealed that they were closely related to positional characteristic, damage size, whereas were little related to materials by fillers. Second, the size of the affected part was the most frequent at the bough sheared part under $0.09m^2$, and the hollow size by position(part) was the biggest at 'root + stem' starting from the behind of the main root and stem As a result of analyzing the correlation, the same result was elicited at the group with low correlation. Third, the problem was serious in charging the fillers (especially urethane) in the big hollow or exposed root produced at the behind of the root and stem part, or surface-processing it. The benefit by charging the hollow part was analyzed as not so much. Fourth, the surface-processing of fillers currently used (artificial bark) is mainly 'epoxy+woven fabric+cork', but it is not flexible, so it has brought forth problems of frequent cracks and cracked surface at the joint part with the treetextured part. Fifth, the correlation with the external status of the operated part was very high with the closeness, surface condition, formation of adhesive tissue and internal survey result. Sixth, the most influential thing on flushing by the wrong management of an old and big tree was banking, and a wrong pruning was the source of the ground part damage. In pruning a small bough can easily recover itself from its damage as its formation of adhesive tissue when it is cut by a standard method. Seventh, the parameters affecting the times of related business handling of an old and big tree are 'the need of the conscious reform of the manager and related business'. Eighth, a reform plan in an institutional aspect can include the arrangement of the law and organization of the old and big tree management and preservation at an institutional aspect. This study for preparing a reform plan through the status survey of the designated old and big tree, has a limit inducing a reform plan based on the status survey through individual research, and a weak point suggesting grounds by any statistical data. This can be complemented by subsequent studies.

A Study on the Daesoon Cosmology of the Correlative Relation between Mugeuk and Taegeuk (무극과 태극 상관연동의 대순우주론 연구)

  • Kim, Yong-hwan
    • Journal of the Daesoon Academy of Sciences
    • /
    • v.33
    • /
    • pp.31-62
    • /
    • 2019
  • The purpose of this article is to study on the Daesoon Cosmology of the Correlative Relation between Mugeuk and Taegeuk. Daesoon cosmology is a cosmology based on the juxtaposition between the Gucheon Sangje and the world. In this article, I would like to say that this theory in Daesoon Thought was developed in three stages: the phase of the Mugeuk Transcendence of Gucheon Sangje, the stage of the Taegeuk Immanence, and the phase of the Grand Opening of the Later World between Mugeuk and Taegeuk as a correlative gentle reign. First of all, the phase of the Mugeuk Transcendence of Gucheon Sangje has been revealed as a yin-yang relationship. The stage of the Taegeuk Immanence represents the togetherness of harmony and co-prosperity between yin and yang, and the phase of the Grand Opening of the Later World between Mukeuk and Taegeuk refers to the unshakable accomplishment of its character and energy. It will be said that this is due to the practical mechanism in the correct balance of yin-yang making a four stage cycle of birth, growth, harvest, and storage. In addition, the Daesoon stage of the settlement of yin and yang is revealed as a change in the growth of all things and the formation of the inner circle. The mental growth reveals the characteristics of everything in the world, each trying to shine at the height of their own respective life as they grow up energetically. The dominant culture of cerebral communion renders a soft and elegant mood and combines yin and yang to elevate the heavenly and earthly period through transcendental change into sympathetic understanding. The stage of the Grand Opening of the Later World between Mugeuk and Taegeuk is one of the earliest days of the lunar month and also the inner circle of Taegeuk. It is in line with Ken Wilbur's integrated model as a step to the true degrees to develop into a world with brightened degrees. It is a beautiful and peaceful scene where celestial maidens play music, the firewood burns, and the scholars command thunder and lightning playfully. Human beings achieve a state of happiness as a free beings who lives as gods upon the earth. This is the world of theGrand Opening of the Later World between Mugeuk and Taegeuk. Daesoon Thought was succeeded by Dojeon in 1958, when Dojeon emerged as the successor in the lineage of religious orthodoxy and was assigned the task of handling Dao in its entirety. In addition, Daesoon is a circle and represents freedom and commonly shared happiness among the populous. Cosmology in the Daesoon Thought will enable us to understand deep dimensions and the identity of members as individuals within an inner circle of correlation between transcendence and immanence. This present study tries to analyze the public effects philologically and also the mutual correlation by utilizing the truthfulness of literature and rational interpretation. The outlook for the future in Daesoon Thought also leads to the one-way communication of Daesoon as a circle.

A Study on the Dimensions, Surface Area and Volume of Grains (곡립(穀粒)의 치수, 표면적(表面積) 및 체적(體積)에 관(關)한 연구(硏究))

  • Park, Jong Min;Kim, Man Soo
    • Korean Journal of Agricultural Science
    • /
    • v.16 no.1
    • /
    • pp.84-101
    • /
    • 1989
  • An accurate measurement of size, surface area and volume of agricultural products is essential in many engineering operations such as handling and sorting, and in heat transfer studies on heating and cooling processes. Little information is available on these properties due to their irregular shape, and moreover very little information on the rough rice, soybean, barley, and wheat has been published. Physical dimensions of grain, such as length, width, thickness, surface area, and volume vary according to the variety, environmental conditions, temperature, and moisture content. Especially, recent research has emphasized on the variation of these properties with the important factors such as moisture content. The objectives of this study were to determine physical dimensions such as length, width and thickness, surface area and volume of the rough rice, soybean, barley, and wheat as a function of moisture content, to investigate the effect of moisture content on the properties, and to develop exponential equations to predict the surface area and the volume of the grains as a function of physical dimensions. The varieties of the rough rice used in this study were Akibare, Milyang 15, Seomjin, Samkang, Chilseong, and Yongmun, as a soybean sample Jangyeobkong and Hwangkeumkong, as a barley sample Olbori and Salbori, and as a wheat sample Eunpa and Guru were selected, respectively. The physical properties of the grain samples were determined at four levels of moisture content and ten or fifteen replications were run at each moisture content level and each variety. The results of this study are summarized as follows; 1. In comparison of the surface area and the volume of the 0.0375m diameter-sphere measured in this study with the calculated values by the formula the percent error between them showed least values of 0.65% and 0.77% at the rotational degree interval of 15 degree respectively. 2. The statistical test(t-test) results of the physical properties between the types of rough rice, and between the varieties of soybean and wheat indicated that there were significant difference at the 5% level between them. 3. The physical dimensions varied linearly with the moisture content, and the ratios of length to thickness (L/T) and of width to thickness (W/T) in rough rice decreased with increase of moisture content, while increased in soybean, but uniform tendency of the ratios in barley and wheat was not shown. In all of the sample grains except Olbori, sphericity decreased with increase of moisture content. 4. Over the experimental moisture levels, the surface area and the volume were in the ranges of about $45{\sim}51{\times}10^{-6}m^2$, $25{\sim}30{\times}10^{-9}m^3$ for Japonica-type rough rice, about $42{\sim}47{\times}10^{-6}m^2$, $21{\sim}26{\times}10^{-9}m^3$ for Indica${\times}$Japonica type rough rice, about $188{\sim}200{\times}10^{-6}m^2$, $277{\sim}300{\times}10^{-9}m^3$ for Jangyeobkong, about $180{\sim}201{\times}10^{-6}m^2$, $190{\sim}253{\times}10^{-9}m^3$ for Hwangkeumkong, about $60{\sim}69{\times}10^{-6}m^2$, $36{\sim}45{\times}10^{-9}m^3$ for Covered barley, about $47{\sim}60{\times}10^{-6}m^2$, $22{\sim}28{\times}10^{-9}m^3$ for Naked barley, about $51{\sim}20{\times}10^{-6}m^2$, $23{\sim}31{\times}10^{-9}m^3$ for Eunpamill, and about $57{\sim}69{\times}10^{-6}m^2$, $27{\sim}34{\times}10^{-9}m^3$ for Gurumill, respectively. 5. The increasing rate of surface area and volume with increase of moisture content was higher in soybean than other sample grains, and that of Japonica-type was slightly higher than Indica${\times}$Japonica type in rough rice. 6. The regression equations of physical dimensions, surface area and volume were developed as a function of moisture content, the exponential equations of surface area and volume were also developed as a function of physical dimensions, and the regression equations of surface area were also developed as a function of volume in all grain samples.

  • PDF

Design and Implementation of MongoDB-based Unstructured Log Processing System over Cloud Computing Environment (클라우드 환경에서 MongoDB 기반의 비정형 로그 처리 시스템 설계 및 구현)

  • Kim, Myoungjin;Han, Seungho;Cui, Yun;Lee, Hanku
    • Journal of Internet Computing and Services
    • /
    • v.14 no.6
    • /
    • pp.71-84
    • /
    • 2013
  • Log data, which record the multitude of information created when operating computer systems, are utilized in many processes, from carrying out computer system inspection and process optimization to providing customized user optimization. In this paper, we propose a MongoDB-based unstructured log processing system in a cloud environment for processing the massive amount of log data of banks. Most of the log data generated during banking operations come from handling a client's business. Therefore, in order to gather, store, categorize, and analyze the log data generated while processing the client's business, a separate log data processing system needs to be established. However, the realization of flexible storage expansion functions for processing a massive amount of unstructured log data and executing a considerable number of functions to categorize and analyze the stored unstructured log data is difficult in existing computer environments. Thus, in this study, we use cloud computing technology to realize a cloud-based log data processing system for processing unstructured log data that are difficult to process using the existing computing infrastructure's analysis tools and management system. The proposed system uses the IaaS (Infrastructure as a Service) cloud environment to provide a flexible expansion of computing resources and includes the ability to flexibly expand resources such as storage space and memory under conditions such as extended storage or rapid increase in log data. Moreover, to overcome the processing limits of the existing analysis tool when a real-time analysis of the aggregated unstructured log data is required, the proposed system includes a Hadoop-based analysis module for quick and reliable parallel-distributed processing of the massive amount of log data. Furthermore, because the HDFS (Hadoop Distributed File System) stores data by generating copies of the block units of the aggregated log data, the proposed system offers automatic restore functions for the system to continually operate after it recovers from a malfunction. Finally, by establishing a distributed database using the NoSQL-based Mongo DB, the proposed system provides methods of effectively processing unstructured log data. Relational databases such as the MySQL databases have complex schemas that are inappropriate for processing unstructured log data. Further, strict schemas like those of relational databases cannot expand nodes in the case wherein the stored data are distributed to various nodes when the amount of data rapidly increases. NoSQL does not provide the complex computations that relational databases may provide but can easily expand the database through node dispersion when the amount of data increases rapidly; it is a non-relational database with an appropriate structure for processing unstructured data. The data models of the NoSQL are usually classified as Key-Value, column-oriented, and document-oriented types. Of these, the representative document-oriented data model, MongoDB, which has a free schema structure, is used in the proposed system. MongoDB is introduced to the proposed system because it makes it easy to process unstructured log data through a flexible schema structure, facilitates flexible node expansion when the amount of data is rapidly increasing, and provides an Auto-Sharding function that automatically expands storage. The proposed system is composed of a log collector module, a log graph generator module, a MongoDB module, a Hadoop-based analysis module, and a MySQL module. When the log data generated over the entire client business process of each bank are sent to the cloud server, the log collector module collects and classifies data according to the type of log data and distributes it to the MongoDB module and the MySQL module. The log graph generator module generates the results of the log analysis of the MongoDB module, Hadoop-based analysis module, and the MySQL module per analysis time and type of the aggregated log data, and provides them to the user through a web interface. Log data that require a real-time log data analysis are stored in the MySQL module and provided real-time by the log graph generator module. The aggregated log data per unit time are stored in the MongoDB module and plotted in a graph according to the user's various analysis conditions. The aggregated log data in the MongoDB module are parallel-distributed and processed by the Hadoop-based analysis module. A comparative evaluation is carried out against a log data processing system that uses only MySQL for inserting log data and estimating query performance; this evaluation proves the proposed system's superiority. Moreover, an optimal chunk size is confirmed through the log data insert performance evaluation of MongoDB for various chunk sizes.

A Comparative Study of Domestic and International regulation on Mixed-fleet Flying of Flight crew (운항승무원의 항공기 2개 형식 운항관련 국내외 기준 비교 연구)

  • Lee, Koo-Hee
    • The Korean Journal of Air & Space Law and Policy
    • /
    • v.30 no.2
    • /
    • pp.403-425
    • /
    • 2015
  • The Chicago Convention and Annexes have become the basis of aviation safety regulations for every contracting state. Generally, the State's aviation safety regulations refer to the Standards and Recommended Practices(SARPs) provided in the Annexes of the Chicago Convention. In order to properly reflect international aviation safety regulations, constant studies of the aviation fields are of paramount importance. This Paper is intended to identify the main differences between korean and foreign regulation and suggest a few amendment proposals on Mixed-fleet Flying(at or more two aircraft type operation) of flight crew. Comparing with these regulations, the korean regulations and implementations have some insufficiency points. I suggest some amendment proposals of korean regulations concerning Mixed-fleet Flying that flight crew operate aircraft of different types. Basically an operator shall not assign a pilot-in-command or a co-pilot to operate at the flight controls of a type of airplane during take-off and landing unless that pilot has operated the flight controls during at least three take-offs and landings within the preceding 90 days on the same type of airplane or in a flight simulator. Also, flight crew members are familiarized with the significant differences in equipment and/or procedures between concurrently operated types. An operator shall ensure that piloting technique and the ability to execute emergency procedures is checked in such a way as to demonstrate the pilot's competence on each type or variant of a type of airplane. Proficiency check shall be performed periodically. When an operator schedules flight crew on different types of airplanes with similar characteristics in terms of operating procedures, systems and handling, the State shall decide the requirements for each type of airplane can be combined. In conclusion, it is necessary for flight crew members to remain concurrently qualified to operate multiple types. The operator shall have a program to include, as a minimum, required differences training between types and qualification to maintain currency on each type. If the Operator utilizes flight crew members to concurrently operate aircraft of different types, the operator shall have qualification processes approved or accepted by the State. If applicable, the qualification curriculum as defined in the operator's Advanced Qualification Program could be applied. Flight crew members are familiarized with the significant differences in equipment and/or procedures between concurrently operated types. The difference among different types of airpcrafts decrease and standards for these airpcrafts can be applied increasingly because function and performance have been improved by aircraft manufacture company in accordance to basic aircraft system in terms of developing new aircrafts for flight standard procedure and safety of flight. Also, it becomes more necessary for flight crews to control multi aircraft types due to various aviation business and activation of leisure business. Nevertheless, in terms of flight crew training and qualification program, there are no regulations in Korea to be applied to new aircraft types differently in accordance with different levels. In addition, it has no choice different programs based on different levels because there are not provisions to restrict or limit and specific standards to operate at or more than two aircraft types for flight safety. Therefore the aviation authority introduce Flight Standardization and/or Operational Evaluation Board in order to analysis differences among aircraft types. In addition to that, the aviation authority should also improve standard flight evaluation and qualification system among different aircraft types for flight crews to apply reasonable training and qualification efficiently. For all the issue mentioned above, I have studied the ICAO SARPs and some state's regulation concerning operating aircraft of different types(Mixed-fleet flying), and suggested some proposals on the different aircraft type operation as an example of comprehensive problem solving. I hope that this paper is 1) to help understanding about the international issue, 2) to help the improvement of korean aviation regulations, 3) to help compliance with international standards and to contribute to the promotion of aviation safety, in addition.