• Title/Summary/Keyword: 정보검색 시스템

Search Result 5,079, Processing Time 0.039 seconds

Influence analysis of Internet buzz to corporate performance : Individual stock price prediction using sentiment analysis of online news (온라인 언급이 기업 성과에 미치는 영향 분석 : 뉴스 감성분석을 통한 기업별 주가 예측)

  • Jeong, Ji Seon;Kim, Dong Sung;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.37-51
    • /
    • 2015
  • Due to the development of internet technology and the rapid increase of internet data, various studies are actively conducted on how to use and analyze internet data for various purposes. In particular, in recent years, a number of studies have been performed on the applications of text mining techniques in order to overcome the limitations of the current application of structured data. Especially, there are various studies on sentimental analysis to score opinions based on the distribution of polarity such as positivity or negativity of vocabularies or sentences of the texts in documents. As a part of such studies, this study tries to predict ups and downs of stock prices of companies by performing sentimental analysis on news contexts of the particular companies in the Internet. A variety of news on companies is produced online by different economic agents, and it is diffused quickly and accessed easily in the Internet. So, based on inefficient market hypothesis, we can expect that news information of an individual company can be used to predict the fluctuations of stock prices of the company if we apply proper data analysis techniques. However, as the areas of corporate management activity are different, an analysis considering characteristics of each company is required in the analysis of text data based on machine-learning. In addition, since the news including positive or negative information on certain companies have various impacts on other companies or industry fields, an analysis for the prediction of the stock price of each company is necessary. Therefore, this study attempted to predict changes in the stock prices of the individual companies that applied a sentimental analysis of the online news data. Accordingly, this study chose top company in KOSPI 200 as the subjects of the analysis, and collected and analyzed online news data by each company produced for two years on a representative domestic search portal service, Naver. In addition, considering the differences in the meanings of vocabularies for each of the certain economic subjects, it aims to improve performance by building up a lexicon for each individual company and applying that to an analysis. As a result of the analysis, the accuracy of the prediction by each company are different, and the prediction accurate rate turned out to be 56% on average. Comparing the accuracy of the prediction of stock prices on industry sectors, 'energy/chemical', 'consumer goods for living' and 'consumer discretionary' showed a relatively higher accuracy of the prediction of stock prices than other industries, while it was found that the sectors such as 'information technology' and 'shipbuilding/transportation' industry had lower accuracy of prediction. The number of the representative companies in each industry collected was five each, so it is somewhat difficult to generalize, but it could be confirmed that there was a difference in the accuracy of the prediction of stock prices depending on industry sectors. In addition, at the individual company level, the companies such as 'Kangwon Land', 'KT & G' and 'SK Innovation' showed a relatively higher prediction accuracy as compared to other companies, while it showed that the companies such as 'Young Poong', 'LG', 'Samsung Life Insurance', and 'Doosan' had a low prediction accuracy of less than 50%. In this paper, we performed an analysis of the share price performance relative to the prediction of individual companies through the vocabulary of pre-built company to take advantage of the online news information. In this paper, we aim to improve performance of the stock prices prediction, applying online news information, through the stock price prediction of individual companies. Based on this, in the future, it will be possible to find ways to increase the stock price prediction accuracy by complementing the problem of unnecessary words that are added to the sentiment dictionary.

A study on the classification of research topics based on COVID-19 academic research using Topic modeling (토픽모델링을 활용한 COVID-19 학술 연구 기반 연구 주제 분류에 관한 연구)

  • Yoo, So-yeon;Lim, Gyoo-gun
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.155-174
    • /
    • 2022
  • From January 2020 to October 2021, more than 500,000 academic studies related to COVID-19 (Coronavirus-2, a fatal respiratory syndrome) have been published. The rapid increase in the number of papers related to COVID-19 is putting time and technical constraints on healthcare professionals and policy makers to quickly find important research. Therefore, in this study, we propose a method of extracting useful information from text data of extensive literature using LDA and Word2vec algorithm. Papers related to keywords to be searched were extracted from papers related to COVID-19, and detailed topics were identified. The data used the CORD-19 data set on Kaggle, a free academic resource prepared by major research groups and the White House to respond to the COVID-19 pandemic, updated weekly. The research methods are divided into two main categories. First, 41,062 articles were collected through data filtering and pre-processing of the abstracts of 47,110 academic papers including full text. For this purpose, the number of publications related to COVID-19 by year was analyzed through exploratory data analysis using a Python program, and the top 10 journals under active research were identified. LDA and Word2vec algorithm were used to derive research topics related to COVID-19, and after analyzing related words, similarity was measured. Second, papers containing 'vaccine' and 'treatment' were extracted from among the topics derived from all papers, and a total of 4,555 papers related to 'vaccine' and 5,971 papers related to 'treatment' were extracted. did For each collected paper, detailed topics were analyzed using LDA and Word2vec algorithms, and a clustering method through PCA dimension reduction was applied to visualize groups of papers with similar themes using the t-SNE algorithm. A noteworthy point from the results of this study is that the topics that were not derived from the topics derived for all papers being researched in relation to COVID-19 (

    ) were the topic modeling results for each research topic (
    ) was found to be derived from For example, as a result of topic modeling for papers related to 'vaccine', a new topic titled Topic 05 'neutralizing antibodies' was extracted. A neutralizing antibody is an antibody that protects cells from infection when a virus enters the body, and is said to play an important role in the production of therapeutic agents and vaccine development. In addition, as a result of extracting topics from papers related to 'treatment', a new topic called Topic 05 'cytokine' was discovered. A cytokine storm is when the immune cells of our body do not defend against attacks, but attack normal cells. Hidden topics that could not be found for the entire thesis were classified according to keywords, and topic modeling was performed to find detailed topics. In this study, we proposed a method of extracting topics from a large amount of literature using the LDA algorithm and extracting similar words using the Skip-gram method that predicts the similar words as the central word among the Word2vec models. The combination of the LDA model and the Word2vec model tried to show better performance by identifying the relationship between the document and the LDA subject and the relationship between the Word2vec document. In addition, as a clustering method through PCA dimension reduction, a method for intuitively classifying documents by using the t-SNE technique to classify documents with similar themes and forming groups into a structured organization of documents was presented. In a situation where the efforts of many researchers to overcome COVID-19 cannot keep up with the rapid publication of academic papers related to COVID-19, it will reduce the precious time and effort of healthcare professionals and policy makers, and rapidly gain new insights. We hope to help you get It is also expected to be used as basic data for researchers to explore new research directions.

  • Analyzing the Issue Life Cycle by Mapping Inter-Period Issues (기간별 이슈 매핑을 통한 이슈 생명주기 분석 방법론)

    • Lim, Myungsu;Kim, Namgyu
      • Journal of Intelligence and Information Systems
      • /
      • v.20 no.4
      • /
      • pp.25-41
      • /
      • 2014
    • Recently, the number of social media users has increased rapidly because of the prevalence of smart devices. As a result, the amount of real-time data has been increasing exponentially, which, in turn, is generating more interest in using such data to create added value. For instance, several attempts are being made to analyze the relevant search keywords that are frequently used on new portal sites and the words that are regularly mentioned on various social media in order to identify social issues. The technique of "topic analysis" is employed in order to identify topics and themes from a large amount of text documents. As one of the most prevalent applications of topic analysis, the technique of issue tracking investigates changes in the social issues that are identified through topic analysis. Currently, traditional issue tracking is conducted by identifying the main topics of documents that cover an entire period at the same time and analyzing the occurrence of each topic by the period of occurrence. However, this traditional issue tracking approach has two limitations. First, when a new period is included, topic analysis must be repeated for all the documents of the entire period, rather than being conducted only on the new documents of the added period. This creates practical limitations in the form of significant time and cost burdens. Therefore, this traditional approach is difficult to apply in most applications that need to perform an analysis on the additional period. Second, the issue is not only generated and terminated constantly, but also one issue can sometimes be distributed into several issues or multiple issues can be integrated into one single issue. In other words, each issue is characterized by a life cycle that consists of the stages of creation, transition (merging and segmentation), and termination. The existing issue tracking methods do not address the connection and effect relationship between these issues. The purpose of this study is to overcome the two limitations of the existing issue tracking method, one being the limitation regarding the analysis method and the other being the limitation involving the lack of consideration of the changeability of the issues. Let us assume that we perform multiple topic analysis for each multiple period. Then it is essential to map issues of different periods in order to trace trend of issues. However, it is not easy to discover connection between issues of different periods because the issues derived for each period mutually contain heterogeneity. In this study, to overcome these limitations without having to analyze the entire period's documents simultaneously, the analysis can be performed independently for each period. In addition, we performed issue mapping to link the identified issues of each period. An integrated approach on each details period was presented, and the issue flow of the entire integrated period was depicted in this study. Thus, as the entire process of the issue life cycle, including the stages of creation, transition (merging and segmentation), and extinction, is identified and examined systematically, the changeability of the issues was analyzed in this study. The proposed methodology is highly efficient in terms of time and cost, as it sufficiently considered the changeability of the issues. Further, the results of this study can be used to adapt the methodology to a practical situation. By applying the proposed methodology to actual Internet news, the potential practical applications of the proposed methodology are analyzed. Consequently, the proposed methodology was able to extend the period of the analysis and it could follow the course of progress of each issue's life cycle. Further, this methodology can facilitate a clearer understanding of complex social phenomena using topic analysis.

    A study of Artificial Intelligence (AI) Speaker's Development Process in Terms of Social Constructivism: Focused on the Products and Periodic Co-revolution Process (인공지능(AI) 스피커에 대한 사회구성 차원의 발달과정 연구: 제품과 시기별 공진화 과정을 중심으로)

    • Cha, Hyeon-ju;Kweon, Sang-hee
      • Journal of Internet Computing and Services
      • /
      • v.22 no.1
      • /
      • pp.109-135
      • /
      • 2021
    • his study classified the development process of artificial intelligence (AI) speakers through analysis of the news text of artificial intelligence (AI) speakers shown in traditional news reports, and identified the characteristics of each product by period. The theoretical background used in the analysis are news frames and topic frames. As analysis methods, topic modeling and semantic network analysis using the LDA method were used. The research method was a content analysis method. From 2014 to 2019, 2710 news related to AI speakers were first collected, and secondly, topic frames were analyzed using Nodexl algorithm. The result of this study is that, first, the trend of topic frames by AI speaker provider type was different according to the characteristics of the four operators (communication service provider, online platform, OS provider, and IT device manufacturer). Specifically, online platform operators (Google, Naver, Amazon, Kakao) appeared as a frame that uses AI speakers as'search or input devices'. On the other hand, telecommunications operators (SKT, KT) showed prominent frames for IPTV, which is the parent company's flagship business, and 'auxiliary device' of the telecommunication business. Furthermore, the frame of "personalization of products and voice service" was remarkable for OS operators (MS, Apple), and the frame for IT device manufacturers (Samsung) was "Internet of Things (IoT) Integrated Intelligence System". The econd, result id that the trend of the topic frame by AI speaker development period (by year) showed a tendency to develop around AI technology in the first phase (2014-2016), and in the second phase (2017-2018), the social relationship between AI technology and users It was related to interaction, and in the third phase (2019), there was a trend of shifting from AI technology-centered to user-centered. As a result of QAP analysis, it was found that news frames by business operator and development period in AI speaker development are socially constituted by determinants of media discourse. The implication of this study was that the evolution of AI speakers was found by the characteristics of the parent company and the process of co-evolution due to interactions between users by business operator and development period. The implications of this study are that the results of this study are important indicators for predicting the future prospects of AI speakers and presenting directions accordingly.

    A Study on the Development Trend of Artificial Intelligence Using Text Mining Technique: Focused on Open Source Software Projects on Github (텍스트 마이닝 기법을 활용한 인공지능 기술개발 동향 분석 연구: 깃허브 상의 오픈 소스 소프트웨어 프로젝트를 대상으로)

    • Chong, JiSeon;Kim, Dongsung;Lee, Hong Joo;Kim, Jong Woo
      • Journal of Intelligence and Information Systems
      • /
      • v.25 no.1
      • /
      • pp.1-19
      • /
      • 2019
    • Artificial intelligence (AI) is one of the main driving forces leading the Fourth Industrial Revolution. The technologies associated with AI have already shown superior abilities that are equal to or better than people in many fields including image and speech recognition. Particularly, many efforts have been actively given to identify the current technology trends and analyze development directions of it, because AI technologies can be utilized in a wide range of fields including medical, financial, manufacturing, service, and education fields. Major platforms that can develop complex AI algorithms for learning, reasoning, and recognition have been open to the public as open source projects. As a result, technologies and services that utilize them have increased rapidly. It has been confirmed as one of the major reasons for the fast development of AI technologies. Additionally, the spread of the technology is greatly in debt to open source software, developed by major global companies, supporting natural language recognition, speech recognition, and image recognition. Therefore, this study aimed to identify the practical trend of AI technology development by analyzing OSS projects associated with AI, which have been developed by the online collaboration of many parties. This study searched and collected a list of major projects related to AI, which were generated from 2000 to July 2018 on Github. This study confirmed the development trends of major technologies in detail by applying text mining technique targeting topic information, which indicates the characteristics of the collected projects and technical fields. The results of the analysis showed that the number of software development projects by year was less than 100 projects per year until 2013. However, it increased to 229 projects in 2014 and 597 projects in 2015. Particularly, the number of open source projects related to AI increased rapidly in 2016 (2,559 OSS projects). It was confirmed that the number of projects initiated in 2017 was 14,213, which is almost four-folds of the number of total projects generated from 2009 to 2016 (3,555 projects). The number of projects initiated from Jan to Jul 2018 was 8,737. The development trend of AI-related technologies was evaluated by dividing the study period into three phases. The appearance frequency of topics indicate the technology trends of AI-related OSS projects. The results showed that the natural language processing technology has continued to be at the top in all years. It implied that OSS had been developed continuously. Until 2015, Python, C ++, and Java, programming languages, were listed as the top ten frequently appeared topics. However, after 2016, programming languages other than Python disappeared from the top ten topics. Instead of them, platforms supporting the development of AI algorithms, such as TensorFlow and Keras, are showing high appearance frequency. Additionally, reinforcement learning algorithms and convolutional neural networks, which have been used in various fields, were frequently appeared topics. The results of topic network analysis showed that the most important topics of degree centrality were similar to those of appearance frequency. The main difference was that visualization and medical imaging topics were found at the top of the list, although they were not in the top of the list from 2009 to 2012. The results indicated that OSS was developed in the medical field in order to utilize the AI technology. Moreover, although the computer vision was in the top 10 of the appearance frequency list from 2013 to 2015, they were not in the top 10 of the degree centrality. The topics at the top of the degree centrality list were similar to those at the top of the appearance frequency list. It was found that the ranks of the composite neural network and reinforcement learning were changed slightly. The trend of technology development was examined using the appearance frequency of topics and degree centrality. The results showed that machine learning revealed the highest frequency and the highest degree centrality in all years. Moreover, it is noteworthy that, although the deep learning topic showed a low frequency and a low degree centrality between 2009 and 2012, their ranks abruptly increased between 2013 and 2015. It was confirmed that in recent years both technologies had high appearance frequency and degree centrality. TensorFlow first appeared during the phase of 2013-2015, and the appearance frequency and degree centrality of it soared between 2016 and 2018 to be at the top of the lists after deep learning, python. Computer vision and reinforcement learning did not show an abrupt increase or decrease, and they had relatively low appearance frequency and degree centrality compared with the above-mentioned topics. Based on these analysis results, it is possible to identify the fields in which AI technologies are actively developed. The results of this study can be used as a baseline dataset for more empirical analysis on future technology trends that can be converged.

    Development of JPEG2000 Viewer for Mobile Image System (이동형 의료영상 장치를 위한 JPEG2000 영상 뷰어 개발)

    • 김새롬;정해조;강원석;이재훈;이상호;신성범;유선국;김희중
      • Progress in Medical Physics
      • /
      • v.14 no.2
      • /
      • pp.124-130
      • /
      • 2003
    • Currently, as a consequence of PACS (Picture Archiving Communication System) implementation many hospitals are replacing conventional film-type interpretations of diagnostic medical images with new digital-format interpretations that can also be saved, and retrieve However, the big limitation in PACS is considered to be the lack of mobility. The purpose of this study is to determine the optimal communication packet size. This was done by considering the terms occurred in the wireless communication. After encoding medical image using JPGE2000 image compression method, This method embodied auto-error correction technique preventing the loss of packets occurred during wireless communication. A PC class server, with capabilities to load, collect data, save images, and connect with other network, was installed. Image data were compressed using JPEG2000 algorithm which supports the capability of high energy density and compression ratio, to communicate through a wireless network. Image data were also transmitted in block units coeded by JPEG2000 to prevent the loss of the packets in a wireless network. When JPGE2000 image data were decoded in a PUA (Personal Digital Assistant), it was instantaneous for a MR (Magnetic Resonance) head image of 256${\times}$256 pixels, while it took approximately 5 seconds to decode a CR (Computed Radiography) chest image of 800${\times}$790 pixels. In the transmission of the image data using a CDMA 1X module (Code-Division Multiple Access 1st Generation), 256 byte/sec was considered a stable transmission rate, but packets were lost in the intervals at the transmission rate of 1Kbyte/sec. However, even with a transmission rate above 1 Kbyte/sec, packets were not lost in wireless LAN. Current PACS are not compatible with wireless networks. because it does not have an interface between wired and wireless. Thus, the mobile JPEG2000 image viewing system was developed in order to complement mobility-a limitation in PACS. Moreover, the weak-connections of the wireless network was enhanced by re-transmitting image data within a limitations The results of this study are expected to play an interface role between the current wired-networks PACS and the mobile devices.

    • PDF

    An Analysis of Research Trends Related to Software Education for Young Children in Korea (유아의 소프트웨어 교육 관련 국내 최근 연구의 경향 분석)

    • Chun, Hui Young;Park, Soyeon;Sung, Jihyun
      • Korean Journal of Child Education & Care
      • /
      • v.19 no.2
      • /
      • pp.177-196
      • /
      • 2019
    • Objective: This study aims to analyze research trends related to software education for young children, focusing on studies published in Korea from 2016 to 2019 March. Methods: A total of 26 research publications on software education for young children, searched from Korea Citation Index and Research Information Sharing Service were identified for the analysis. The trend in these publications was classified and examined respectively by publication dates, types of publications, and the fields of study. To investigate a means of research, the analysis included key topics, types of research methods, and characteristics of the study variables. Results: The results of the analysis show that the number of publications on the topic of software education for young children has increased over the three years, of which most were published as a scholarly journal article. Among the 26 research studies analyzed, 16 (61.5%) are related to the field of early childhood education or child studies. Key topics and target subjects of the most research include the curriculum development of software education for young children or the effectiveness of software education on 4- and 5-year-old children. Most of the analyzed studies are experimental research designs or in the form of literature reviews. The most frequently studied research variable is young children's cognitive characteristics. For the studies that employ educational programs, the use of a physical computing environment is prevalent, and the most frequently used robot as a programming tool is "Albert". The duration of the program implementation varies, ranging from 5 weeks to 48 weeks. In the analyzed research studies, computational thinking is conceptualized as a problem-solving skill that can be improved by software education, and assessed by individual instruments measuring sub-factors of computational thinking. Conclusion/Implications: The present study reveals that, although the number of research publications in software education for young children has increased, the overall sufficiency of the accumulated research data and a variety of research methods are still lacking. An increased interest in software education for young children and more research activities in this area are needed to develop and implement developmentally appropriate software education programs in early childhood settings.

    A Development of Facility Web Program for Small and Medium-Sized PSM Workplaces (중·소규모 공정안전관리 사업장의 웹 전산시스템 개발)

    • Kim, Young Suk;Park, Dal Jae
      • Korean Chemical Engineering Research
      • /
      • v.60 no.3
      • /
      • pp.334-346
      • /
      • 2022
    • There is a lack of knowledge and information on the understanding and application of the Process Safety Management (PSM) system, recognized as a major cause of industrial accidents in small-and medium-sized workplaces. Hence, it is necessary to prepare a protocol to secure the practical and continuous levels of implementation for PSM and eliminate human errors through tracking management. However, insufficient research has been conducted on this. Therefore, this study investigated and analyzed the various violations in the administrative measures, based on the regulations announced by the Ministry of Employment and Labor, in approximately 200 small-and medium-sized PSM workplaces with fewer than 300 employees across in korea. This study intended to contribute to the prevention of major industrial accidents by developing a facility maintenance web program that removed human errors in small-and medium-sized workplaces. The major results are summarized as follows. First, It accessed the web via a QR code on a smart device to check the equipment's specification search function, cause of failure, and photos for the convenience of accessing the program, which made it possible to make requests for the it inspection and maintenance in real time. Second, it linked the identification of the targets to be changed, risk assessment, worker training, and pre-operation inspection with the program, which allowed the administrator to track all the procedures from start to finish. Third, it made it possible to predict the life of the equipment and verify its reliability based on the data accumulated through the registration of the pictures for improvements, repairs, time required, cost, etc. after the work was completed. It is suggested that these research results will be helpful in the practical and systematic operation of small-and medium-sized PSM workplaces. In addition, it can be utilized in a useful manner for the development and dissemination of a facility maintenance web program when establishing future smart factories in small-and medium-sized PSM workplaces under the direction of the government.

    Management and Use of Oral History Archives on Forced Mobilization -Centering on oral history archives collected by the Truth Commission on Forced Mobilization under the Japanese Imperialism Republic of Korea- (강제동원 구술자료의 관리와 활용 -일제강점하강제동원피해진상규명위원회 소장 구술자료를 중심으로-)

    • Kwon, Mi-Hyun
      • The Korean Journal of Archival Studies
      • /
      • no.16
      • /
      • pp.303-339
      • /
      • 2007
    • "The damage incurred from forced mobilization under the Japanese Imperialism" means the life, physical, and property damage suffered by those who were forced to lead a life as soldiers, civilians attached to the military, laborers, and comfort women forcibly mobilized by the Japanese Imperialists during the period between the Manchurian Incident and the Pacific War. Up to the present time, every effort to restore the history on such a compulsory mobilization-borne damage has been made by the damaged parties, bereaved families, civil organizations, and academic circles concerned; as a result, on March 5, 2004, Disclosure act of Forced Mobilization under the Japanese Imperialism[part of it was partially revised on May 17, 2007]was officially established and proclaimed. On the basis of this law, the Truth Commission on Forced Mobilization under the Japanese Imperialism Republic of Korea[Compulsory Mobilization Commission hence after] was launched under the jurisdiction of the Prime Minister on November 10, 2004. Since February 1, 2005, this organ has begun its work with the aim of looking into the real aspects of damage incurred from compulsory mobilization under the Japanese Imperialism, by which making the historical truth open to the world. The major business of this organ is to receive the damage report and investigation of the reported damage[examination of the alleged victims and bereaved families, and decision-making], receipt of the application for the fact-finding & fact finding; fact finding and matters impossible to make judgment; correction of a family register subsequent to the damage judgement; collection & analysis of data concerning compulsory mobilization at home and from abroad and writing up of a report; exhumation of the remains, remains saving, their repatriation, and building project for historical records hall and museum & memorial place, etc. The Truth Commission on Compulsory Mobilization has dug out and collected a variety of records to meet the examination of the damage and fact finding business. As is often the case with other history of damage, the records which had already been made open to the public or have been newly dug out usually have their limits to ascertaining of the diverse historical context involved in compulsory mobilization in their quantity or quality. Of course, there may happen a case where the interested parties' story can fill the vacancy of records or has its foundational value more than its related record itself. The Truth Commission on Compulsory mobilization generated a variety of oral history records through oral interviews with the alleged damage-suffered survivors and puts those data to use for examination business, attempting to make use of those data for public use while managing those on a systematic method. The Truth Commission on compulsory mobilization-possessed oral history archives were generated based on a drastic planning from the beginning of their generation, and induced digital medium-based production of those data while bearing the conveniences of their management and usage in mind from the stage of production. In addition, in order to surpass the limits of the oral history archives produced in the process of the investigating process, this organ conducted several special training sessions for the interviewees and let the interviewees leave their real context in time of their oral testimony in an interview journal. The Truth Commission on compulsory mobilization isn't equipped with an extra records management system for the management of the collected archives. The digital archives are generated through the management system of the real aspects of damage and electronic approval system, and they plays a role in registering and searching the produced, collected, and contributed records. The oral history archives are registered at the digital archive and preserved together with real records. The collected oral history archives are technically classified at the same time of their registration and given a proper number for registration, classification, and keeping. The Truth Commission on compulsory mobilization has continued its publication of oral history archives collection for the positive use of them and is also planning on producing an image-based matters. The oral history archives collected by this organ are produced, managed and used in as positive a way as possible surpassing the limits produced in the process of investigation business and budgetary deficits as well as the absence of records management system, etc. as the form of time-limit structure. The accumulated oral history archives, if a historical records hall and museum should be built as regulated in Disclosure act of forced mobilization, would be more systematically managed and used for the public users.